Proceedings of the Third National Conference on RECENT TRENDS IN INFORMATION AND COMMUNICATION ...

516
PROCEEDINGS of THIRD NATIONAL CONFERENCE ON RECENT TRENDS IN INFORMATION AND COMMUNICATION TECHNOLOGY (RTICT 2010) 9 10 APRIL 2010 Editors Dr. Amitabh Wahi Mr. S. Daniel Madan Raja Mr. S. Sundaramurthy Organised by DEPARTMENT OF INFORMATION TECHNOLOGY BANNARI AMMAN INSTITUTE OF TECHNOLOGY (An Autonomous Institution Affiliated to Anna University Coimbatore | Approved by AICTE New Delhi | Accredited by NBA New Delhi and NAAC with 'A' Grade | ISO 9000:2001 Certified Institute) Sathyamangalam - 638 401 Erode District Tamil Nadu Phone: 04295-221289 Fax: 04295-226666 Email : [email protected] www.bitsathy.ac.in

Transcript of Proceedings of the Third National Conference on RECENT TRENDS IN INFORMATION AND COMMUNICATION ...

PROCEEDINGS

of

THIRD NATIONAL CONFERENCE ON

RECENT TRENDS IN INFORMATION AND COMMUNICATION

TECHNOLOGY

(RTICT 2010)

9 – 10 APRIL 2010

Editors

Dr. Amitabh Wahi

Mr. S. Daniel Madan Raja

Mr. S. Sundaramurthy

Organised by

DEPARTMENT OF INFORMATION TECHNOLOGY

BANNARI AMMAN INSTITUTE OF TECHNOLOGY (An Autonomous Institution Affiliated to Anna University Coimbatore | Approved by AICTE New Delhi |

Accredited by NBA New Delhi and NAAC with 'A' Grade | ISO 9000:2001 Certified Institute)

Sathyamangalam - 638 401 Erode District Tamil Nadu Phone: 04295-221289 Fax: 04295-226666 Email : [email protected]

www.bitsathy.ac.in

PREFACE

Almost everybody today believes that nothing in economic history has ever moved as fast

as, or had a greater impact than, the Information Revolution. Not all problems have a

technological answer, but when they do, that is the more lasting solution. Information and

communications technology unlocks the value of time, allowing and enabling multi-

tasking, multi-channels, multi-this and multi-that.

Communication has allowed the sharing of ideas, enabling human beings to work

together to build the impossible. Our greatest hopes could become reality in the future.

With the technology at our disposal, the possibilities are unbounded.

This conference acts as a platform for the blazing minds of the nation to disseminate

knowledge in their field of expertise. We are indeed at great pleasure in the response

from various researchers across the country.

We are delighted to acknowledge the sponsors of this conference. The work of the

authors and the reviewers of technical papers is commendable. We sincerely thank the

leading lights who have consented to deliver the lectures. We are grateful to our

Chairman, Dr.S.V.Balasubramaniam for his patronage. We are very much thankful to our

beloved Director, Dr.S.K.Sundararaman for his kind encouragement and support in all

our endeavors. We are indebted to our Chief Executive Dr.A.M.Natarajan for his

ebullient support in all the activities of the department. We are obliged to our Principal

Dr.A.Shanmugam, for his kind and support. We shall extend our hearty appreciation to

the members of the technical committee and the organizing committee for their zeal and

involvement.

Editors

Dr. Amitabh Wahi

Mr. S. Sundaramurthy

Mr. S. Daniel Madan Raja

latha ThiagarajanHema

ORGANISING COMMITTEE

Chief Patrons

Dr. S. V. Balasubramaniam, Chairman, BIT

Dr. S. K. Sundararaman, Director, BIT

Patron

Dr. A. M. Natarajan, CEO, BIT

Advisory Committee

Dr. S. Babusundar, King Khalid University, Saudi Arabia

Dr. P. Nagabhusan, University of Mysore, Mysore

Dr. , NIT, Tiruchirappalli

Dr. K. K. Shukla, BHU, Varanasi

Dr. Umapada Pal, Indian Statistical Institute, Kolkata

Dr. B. H. Shekar, University of Mangalore, Mangalore

Dr. R. Bhakar, GCT, Coimbatore, TN

Dr. V. P. Shukla, MITS Deemed University, Rajasthan

Dr. H. N. Upadhaya, SASTRA University, Tanjore, TN

Dr. A. K. Srivastava, NanoSonix Inc, USA

Dr. M. V. N. K. Prasad, IDRBT, Hyderabad, AP

Mr. Venugopal Kaliyannan, Sun Microsystems, Bangalore

Mr. T. R. Sridhar, Alcatel-Lucent, Bangalore

Mr. M. Karappusamy, Wipro, Bangalore

Mr. K. S. Sundar, Infosys, Mysore

Chairman

Dr. A. Shanmugam, Principal, BIT

Convener

Dr. Amitabh Wahi, Prof. & Head

Organising Secretaries

Mr. S. Sundaramurthy & Mr. S. Daniel Madan Raja

Joint Organising Secretaries

Ms. A. Valarmozhi & Mr. R. LokeshKumar

Programme Committee

Dr. C. Palanisamy, Mrs. D. Sharmila, Mrs. A. Bharathi, Mrs. O. S. Shanmugapriya,

Mrs. M. Alamelu, Ms. N. R. Gayathri, Mrs. G. Srinitya, Mr. P. Sengottuvelan, Mrs. R.

Brindha, Mr. T. Vijayakumar, Ms. K. Gandhimathi, Ms. M. Akilandeeswari, Mr. M.

Ravichandran, Mr. R. Vinoth Saravanan, Mr. T. Gopalakrishnan, Ms. S. Abirami, Ms. M.

Manimegalai, Ms. M. Priya, Mr. R. Lokesh Kumar, Mr. A. Karthikeyan, Ms. A.

Valarmozhi, Mr. S. Karthikeyan, Mr. M. S. Nagarajan

OUR SPONSORS

CSIR, New Delhi

CONTENTS

IMAGE PROCESSING

S.NO PAPER TITLE AUTHOR NAME PAGE NO

1 A Steganography Scheme for

JPEG2000

P.Krishnamoorthy

A.Vasuki.

1

2 Automatic image structure-

texture restoration with text

removal and impainting system

G.Thandavan

D.Mohana Geetha.

6

3 Optical Training Sequence and Precoding Method for MIMO

Frequency-Selective Fading

Channels

J.Jeraldroy. V.Jeyasri Arokiamary.

11

4 Face Recognition from Sparse

Representation

C.K.Shahnazeer.

Jeyavel.J.

17

5 Enhanced Ensemble Kalman

Filter for Tomographic Imaging

D.Dhanya

S.Selvadhayanithy.

24

6 Designing of Meta-Caching

Based Adaptive Transcoding for

Multimedia Streaming

N.Duraimurugan.

Sivakumar.R.

32

7 Digital image processing(dip)

Jpeg compression with client

browser plug-in approach for efficient navigation over internet

Preethy Balakrishnan

,K.Kavin prabhu.

39

8 Image Restoration Using

Regularized Multichannel Blind

Deconvolution Technique

Munia Selvan L,

Vinoth Kumar C

49

9 Text extraction in video

J.Dewakhar,

K.Prabhakaran

55

10 Information retrieval using

moment techniques

N. Anilkumar

Dr.Amitabh Wahi

66

IMAGE PROCESSING

S.NO

PAPER TITLE

AUTHOR NAME

PAGE NO

11

Intelligent face parts generation

system from

Finger prints

A.Gowri

Daniel Madan Raja.S.

73

12

A Survey on Ontology Integration

S.Varadharajan,

Sunitha.R.

83

13

An Efficient Multistage Motion

Estimation Scheme for Video

Compression Using Content Based

Adaptive Search Technique

P.Dinesh Khanna

T.SarathBabu

89

14

Framework for Adaptive

Composition and Provisioning of

Web Services

S.Thilagavathy

R.Sivaraman.

95

15

An Enhanced approach for class-

dependent feautre selection

J.Arumugam

K.Shanmugasundaram.

100

16

Design and implementation of a

multiplier

with spurious power suppression

technique

A.Jeyapraba

108

17

Robotics mobile surveillance using

zigbee

V.Suresh

H.Saravanakumar

H.Senthil kumar.

S.T.Prabhu Sowndharya.

113

18

Effective Electrocardiographic

Measurements of The Heart’s

Rhythm Using Hilbert Transform

P.Muthu

Manoj kumar Varma.

118

19

Visual Analysis for object tracking

using Filters and mean shift

Tracker

Vijayalakshmi.B.

K.Menaga.

121

20

Development of data acqusition

system using serial communication

K.Eunice, S. Smys

128

DATA MINING

S.NO

PAPER TITLE

AUTHOR NAME

PAGE NO

21

Analysis of Health care data using

different classifiers

Soby Abraham

131

22

Mining In Distributed System

Using Generic Local Algorithm

Merin Jojo

V.Venkatesh Kumar.

141

23

Province based web search

M.Gowdhaman

A.Suganthi.

147

24

Divergence Control of Data in

Replicated Database systems

P.K.Karunakaran

R.Sivaraman.

151

25

Similarity profiled spatio temporal association Mining

E.Esther Priscilla

156

26

Effective Integration anyInter

Attribute Dependancy Graphs of

Schema Matching

M.Ramkumar

V.S.Akshaya.

163

27

Designed a focused crawler using

multi-agents system with fault

tolerance involving learning

component

G.DeenaDayalan

T.Goutham,

P.Arunkumar

169

28

Secure banking using bluetooth

B. Karthika, D.

Sharmila

and R.

Neelaveni

173

29

An Artificial Device of

Neuroscience

R.Anand Raj

177

30

Issues in implementing E-learning

in Secondary Schools in Western Maharashtra

Hanmant namdeo

renushe

Abhijit S Desai.

Prasanna R .Rasal

181

BROADBAND COMMUNICATION

S.NO

PAPER TITLE

AUTHOR NAME

PAGE NO

31

Remote patient monitoring –

an

implementation in icu ward

E.Arun, V.Marimunthu,

D. Sharmila

and R. Neelaveni

188

32

An efficient search in unstructured Peer-

to-Peer networks

A.Tamilmaran

V.Vigilesh.

197

33

An Ethical Hacking Technique for

preventing malicious imposter email

Rajesh Perumal.R.

202

34

Sterling network security–

(New

dimension in computer Security)

M.Ranjith

K.Yamini

207

35

Automation of irrigation using wireless

R.Divya Priya,

R.Lavanya

217

36

Key management and distribution for

authenticating group communication

K.S.Krisshnakumar,

M.Jaisakthiraman

225

37

Comparison of cryptographic algorithms

for secure data transmission in wireless

sensor network

R.Vigneswaran,

R.Vinod mohan

232

38

Frequency synchronization in 4G wireless

communication

C.Bhuvaneshwaran,

M.Vasanthkumar

242

39

Network security (Binary level encryption)

D.Ashok

T.Arunraj

245

INFORMATION SECURITY

S.NO

PAPER TITLE

AUTHOR NAME

PAGE NO

40

Improving Broadcasting Efficiency in MANET by

Intelligent Flooding Algorithm

Madala

V.Satyanarayana

253

41

Implementation of SE Decoding Algorithm in

FPGA for MIMO Detection

S. Sharmila

Amirtham, S. Karthie

254

42

Wireless Networks Accident

Response Server “An

Automatic Accident Notification System”

S.Renuga,

R.Ramesh krishnan

257

43

Secure Dynamic Routing in Networks

M.Akilandeeswari,

G.Marimuthu,

V.Saranya, & S.Thiripurasundari

266

44

MA-MK Cryptographic Approach for Mobile Agent

Security with Mitigation of System and Information

Oriented Attacks

N.E.Vengatesh,

K.Jeykumar

271

45

New steganography method for secure data

communication

T.ThambiDurai,

K.Nagaraj

274

46

Audio steganography by reverse forward algorithm

using wavelet

based fusion

P.Arthi

G.Kavitha

281

47

Feature selection for microarray datasets using svm

& anova

Janani.G

A.Bharathi

A.M.Natrajan

288

48

Digital Fragile Watermarking using Integer Wavelet

Transform

C. Palanisamy,

Amitabh

Wahi,V.Kavitha

293

49

Testing Polymorphism in object oriented Systems at Dynamic phases for improving software Quality

Rondla vinod kumar

G.Annapoorani

299

MOBILE COMPUTING

S.NO

PAPER TITLE

AUTHOR NAME

PAGE NO

50

A paper on Efficient Reverse

Turing Test in

Artificial Intelligence

G.Geetha

305

51

Bomb detection using wireless

sensor and neural network

S.Sharon Rosy

Shakena Grace.S.

310

52

Warrior robots

A.Sandeep Kumar & S.M.Siddharth

316

53

Performance Analysis of ATM

networks using modified particle

approach

L.Nitisshish Anand

M.Sundarambal.

M.Dhivya

325

54

Curtail the cost and impact in

wireless sensor network

T.Muniraj

Vijeya kumar.K.

331

55

Data Allocation Optimization Algorithm in Mobile

P.V.Aiswarya

Sabireen.H.

337

56

Greedy Resource Allocation

Algorithm for OFDMA Based

Wireless Network

P.Subramanian

Dhanajay kumar.

346

57

Optimal power path routing in

wireless sensor networks

M.Shrilekha

Sathya priyen.V.B.

351

58

Analysis of Propagation Models

used in Wireless Communication

System Design

K.Phani Srinivas

355

59

GSM Mobile technology for

recharge systems

R.Shanthi,

G.Revathy,

J.Anandavalli

362

60

Mitigating routing misbehaviour in mobile ad hoc Networks

M.Sangeetha

,R.Vidhya prakash

M.Rajesh Babu.

368

MOBILE COMPUTING

S.NO

PAPER TITLE

AUTHOR NAME

PAGE NO

61

Automatic call transfer system Based on

location prediction using Wireless sensor

network

C.Somasudaram

Vairavamoorthy.A.

Jesvin veanc

373

62

An Energy Efficient relocation of gateway for improving Timeliness in Wireless Sensor

Networks

S.S.Rahmath ameena

Dr.B.Paramasivan

376

63

Improving stability of shortest multipath routing

algorithm in manet

S.Tamilzharasu

M.Newlin Rajkumar.

382

64

Broadcast Scheduling in wireless network

A.Grurshakthi meera

G.Prema

387

65

Performance Analysis of Security and QoS self-

Optimization in Mobile AD-Hoc Networks

R.R.Ramya

T.SreeSharmila

392

66

Security in Mobile Adhoc Network

Amit Kumar Jaiswal,

Kamal Kant

Pradeep Singh

396

67

Energy Efficient Hybrid Topology Management

Scheme for Wireless Sensor Networks

S. Vadivelan

, A.

Jawahar

400

68

Rateless Forward error correction using LDPC

for MANETS

N.Khadirkumar

406

69

Application Of Genetic Algorithm For The Design Of Circular Microstrip

Antenna

Rajendra Kumar Sethi

Nirupama Tirupathy.

411

70

Automated Water quality monitoring system

using GSM service

R.Bhaarath Vijay,

A.Kapil

416

71

U Slotted Rectangular Microstrip Antenna For

Bandwidth Enhancement

Shalet Sydney

D.Sugumar

422

MODERN DIGITAL COMMUNICATION TECHNIQUES

S.NO

PAPER TITLE

AUTHOR NAME

PAGE NO

72

Implementation of FIR filter using VEDIC Mathematics in VHDL

Arun Bhatia

427

73

Two microphone enhancement of reverberant

speech by SETA-TCLMS algorithm

M.Johnny

Manikandan.A.

Ashok.M.B

Kashi vishwanath.R.G.

434

74

A Cooperative Communication method for

noise Elimination Based on LDPC Code

S.Sasikala

439

75

Data Transmission using OFDM Cognitive

Radio under interference environment

Vinothkumar J,

Kalidoss R

446

76

Wireless communication of multi-user MIMO

system with ZFBF and sum rate analysis

K.Srinivasan

450

77

Low cost Embedded based Braille Reader

J.Raja,M.Chozhaventhar,

V.Eswaran,K.Vinothkumar

452

78

Unbiased Sampling in Ad-hoc( Peer-to-Peer)

Networks

M.Deiveegan,A.Saravanan,

S.Vinothkannan

457

79

Global chaos synchronization of identical and

different chaotic systems by nonlinear control

V. Sundarapandian,

R. Suresh

464

80

Reducing Interference by Using Synthesized

Hopping Techniques

A.Priya

B.V.Srilaxmi priya.

468

81

Real time image processing

applied to traffic

–queue detection algorithm

R.Ramya , P.Saranya

473

82

A fuzzy logic clustering technique to

analyse holy ground water

D.Anugraha,

Dharani

483

83

Image segmentation using clustering

algorithm in six color spaces

S.Venkatesh kumar

488

84

A Novel Approach for Web Service

Composition in Dynamic Environment Using

AR -based Techniques

Rebecca.R

494

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

A Steganography Scheme for JPEG2000

Krishnamoorthy .P

1, Vasuki .A

2

1 PG Student, 2 Asst.Professor

Department of Electronics and Communication Engineering Kumaraguru College of Technology, Coimbatore-641006

Tamilnadu India.

Email: [email protected], [email protected]

Abstract – Steganography is a technique to

hide secret information in an image. Hiding

capacity is a very important aspect in all

steganographic schemes. Hiding capacity in

JPEG2000 compressed images is very limited

due to bit stream truncation, so it is necessary

to increase the hiding capacity. In this paper,

a new steganography scheme is introduced for

JPEG2000, which uses bit plane coding and

rate distortion optimization technique.

Moreover the embedding points and their

intensity are determined using redundancy

evaluation in order to increase the hiding

capacity and the data is embedded in the

lowest bit plane to keep the message integrity.

The performance of the algorithm are

evaluated for two different images with hiding

capacity in bits and PSNR (dB) of the

reconstructed image as the parameters

Index terms – JPEG2000, bit plane coding,

steganography, data hiding

1. INTRODUCTION

Data hiding in covert communication is

possible due to the characteristics of human

visual system. The hiding technology which is

used in covert communication is named as steganography. Steganography is different from

encryption because encryption hides information

contents whereas steganography hides

information existence. The main aspects that are

considered in data hiding are: capacity, security

and robustness. Capacity means the amount of

data that can be sent secretly, security means the

inability to detect the hidden information and

robustness means the amount of data that a cover

medium can withstand before an intruder can

destroy the hidden information [1]. Robustness is

mainly considered in the digital watermarking

applications rather than steganography. The

steganography algorithms are concerned about

the hiding capacity and security.

The most widely used method is the Least

Significant Bits (LSB) method [2]. Here the

message bits are embedded in the LSB of image

pixel. This method is used because of its

simplicity and ease of implementation. The main

disadvantage is that when we go for truncation

operation then the secret message will be destroyed and thus the extraction of the secret

message is very difficult.

JPEG2000 standard is based on discrete

wavelet transform (DWT) and embedded block

coding and optimized truncation (EBCOT)

algorithms. The bit streams is rate distortion

optimized truncated after bit plane encoding. If

the secret message is embedded directly into the

lowest bit plane of the quantized wavelet

coefficients, then the secret message will be

destroyed by truncation operation [3], [4].

The spread spectrum hiding technique can be

applied in JPEG2000 without considering the bit

stream truncation problem [5]. Here the hiding

capacity is minimized by the spread spectrum pre

processing, so it is often used in digital

watermarking techniques rather than

steganography. After analyzing the challenge of

covert communication, Su and Kuo [6] presented a steganography scheme to hide high

volumetric data in JPEG2000 compressed

images. In order to avoid the bit stream

truncation, the entropy coding was completely

bypassed and is named as ―lazy mode‖. It is

limited to the simplified version of JPEG2000

and not designed for standard JPEG2000.

In this paper a steganography scheme is

proposed for the standard JPEG2000 systems,

2

the bit plane coding procedure is used twice to

solve the problem of bit stream truncation.

Moreover, the embedding points and their

intensity are determined using the redundancy

evaluation technique in order to increase the

hiding capacity. The rest of the paper is structured as follows. Section 2 describes the

details of JPEG2000 standard. Section 3

describes the procedures of proposed

steganography scheme. Section 4 shows the

simulation results. Section 5 gives the

conclusion.

2. JPEG2000 STANDARD

The JPEG2000 encoder is shown in Fig. 1.

First the source image is DC level shifted, by

subtracting all the samples of the image by , where p is the component precision.

Source image

Code stream

Fig. 1. JPEG2000 encoder

The discrete wavelet transform is applied on

the level shifted image coefficients, implemented

by Daubechies 9-tap/7-tap filter [7], which

decompose the image into different resolution

levels. After taking transform all the coefficients

are quantized using different quantization step

size for different levels.

(1)

where

(2)

where BSS is the basic step size, l is the

decomposition level, is the quantization step

size [8], is the transformed coefficient

of the sub band b, is the quantized value.

After quantization each sub band of different

resolution levels are further divided into code

blocks of size smaller than the sub bands. Then

bit plane coding is applied separately on each

code block. It operates on the bit plane in the

order of decreasing importance, to produce an

independent bit stream for each code block. Each

bit plane is encoded in a sequence of three coding passes namely, significance propagation

pass, magnitude refinement pass, clean up pass

[9].

The coefficients of each code block are

scanned as shown in Fig. 2. Starting from the top

left the first four coefficients of the first column

are scanned, then four coefficients of the second

column are scanned until the width of the code

block are scanned and so on. A similar vertical

scan is continued for any leftover rows on the

code blocks in the sub band. The scanning order is same for all the three coding passes.

Fig. 2. Scanning order

After bit plane encoding, rate distortion

optimized truncation is done on coded bit

streams. The EBCOT algorithm produces a bit

stream with many useful truncation points. The

bit stream can be truncated at the end of any

coding passes to get desired compression ratio.

The truncation point of every code block is

determined by rate-distortion optimization [9].

At the decoder side bit plane decoding,

dequantization and wavelet reconstruction is

done to reconstruct the image.

3. PROPOSED SCHEME

The block diagram of the proposed

steganography scheme is shown in Fig. 3. The

steps which are marked by dashed lines in Fig. 3 represent the proposed steganography scheme. The steps involved in the determination of the

embedding points and their intensity for a code

block are as follows:

The wavelet coefficients greater than a

certain threshold are chosen as the candidate embedding points.

Wavelet decomposition

Quantization Bit plane encoding

Rate distortion optimization

3

According to the optimization

algorithm, the lowest bit plane which is

unchanged even after the bit stream

truncation is determined as the lowest

embed-allowed bit plane of that code

block.

Then the embedding points and their

intensity are adjusted on the basis of

redundancy evaluation.

The redundancy of each wavelet coefficient is

calculated in order to determine the embedding

points in each code block. The visual masking

effect and brightness sensitivity are considered in

order to calculate the redundancy of each

coefficient, which are calculated as follows:

First, the Self contrast masking effect of each

coefficient is given by

(3)

where is the quantized wavelet coefficient,

is the quantization step size, the parameter α assumes a value between ‗0‘ and ‗1‘, the typical

value of α is 0.7, is the output of the self-contrast masking effect.

The next step is the neighborhood masking

effect given by

(4)

The neighborhood contains wavelet

coefficients within a window of N by N, centered

at the current position. The parameter β assumes

a value between ‗0‘ and ‗1‘, total number of wavelet coefficient in the neighborhood. The

parameter is a normalization factor with a

constant value of , where β is set to 0.2, d is the bit depth of the image component.

The third step is the brightness sensitivity

determination. denotes the sub band at

resolution level and with orientation

. The symbol

denotes the wavelet coefficient located at (i,j) in

sub band , K is the levels of wavelet

decomposition. The local brightness weighting

factor is given by

(5)

(6)

Before wavelet decomposition the image is

shifted to a dynamic range of [-128 127]. The

local average brightness is normalized by

dividing 128 in (6). The normalized value is

given by

(7)

Finally the redundancy is calculated by

(8)

In order to reduce the image degradation message bits are embedded in the wavelet

coefficients whose redundancy is not less than 2.

The rule of adjustment on embedding points and

intensity is as follows:

1) If , then the candidate embedding point should be removed.

2) If , then the embedding capacity of this point is determined to

be n bits.

Source image

Code stream

Fig. 3. Message embedding

After determining the embedding points, secret

message is embedded in to the selected

embedding points from the lowest embed

allowed bit plane to the higher ones. Secondary

bit plane encoding is done after the message

embedding which is same as that of the previous

one.

Message extraction process is illustrated in

Fig. 4. The first step is the bit plane decoding,

Wavelet decomposition Quantization

Bit plane encoding Rate distortion optimization

Determine embedding points and intensity

Message

embedding

Secondary bit plane

encoding

4

followed by the determination of embedding

point and intensity. Then the secret message is

extracted. Finally dequntization and IDWT is

taken to reconstruct the image.

code stream

Reconstructed image

Fig. 4. Message extraction

4. SIMULATION RESULTS

The original image is the gray scale image

―lena‖ of size 256 x 256 as shown in Fig. 5 (a).

The secret image to be embedded is the binary

image of size 64 x 64 shown in Fig. 5 (b).

The source image is decomposed using four-

level DWT. The coefficients of each sub bands

are quantized and partitioned into code blocks of

size 16 x 16. The bit plane encoding procedure is operated on each code block to produce the bit

streams. According to rate-distortion

optimization, the bit streams of each code block

should be truncated after certain coding passes in

order to get the desired compression ratio. Then

the redundancy evaluation is done to determine

the embedding points in each code block and the

secret message is embedded in those points.

After embedding the secondary bit plane coding

is done. At the receiver side the reverse process

is done to reconstruct the image and also the

secret message is retrieved.

In order to test the performance of the

proposed algorithm, two methods are evaluated.

Method 1: with redundancy evaluation

Method 2: without redundancy

evaluation

The two methods are tested for two different

images with same size. For a threshold of 8, the

hiding capacities of the image ―lena‖, for

different compression ratios are compared in Table 1. Hiding capacity is the total number of

bits that can be embedded in the image. The

embedded secret image is shown in Fig. 5 (b).

The reconstructed image and the retrieved logo is

shown in Fig. 6. After reconstructing the image

the PSNR is calculated between the original and

the reconstructed image.

(9)

where

MSE = (original image—reconst.image)2 (10)

TABLE 1

COMPARISON OF HIDING CAPACITY OF IMAGE

―LENA‖ FOR THRESHOLD ‗8‘

Bits per

pixel

Capacity(bits) PSNR

(dB) Method 1 Method 2

0.4 5381 2117 30.3819

0.6 6270 2502 32.3748

0.8 6576 2640 35.2095

(a) (b)

Fig. 5 (a) Original image (b) logo image

(a) (b)

Fig. 6 (a) Reconstructed image (b) retrieved logo

For the same threshold value the hiding

capacity of the image ―Goldhill‖, for different

compression ratios are compared in Table 2. The

secret message is the same logo image. The

PSNR value of the reconstructed image is also

Bit plane Decoding

Determine embedding points and intensity

Message extraction Dequantization

Wavelet reconstruction

5

calculated. The original and the reconstructed

image is shown in Fig. 7

TABLE 2

COMPARISON OF HIDING CAPACITY OF IMAGE

―GOLDHILL‖ FOR THRESHOLD ‗8‘

Bits per

pixel

Capacity(bits) PSNR

(dB) Method 1 Method 2

0.4 9175 3542 26.9621

0.6 13636 5423 28.1836

0.8 14264 5701 28.9238

(a) (b)

Fig. 7 (a) original image (b) reconstructed image

Thus it is clear from the Table 1 and Table 2

that the hiding capacity of the proposed

algorithm is higher and also the PSNR value of

the reconstructed image will vary.

5. CONCLUSION

In this paper a new steganography scheme is

proposed for the standard JPEG2000 systems,

which uses redundancy evaluation technique to

embed the secret message and thus the hiding

capacity of the JPEG2000 compressed images is

increased. But the bit plane coding procedure is

used twice, which will not have much effect on

the computational complexity.

REFERENCES

[1] N. Provos and P. Honeyman, ―Hide and seek: An introduction to steganography,‖

IEEE Security and Privacy Mag., vol. 1, no.

3, pp.32–44, 2003

[2] C. K. Chan and L. M. Cheng, ―Hiding data

in images by simple LSB substitution,‖

Pattern Recognit., vol. 37, no. 3, pp. 469–

474, 2004.

[3] JPEG2000 Part 1: Final Committee Draft

Version 1.0, ISO/IEC. FCD 15444-1, 2000.

[4] JPEG2000 Part 2: Final Committee Draft,

ISO/IEC FCD 15444-2, 2000.

[5]R. Grosbois and T. Ebrahimi, ―Watermarking in the JPEG2000 domain,‖ in Proc. IEEE

4th Workshop on Multimedia Signal, 2001,

pp.339–344

[6] P. C. Su and C. C. J. Kuo, ―Steganography

in JPEG2000 compressed images,‖ IEEE

Trans. Consum. Electron., vol. 49, no. 4, pp.

824–832, Apr. 2003.

[7] C. Christopoulos, A. Skodas, T. Ebrahimi,

―The JPEG-2000 Still Image Coding

System: An Overview,‖ IEEE Trans.

Consumer Electronics, vol. 46, no. 4, pp. 1103-1127, Nov. 2000.

[8] B.E. Usevitch, ―A tutorial on modern lossy

wavelet image compression: Foundations of

JPEG 2000,‖ IEEE Signal Processing Mag.,

vol. 18, pp.22-35, Sept. 2001.

[9] D. Taubman, ―High performance scalable

image compression with EBCOT,‖ IEEE

Trans. Image Processing, vol. 9, pp. 1158-

1170, July 2000.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

6

AUTOMATIC IMAGE STRUCTURE-TEXTURE RESTORATION WITH TEXT

REMOVAL AND INPAINTING SYSTEM

G. Thandavan, Mrs. D. Mohana Geetha

PG Scholor, Assistant professor

Department of Electronics and Communicaton Engineering

Kumaraguru College of technology, Coimbatore

[email protected] , [email protected]

ABSTRACT - This paper deal with the problem of

finding text in images and with the inpainting

problem. Many of the methods are used for structure

and texture inpaintings. Here we used Field Of

Experts (FOE) method. This new approach provides

a practical method for learning higher-order Makov

Random Field (MRF) models with potential function

that extend over large pixels neighborhoods. These

cliques potential are modeled using the Product of

Experts framework that uses non-linear filter

responses. In contrast to the previous approaches all

parameter, including linear filters themselves, are

learned from training data. Field Of Experts model

mainly used two applications, image denoising and

image inpainting. The novel contribution of this

paper is the combination of the inpainting techniques

with the techniques of finding text in images and a

simple morphing algorithm links them. This

combination of results in a automatic system for text

removal and image restoration that requires no user

interface at all.

Key words: Inpainting, Field of Experts, Text

detection, mathematical morphology.

1. INTRODUCTION

Inpainting is the process of filling in missing

or damaged image information. Its applications

involve include the removal of scratches in a

photograph, repairing damaged areas in ancient

paintings, recovering lost blocks caused by wireless

image transmission, image zooming and super

resolution , removing undesired objects from an

image, even perceptual filtering. Text extraction from

an images and video sequences finds many useful

applications in document processing , detection of

vehicle license plate, analyses of technical papers

with tables, maps, charts, and electric circuits and

content based image/video data base. Educational

training and TV programs such as news contains

mixed text-picture-graphics region. In the inpainting

problem, the user has to select the area to be filled in,

inpainting region, since the area missing or damaged

in an image cannot be easily classified in an objective

way. However, there are some occasions where this

can be done. One such a example detecting and

recognizing these characters can be very important in bi-model search of internet data (image and text), and

removing these is important in the context of

removing indirect advertisement, and for aesthetic

reasons.

This paper deal with a problem of automatic

inpainting- based image restoration after text

detection and removal from `images that require no user interaction. This system is important because the

selection of the area to be inpainted has been done

manually by previous inpainting systems.

2.BACKGROUND AND REVIEW OF

INPAINTING

For the in painting problem it is essential to

proceed to the discrimination between the structure

and the texture of an image. Structure define the main

parts - objects of an image, whose surface is

homogeneous without having any details. Texture

defines the details on the surface of the objects which

make the images more realistic.

2.1. Structure Inpainting

The term inpainting was first used by Bertalmio

et al .The idea is to propagate the isophotes (lines

with the same intensity) that arrive at the boundary of

the inpainting region, smoothly inside the region

while preserving the arrival angle. In the same

context of mimicking a natural process, Bertalmioet al. suggested another similar model, where the

evolution of the isophotes is based on the Navier

7

Stokes equations that govern the evolution of fluid

dynamics.

Apart from physical processes, images can also be

modeled as elements of certain function spaces. An

early related work under the word ”disocclusion”

rather than inpainting was done by Masnou and

Morel . In Chan and Shenderived an inpainting model

by considering the image as an element of the space

of Bounded Variation (BV) images, endowed with

the Total Variation (TV) norm .

2.2. Texture Inpainting

The problem of texture inpainting is highly connected with the problem of texture synthesis. A

very simple and highly effective algorithm was

presented by Efros and Leung . In this algorithm the

image is modeled as a Markov Random Field and

texture is synthesized in a pixel by pixel way, by

picking existing pixels with similar neighborhoods in

a randomized fashion. This algorithm performs very

well but it is very slow since the filling-in is being

done pixel by pixel. In the speed was greatly

improved by using a similar and Simpler algorithm,

which filled in the image in a block by block of pixels way.

2.3. Fields of Experts Model Inpainting

To overcome the limitations of pair wise

MRFs and patch-based models we define a high-

order Markov random field for entire images X using

a neighborhood system that connects all nodes in an

m×m square region. This is done for all overlapping

m×m regions of x, which now denotes an entire

image rather than a small image patch. Every such

neighborhood centered on a node (pixel) k = 1… k defines a maximal clique x (k) in the graph. Without

loss of generality we usually assume that the

maximal cliques in the MRF are square pixel patches

of a fixed size. Other, non-square, neighborhoods can

be used.

(1)

where

This equation (1) that the potentials are defined

with a set of expert functions that model filter

responses to a bank of linear filters. This global prior

for low-level vision is a Markov random field of

“experts", or more concisely a Field of Experts

(FoE).More formally, Eq. (1) is used to define the

potential function (written as factor):

(2)

Each Ji is a linear filter that defines the direction (in

the vector space of the pixel values in x(k)) that the

corresponding expert Ф(.;.) Is modeling, and αi is

its corresponding (set of) expert parameter(s). θ =

Ji,αi | i=1,..N is the set of all model parameters. The

number of experts and associated filters, N, is not

prescribed in a particular way. Since each factor can

be unnormalized, we neglect the normalization

component of Eq. (1) for simplicity. Overall, the Field-of-Experts model is thus

defined as

(3)

All components retain their definitions from

above. It is very important to note here that this

definition does not imply that we take a trained PoE

model with fixed parameters θ and use it directly to

model the potential function. This would be incorrect,

because the PoE model described was trained on

independent patches. In case of the FoE, the pixel

regions x(k) that correspond to the maximal cliques

are overlapping and thus not independent. Instead, we

use the untrained PoE model to define the potentials,

and learn the parameters θ in the context of the full MRF Model. What distinguishes this model is that it

explicitly models the overlap of image patches and

the resulting statistical dependence.

The filters Ji, as well as the expert

parameters αi must account for this dependence (to

the extent they can). It is also important to note that

the FoE parameters θ = Ji,αi | i=1,..N are shared

between all maximal cliques and their associated

factors.The model applies to images of an arbitrary

size and is translation invariant because of the homogeneity of the potential functions. This means

that the FoE model can be thought of as a translation-

invariant PoE model.

3 . NEW SYSTEMS

This system aims in the automatic detection of

text and its removal from images. First we make an

8

initial detection of the text characters in the image.

The algorithm assumes that the intensity of the text

characters is different by their background by at least

a certain threshold, for example 30 in an image with

8-bit intensity, i.e., the characters correspond to the

strong edges of the image. After that a morphological closing operator is being applied and the connected

components that represent regions with text are

derived. The character positions are then derived by a

second thresholding of each connected component

Here need to dilate the text region that was derived

from the algorithm in order to cover all the text

character pixels and only those. This goal leads us to

two morphological operators: Conditional dilation

and reconstruction opening.

Conditional dilation is defined as:

where X is the reference frame inside which

the dilations are allowed, and M plays the role of a

marker to be expanded inside the frame. The

reconstruction opening is the limit of iterative

conditional dilation with the same reference frame

and it is defined as the operator Here apply iterative

conditional dilation but with an adaptive reference

frame

Fig. 1. A new system for automatic text removal and image inpainting

Fig. 2. FOE Model Filters

To calculate the reference frame at each iteration of

the algorithm, the hypothesis that the intensity of the pixels in the characters is different is different by at

least a certain threshold γ than their background and

the hypothesis that the all the pixels that correspond

to characters have similar intensity.

The algorithm is as follows: Let M the binary

picture that represents the initial guess of the text

position, and let u be the original image.

1. Initialization: M1=M

2. Calculation of reference frame Xn:

Calculation of the mean value of the original

image in the mask area:

Dilation with unitary disk B , of the image

Mn

Where γ is the threshold mentioned above (eg. γ=30).

3. Conditional dilation for the new prediction:

Mn+1 = δB(Mn|Xn).

4. Repeat step 2-3 until Mn+1=Mn

Experimentally found that a good empirical value for

the threshold is γ = 30 and that the algorithm

converges quickly, e.g. in less than 10 repetitions.

After the termination of this algorithm we have the

exact positions of

9

Fig3(a) Original image

Fig3(b)Initial Text guess

Fig3(c) Final Text guess

Fig3(d) Reconstructed image

Fig4(a) Original image

Fig4 (b) initial text guess

Fig4 (c) Final Text guess

Fig4(d) Reconstructed image

10

the text characters in the image. We can also apply an

unconditional morphological dilation operation with

a unitary disk, to ensure that we have captured all the

pixels that correspond to text characters In image

inpainting the goal is to remove certain parts of an image, for example scratches on a photograph or

unwanted occluding objects, without disturbing the

overall visual appearance. Typically, the user

supplies a mask, M, of pixels that are to be filled in

by the algorithm.

To perform inpainting, we use a simple

gradient ascent procedure, in which we leave the

unmasked pixels untouched, while modifying the

masked pixels only based on the FoE prior. We can

do this by defining a mask matrix M that sets the

gradient to zero for all pixels outside of the masked region M:

(5)

Here, η is the step size of the gradient ascent

procedure. In contrast to other algorithms, we make

no explicit use of the local image gradient direction;

local structure information only comes from the responses to the learned filter bank. The filter bank as

well as the αi are the same as in the denoising

experiments. A schematic representation of our

system is given in Fig. 1.

4. EXPERIMENTAL RESULTS

This paper present a results of the system that

we described above. It is important to note that, the

system requires no user interaction, since both the

text detection (inpainting region) and the inpainting process are being done automatically. In Fig. 3 we

used texture inpainting and the algorithm area around

the text characters is an area with texture and no

structure at all. In Fig. 4 the image contains both

structure and texture; thus, both the decomposition of

the image in the two components and the separate

processing of each one are needed. In general, an

automatic system is needed to decide about the

appropriate inpainting method In Fig. 3(b) we can

easily recognize the text, but this area does not

contain all the character pixels. After applying the simple recursive algorithm we can see that the final

guess contains all the pixels, and thus the inpainting

algorithms can be applied effectively. When we

apply the inpainting over the initial textguess area,

we see that the whole area is not restored because the

text has not been captured completely. However the

text area that has been captured is restored

satisfactory.

This is not the case in Fig. 4. As we can see,

the image inpainting, when the inpainting area is the initial text gives very poor results. The reason is that

in this example we applied simultaneous structure

and texture inpainting. Structure inpainting uses the

neighboring pixels to generate new information.

Since the initial text guess has not captured all the

text-pixels, the neighboring pixels are still text-pixels

and thus they don’t generate the appropriate

information characters, and therefore this area

vanished gradually from the reference frame

5. CONCLUSION

This paper dealt with the combined

problems of inpainting and finding text in images.

Have proposed a new simple morphological

algorithm inspired from the reconstruction opening

operation. This algorithm captures all the pixels that

correspond to text characters and thus its output can

be considered as the inpainting region. By applying

then an appropriate inpainting method, have

developed a system for automatic text removal and

image inpainting.

6. REFERENCES

1. A.Pnevmatikakis and Petros Maragros “An

Inpainting system for Image structure-

texture restortion with text removal” IEE

.Trans.Im.Proc-2008.

2. Stefan Roth and Michael J. Black “ Field of

Experts”. Int.J.Comput.Vision, Nov-2008

3. Y. Hasan and L. Karam, “Morphological

text extraction from images,” IEEE

Trans.Im. Proc., vol. 9, no. 11, pp. 1978–

1983,2000. 4. M. Dimmiccoli and P. Salembier,

“Perceptual filtering with connected

operators and image inpainting,” in Proc.

ISMM,2007.

5. M. Bertalmio, L. Vese, G. Sapiro, and S.

Osher, “Simultaneous structure and texture

image inpainting,” IEEE Trans. Im.

Proc.,vol. 12, no. 8, pp. 882–889, 2003..

6. A. Criminisi, P. Perez, and K. Toyama,

“Object removal by exemplar based

inpainting,” in Proc. IEEE-CVPR, 2003.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

11

Optimal Training Sequence and Precoding Method for

MIMO Frequency-Selective Fading Channels

V.JEYASRI AROKIAMARY1,

J.JERALDROY

2

1 Asst. Professor, 2 PG Scholar

Kumarguru College of Technology Department of Electronics and Communication Engineering

Coimbatore-641006, Tamilnadu,India.

E-mail: [email protected], [email protected]

Abstract—A new affine precoding and decoding

method for multiple-input multiple-output (MIMO)

frequency-selective fading channels is proposed.The

affine precoder consists of a linear precoder and a

training sequence, which is superimposed on the

linearly precoded data in an orthogonal manner and the

optimal power allocation between the data and training

signals is also analytically derived.This method is better

compared to other affine precoding methods with

regard to source detection performance and

computational complexity.

Index Terms—Affine precoding, decoding, MIMO

channel estimation, source detection, training Sequence.

I. INTRODUCTION

Multi-input and multiple-output wireless

communications systems is used to achieve the very

high data rates over the wireless links . A major

challenge is how to effectively recover the data trans-

mitted over an unknown frequency-selective fading

channel. Then the training sequence is used for

identifying the unknown channel . But it cannot be

used to guard against the effects of frequency

distortion caused by multi-paths. To avoid these

effects, Redundancy is introduced (Redundancy is

introduced to avoid the probability error) by linearly precoding the source data prior to transmission over

the channel .But the linear precoder can guard against

channel spectral nulls [1], [2], it can be inefficient in

identifying the channel . Therefore both the training

sequence and linear precoder are simultaneously

needed.

In this paper, we propose a new affine precoding method for MIMO frequency-selective fading

channels, Prior to the channel estimation, the

proposed method completely eliminates the linearly

precoded data from the received signal in a simple

manner, so that the channel is effectively estimated.

In contrast to the conventional approaches where the

source data vector s is precoded by a “tall” matrix F

to obtain the precoded dataFs, in our approach a

source data matrix S is first built from the source data

vector and then is postmultiplied by a “fat” pre-coder

matrix P. This yields the precoded data matrix SP,

which is later decoded by postmultiplying with a

decoder matrix QE, designed such that PQE=0. As a

result, the unwanted interference from the unknown

precoded data in the channel estimation process is

easily eliminated. Then, in the source detection

process, the presence of the introduced training

sequence C is completely nulled out by postmultiplying the received signal by another

decoder matrix QD, designed such that CQD=0. This

proposed “postmultiplying”-based decoder is much

more flexible than the previously developed “pre-

multiplying”-based decoders.To further improve the

performances of the channel estimation and source

detection, the optimal power allocation between the

training sequence and data signal is also addressed.

Notation:

Superscript T and H means transposition and the

Hermitian adjoint operators, respectively,while

A-H = (A-1)H .

The symbol ⊗ stands for the Kronecker product.

IN is the N x N identity matrix and the 0N x M

zero matrix.

E is the expectation operation.

tr is the trace of matrix A .

The vectorization operator on a matrix to form a

column vector by vertically stacking the matrix’s

columns is denoted by vec(⋅) .

II. SYSTEM MODEL

Consider the frequency selective block-fading

channel with t transmit antenna and r receive

antennas. The equivalent baseband channel can be

described by the channel coefficients that are constant

during each transmission block and may change from

one block to the next. Let L be the channel order and

the vector hji=[hji(0),…,hji(L)]T

represent the

12

I/P Bit stream

(A). Transmitter

O/P Bit stream

(B). Receiver

Fig.1. System Model

transmission link between the jth receive antenna and

the ith transmit antenna, j=1,2,…,r; i=1,2,…,t. Under

the previous precoding methods , the source symbol

s is pre-multiplied by a “tall” precoder matrix F

before being superimposed with a training vector t.

The received signal is y=HFs+Ht+w, where

represents the effect of additive white Gaussian noise

(AWGN) and is the so-called channel matrix, which

is created from hji . As the precoder matrix F is

Located between the matrix H and the source s, it is

not easy to eliminate the undesired interference of the

superimposed training during in the symbol detection

process. Moreover, as the training sequence and the

precoded data together share a limited power budget,

the tradeoff in power allocation between the data and

the training signals is very important. An increase of

the training power leads to a better channel

estimation but may degrade the source detection due

to the decreased power for the data transmission.

Such a trade]. For a SISO channel, the approximation

AAH=cs I for a “tall” orthogonal matrix and the

assumption that the to-be-estimated matrix H must be

invertible and well conditioned.

In our method, all of the above disadvantages are

overcome, i.e., the unknown data is easily eliminated

in the channel estimation, while the training signal is

completely nulled out in the symbol detection

process. The optimal power allocation between data

and training signals is also derived to enhance the

data detection performance. The key novelty in our

method is that precoding the source data is carried

out by postmultiplying it with a precoder matrix,

which then offers a greater flexibility for channel

estimation and source detection.Our proposed

method is now described by referring to Fig. 1. In

each transmission block, the source sequence is

divided into k vectors of fixed length p, each is

denoted by si, i=1,…,k. The k source vectors si are

gathered to form the following source data matrix:

S=[s1,s2,…,sk]p*k.This source data matrix is

precoded by postmultiplying it with a precoder

matrix

P= k*(k+n), s p (1)

Form the

source

data

matrix

(pp x k)

Precoding

by post-

multiplying

wi th matrix

(Pk x (k+n))

Transmit

Antennas

Adding

Training

matrix

(Cp x(k+n))

Receive

Antennas

Decoupling

for

channel

estimation

Decoupling

for

Data

detection

Channel

detection

Data

detection

+

13

to obtain the precoded data

X=SP=[x1,x2,…,kk+n]p*(k+n).

Here p1,p2,…,pk1*(k+n). are row vectors and n is the

number of redundant vectors introduced by the

precoder. The training matrix is given by

C=[c1,c2,…,ck+n]p*(k+n).

Is added to X to form the transmitted signal

U=C+X=[u1,u2,…,uk+n]p*(k+n).

As the transmitted signal is U is a sum of the training

signal C and the precoded data X, the orthogonality

between C and P is desirable for effectively decoupling them in channel estimation and data

detection. To avoid inter-block interference

(IBI),each p-dimensional transmitted signal vector ui

is divided into t sub-vectors of length m=p/t >=1.

Each sub-vector is then appended with a sequence of

zeros before being transmitted over one of the

transmit antennas.

W=[w1,w2,…,wk+n]p(L+m)r*(k+n).

Be the matrix that represents the effect of AWGN.

Define

Hji=(L+m)m (2)

Which is a Toeplitz matrix constructed from hji . The

overall channel matrix H p(L+m)r*p can be

represented as

H=

It follows that the received signal matrix

Y (L+m)r*(k+n). The transmitted signal U, the noise

W and the channel H are related by the following

equation:

Y=HU+W=HC+HSP+W

A. The channel estimation problem

At the receiver , the matrix Y is postmultiplied

by an estimator matrix QE (k+n)*p, designed such

that

PQE=0kxp (5)

and QE

HQE=Ipxp (6)

i.e., the columns of QE from an orthogonal basis for

the null space of P. The result of this

postmultiplication is

YQE=HCQE+HSPQE+WQE

=HCQE+WQE. (7)

Condition (5) is to ensure that the channel can be

estimated without the presence of the interference

from the data, while condition (6) ensures that the

noise is not amplified with the multiplication operation. Based on the observation YQE in (7), the

problem is now how to design the training matrix and

the estimator matrix QE such that the channel

estimation error is minimized under the mean-

squared error (MSE) criterion.

B. The source detection process

For data detection, consider the system model in

(4).By postmultiplying both sides of (4) by a decoder

matrix

QD(k+n)*k such that

CQD=0pxk and PQD=Ik (8)

the superimposed training signal is easily eliminated

from the received signal. In particular, the equivalent

system for data detection is given by

YQD=HS+WQD (9)

As can be seen from (8), the decoder matrix QD is

the postpseudoinverse of P and it is contained in the

null-space of C. It is given as

QD=PH(PPH)-1 (10)

With P k*(k+n) designed such that

CPH=0pxk

PCH=0kxp (11)

i.e.,CH is contained in the null-space of P.

14

Equation (9) then becomes

YPH(PPH)-1=HS+WPH(PPH)-1 (12)

Define

Y=YPH(PPH)-1, W=WPH(PPH)-1. (13) (12) is written as

Y=HS+H (14)

A popular measure of the detection performance is

the meansquared error of the source symbol, defined

as

ɛS(QD)=trE(S-S)(S-S)H (15)

where S is the estimate of S. The objective is to find

QD to minimize ɛS(QD)..

Another measure is the effective SNR of the system

(14), defined as

SNR= (16)

One is then interested in finding QD to maximize the

above SNR.It will be shown in Section IV

performance indexes lead to the same design of the

optimal precoder P matrix .

C .The power allocation problem

Let denote the estimate of H . Then the channel

estimation error is

=H- .

Rewrite (14) as

=( + )S + = S + S + . (17)

Since the channel is estimated by the MMSE

estimator, H and H are statistically uncorrelated, and

so are HS and HS [4]. The term Z = HS +W is thus

considered as noise and it is statistically uncorrelated

with the signal HS . It follows that the effective SNR

of the system in (17) can be defined as

SNReff= (18)

As will be seen shortly, this effective SNR depends

on the power allocation between the source data S

and the training signal C. This implies that the

detection performance of the source data S can be

improved by maximizing this effective SNR as a

function of the training power.

III. TRAINING DESIGN AND ALGORITHM FOR

CHANNEL ESTIMATION

This section solves the problem of designing the training matrix C and the estimator matrix QE

that minimize the meansquared error between the

channel vector

h = [ , ,…, , ,…, ,…, ,…, ]T (19)

and its linear estimate . The minimization is under

the constraints

(5) and (6). Define

YQ = YQE (L+m)r*p

CQ = CQE

p*p

WQ = WQE (L+m)r*p

the system model in (7) for channel estimation can be

written as

YQE = HCQ +WQ (20)

Observe from (2) and (3) that the matrix H is sparse

and it is linearly dependent on the channel vector h.

Let p be the jth column of

CQ , j=1,….,p. Divide each vector into t sub-

vectors = = [ , ,…, ]T m,

i=1,2,…,t. Tthen construct a Toeplitz matrix from

as follows:

ji=(L+m)*(L+1) (21)

Equation (20) can now be rewritten as

y = h + (22)

where

= vec(YQ) (L+m)r*p

15

= vec(WQ) (L+m)r*p

Let

j = [ j1 j2 … jt ] (L+m)*(L+1)*t

(23)

= (L+m)pr*(L+1)rt (24)

Equation (22) becomes

h + (25)

On the other hand, with w = vec(W) is given by

= vec(WQ) = ( I(L+m)r) w.

The linear MMSE estimate of h is the solution of the

following optimization problem:

=

= tr

= (26)

Accordingly, in our design the matrices C and QE

are chosen through an arbitrary orthogonal matrix

O (k+n)*(k+n)

by the following expressions:

C = c O (1:p,:) p*(k+n)

(27)

QE = O (1:p,:)H (k+n)*p

(28)

Where O (1:p,:) denotes the 1st to pth rows of O.

With the above designs of C and QE , the equality

CQE CH = CCH

holds true (although QE generally is not the

identity matrix). This means that .

tr = tr = p(k+n) .

Condition (6) is obviously satisfied, while condition

(5) is also satisfied by the design of the optimal

precoder matrix P in Section IV.

Furthermore, since CQE = c , it is clear

that the matrix ji defined by (21) satisfies the

equation shown at the bottom of the previous page.

Therefore, for the matrix defined by (24), one can

readily verify that

H = Ir

= Ir

= m(k+n) I(L+1)rt

which is a scaled identity matrix

Conditions (5) and (11) are automatically satisfied

with the design of the precoder matrix p in Section

IV. LINEAR PRECODER DESIGH FOR DATA

DETECTION

We now address the problem of designing

the precoder matrix P that satisfies conditions (5) and (11) and minimizes the MMSE in (15).

With the same orthogonal matrix

O ∊ as in the definition (26) of the

training matrix C,

let O((p+1):(p+k),:) denote the (p+1)th to (p+k)th rows of O. It is clear that

P=

Finally, it should be pointed out that, as C in (27)

contains p rows of O , while P in (29) contains k rows of O, the size (k+n) of O must be greater than or

equal to (k+p) . This results in the condition that n ≥ p

in (1) required in our method.

V. POWER ALLOCATION FOR DATA

DETECTION ENHANCEMENT

In this section, we derive the optimal power

allocation between the training sequence and the

source data such that SNR in (18) is maximized.

eff =

16

VI.SIMULATION RESULTS:

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 130

32

34

36

38

40

42

44

46

48

50

sigma C

Eff

ecti

ve S

NR

Eff SNR Vs Sigc

SNR=20dB

Fig. 2. Impact of the power allocation in the proposed design on the S Reff performance at different channel SNR

values (SNR = 5,10,and 15 dB): 2 x 2 MIMO frequency-selective fading channel with L=2, p = k = n = 4.

= , where

(30)

eff is actually a quadratic function in :

eff ( ) (31)

Where is given by

(32)

VII. CONCLUSION

We have presented a new method of designing an

affine precoder for MIMO frequency-selective fading

channels based on the concept of orthogonal

superimposed training and linear precoding.In the

proposed design method, the channel is estimated

independently of the affine precoded data while the

interference of the superimposed training signal in

data detection is completely eliminated. Simulation

results show that the proposed method outperforms

the other recently-proposed method in source data detection while its computational complexity is much

reduced.

REFERENCES

[1] G. D. Forney and M. V. Eyuboglu, “Combined

equalization and coding using precoding,” IEEE

Commun. Mag., vol. 29, no. 12, pp. 25–34,Dec. 1991.

[2] G. B. Giannakis, “Filterbanks for blind channel

identification and equalization,” IEEE Signal

Process. Lett., vol. 4, no. 6, pp. 184–187,Jun. 1997. [3] B. Hassibi and B. M. Hochwald, “How much

training is needed in multiple-antenna wireless

links?,” IEEE Trans. Inf. Theory, vol. 49, no. 4, pp.

951–963, Apr. 2003.

[4] X. Ma, L. Yang, and G. B. Giannakis, “Optimal

training for MIMO frequency-selective fading

channels,” IEEE Trans. Wireless Commun., vol. 4,

no. 2, pp. 453–466, Mar. 2005.

[5] J. H. Manton, I. V. Mareels, and Y. Hua, “Affine

precoders for reliable communications,” in Proc.

IEEE Int. Conf. Acoustics, Speech, Signal Processing, Jun. 2000, pp. 2749–2752.

[6] J. H. Manton, “Design and analysis of linear

precoders under a mean square error criterion, Part I:

Foundations and worst case design,” Syst.Control

Lett., vol. 49, pp. 121–130, 2003.

[7] J. H. Manton, “Design and analysis of linear

precoders under a meansquare error criterion, Part II:

MMSE design and conclusions,” Syst.Control Lett.,

vol. 49, pp. 131–140, 2003.

[8] S. Ohno and G. B. Giannakis, “Superimposed

training on redundant precoding for low-complexity

[9] S. Ohno and G. B. Giannakis, “Optimal training and redundant precoding for block transmissions with

application to wireless OFDM,”IEEE Trans.

Commun., vol. 50, no. 12, pp. 2113–2123, Dec. 2002.

[10] D. H. Pham and J. H. Manton, “Orthogonal

superimposed training on linear precoding: A new

affine precoder design,” in Proc. IEEE 6th Workshop

Signal Processing Advance inWireless

Communication, Jun.2005, pp. 445–449.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

17

FACE RECOGNITION FROM SPARSE REPRESENTATION

1Shahnazeer C K

2Jayavel J

1 PG scholar, Dept of IT, Anna University, Coimbatore-47, Tamilnadu, India.

Email:[email protected] 2Lecturer, Dept of IT, Anna University, Coimbatore-47, Tamilnadu, India,

Email: [email protected]

Abstract - This paper provides a problem of automatically recognizing human faces from frontal views with

various facial expressions, occlusion, illumination and pose. There are two underlying motivations for us to write

this paper: the first is to provide an occlusion and various expressions of the existing face recognition and the

second is to offer some insights into the studies of pose and illumination of face recognition. We present a

mathematical formulation and an algorithmic framework to achieve these goals. The existing framework offers a

sparse representation of the test image with respect to the training image. The sparse representation can be

accurately and efficiently computed by the l1

minimization. The proposed framework offers a local translational

model for deformation due to pose with a linear subspace model for lighting variations and also gives competitive

performance for moderate variations in both pose and illumination. Extensive experiments on publicly available

databases verify the efficacy of the proposed method and support the above claims. Index Terms- Face Recognition, Occlusion, Illumination, Pose, Sparse representation, l1-minimization

I INTRODUCTION

Real-world automatic face recognition systems

are confronted with a number of sources of

within-class variation, including pose, expression, and illumination, as well as

occlusion or disguise. Several decades of intense

study within the pattern recognition community have produced numerous methods for handling

each of these factors individually.

In this paper, we exploit the discriminative

nature of sparse representation[2] to perform classification. We represent the test sample in an

over-complete dictionary whose base elements

are the training samples themselves. If sufficient training samples are available from each class, it

will be possible to represent the test samples as a

linear combination of just those training samples from the same class. This representation is

naturally sparse, involving only a small fraction

of the overall training database. We argue that in

many problems of interest, it is actually the sparsest linear representation of the test sample

in terms of this dictionary and can be recovered

efficiently via l1-minimization [3]. Seeking the

sparsest representation therefore automatically

discriminates between the various classes

present in the training set. Sparse representation also provides a simple and surprisingly effective

means of rejecting invalid test samples not

arising from any class in the training database:

these samples sparsest representations tend to

involve many dictionary elements, spanning

multiple classes. We investigate to what extent accurate

recognition are possible using only 2D frontal

images. More specifically, we address the following problem: Given only frontal images

taken under several illuminations recognize

faces despite large variation in both pose and

illumination. Our algorithm will apply to test images with

significantly different illumination conditions,

and pose variations upto ±45. In this setting, a typical test image would have an arbitrary pose

in the given range and also an illumination not

present in the training. Our approach is simple but effective: for each small patch of the test

image, we find a corresponding location and

illumination condition that best approximate it.

The quality of this match is used directly as a statistic for classification, and results from

multiple patches are aggregated by voting. For

comparison, we also propose a second scheme that performs pose-invariant recognition

between the test image and a training image

synthesized from the recovered illumination conditions, by matching deformation-resistant

features such as SIFT keys. We will motivate

and study this new approach to classification

18

within the context of automatic face recognition.

Human faces are arguably the most extensively studied object in image-based recognition. This

is partly due to the remarkable face recognition

capability of the human visual system [4] and

partly due to numerous important applications for face recognition technology [5]. In addition,

technical issues associated with face recognition

are representative of object recognition and even data classification in general.

II REVIEW OF EXISTING SYSTEM

A basic problem in object recognition is to use

labeled training samples from k distinct object classes to correctly determine the class to which

a new test sample belongs. We arrange the given

ni training samples from the ith class as columns

of a matrix Ai= [vi,1,vi,2,……,vi,n]εIRmxn

i . In the context of face recognition, we will identify a w

x h grayscale image with the vector vεIRm

(m=wh) given by stacking its columns; the columns of Ai are then the training face images

of the ith

subject.

A Sparse Representation of training samples

In this section, the training images were

presented in matrices form. It performed a linear feature transform. Given sufficient training

samples of the ith

object class, Ai

=[vi,1,vi,2,….,vi,ni]ε IRmxni

, any new (test) sample yε IR

m from the same class will approximately

lie in the linear span of the training samples

associated with object i: y= αi,1vi,1+αi,2vi,2+….+αi,nivi,ni for some

scalars,αi,jεIR, j = 1, 2, . . . , ni. Since the

membership i of the test sample was initially

unknown, and defined a new matrix A for the entire training set as the concatenation of the n

training samples of all k object classes:

A=[A1,A2,….,Ak]=[v1,1,v1,2,…..,vk,nk]. Then, the linear representation of y can be rewritten in

terms of all training samples as y =Ax0 ε IRm,

where x0=[0,….,0,αi,1,αi,n,0,….,0]T

ε IRn is a

coefficient vector whose entries were zero except those associated with the i

th class.

B Recognition with facial features

In this section, the role of feature extraction

within the new sparse representation framework

for face recognition was reexamined. One

benefit of feature extraction, which carried over to the proposed sparse representation

framework, was reduced data dimension and

computational cost. Our SRC algorithm tested

using several conventional holistic face features, namely, Eigenfaces [6], Laplacianfaces [7], and

Fisher faces, and compares their performance

with two unconventional features: random faces and down-sampled images. In this section, the

stable version of SRC in various lower

dimensional feature spaces were used for solving the reduced optimization problem with

the error tolerance ε=0.05. The Mat lab

implementation of the reduced (feature space)

version of Algorithm 1 took only a few seconds per test image on a typical 3-GHz PC.

C Handling corruption and occlusion.

Occlusion poses a significant obstacle to robust

real-world face recognition. This difficulty is mainly due to the unpredictable nature of the

error incurred by occlusion: it may affect any

part of the image and may be arbitrarily large in

magnitude.

Now, to show how the proposed sparse

representation classification framework can be extended to deal with occlusion. Assume that the

corrupted pixels are a relatively small portion of

the image. The error vector e0, like the vector x0, then has sparse nonzero entries. Since y0= Ax0,

we can rewrite y = y0 +e0 =Ax0+e0 as

Y= [A, I]

0

0

e

x= Bw0.

Here, B= [A, I]ε IR mx(n+m)

, so the system y =Bw

is always underdetermined and does not have a unique solution for w. However, from the above

discussion about the sparsity of x0 and e0, the

correct generating w0=[x0, e0] has at most ni +ρm nonzeros. We might therefore hope to recover

w0 as the sparsest solution to the system y =Bw.

In fact, if the matrix B is in general position, then as long as y =Bŵ for some ŵ with less than

m/2 nonzeros, ŵ is the unique sparsest solution.

Thus, if the occlusion e covers less than (m- ni

)/2 pixels, ≈50 percent of the image, the sparsest

19

solution ŵ to y =Bw is the true generator,

w0=[x0, e0]. More generally, one can assume that the corrupting error e0 has a sparse

representation with respect to some basis Ae ε

IRmxn

e . That is, e0 =Aeu0 for some sparse vector

u0 ε IRm. Here, choosing the special case Ae =I ε

IRmxm

as e0 is assumed to be sparse with respect

to the natural pixel coordinates. If the error e0 is

instead sparser with respect to another basis, the matrix B can simply redefine by appending Ae

(instead of the identity I) to A and instead seek

the sparsest solution w0 to the equation: y =Bw with B =[A, Ae] ε IR

mx(n+ni).

In this way, the same formulation can handle

more general classes of (sparse) corruption. As before, to recover the sparsest solution w0 from

solving the following extended 11-minimization

problem: (l1e ) :ŵ1 =arg min||w||1 subject to Bw=

y. That is, in Algorithm 1, now replace the

image matrix A with the extended matrix B =[A,

I] and x with w =[x, e].

Clearly, whether the sparse solution w0 can be

recovered from the above 11-minimization

depends on the neighborliness of the new polytope P =B (P1)=[A, I](P1). This polytope

contains vertices from both the training images

A and the identity matrix I. The bounds given in imply that if y is an image of subject i, the 1

1-

minimization cannot guarantee to correctly

recover w0 = [x0, e0] if ni +|support(e0)| > d/3. Generally, d » ni, so, c .m < t < [(m+1)/3]

implies that the largest fraction of occlusion.

Algorithm 1 below summarizes the complete recognition procedure. Our implementation

minimizes the l1-norm via a primal-dual

algorithm for linear programming.

Algorithm1. Sparse Representation-based

Classification (SRC)

1. Input: a matrix of training samples A =

[A1,A2, . . .,Ak ]εIRmxn

for k classes, a test sample xεIR

m, (and an optional error

tolerance ε> 0.)

2. Normalize the columns of A to have unit l2-norm.

3. Solve the l1-minimization problem:

ŷ1=arg miny ||y||1 subject to Ay=x (Or alternatively, solve ŷ1=arg miny ||y||1

subject to ||Ay-x||2 <= ε.)

4. Compute the residuals ri(x)=||x-Aδi

(ŷ1)||2 for i = 1, . . . ,k.

5. Output: identity(x)= arg mini ri(x).

III PROPOSED SYSTEM

In this section, we discuss the difficulties

associated with variations in pose and

illumination, and why state-of-the-art methods that are quite effective at handling one of these

modes of variability tend to fail when both are

present simultaneously [8].

We assume that the face is well-localized in the test image, i.e., any gross misalignment or

scaling has already been compensated for. In

practice, this is accomplished by applying an affine or projective transformation to

approximately map eye and mouth corners of the

test image to eye and mouth corners of training image. We will denote the so-aligned test image

by . We will assume access to a

gallery of registered frontal training images

for each subject i, taken at k

distinct illuminations.

The figure 1 represents the system architecture

of face recognition. It has two phases: (i) Training Phase (ii) Testing Phase. In training

phase there can be four or more input face

images. These inputs extract the features that are expression, occlusion, illumination and pose.

These details were stored in a database. Put one

image in the testing phase and it will compare features on the database. If the feature matches

with the test image then it will display the

identified face. Otherwise it will not display any

image, because there is no corresponding image in the database.

20

Fig 1: System Architecture

A Matching Image Patches

We first select a set of feature point locations

in the test image for matching. In

Section , we will see that the choice of feature points is not essential – a sufficient number of

randomly selected or evenly distributed feature

points work as well as the scale-space extrema

popular in the image matching literature [9].

Around each feature point in the test

image, we select a square window

of pixels for matching. If this feature is also visible in the training frontal view, there is a

corresponding point in each training

image . If all images were taken under

ambient illumination, we would expect the

corresponding patches to be quite similar in

appearance: .

However, if the test image is taken at more extreme illumination, it is more appropriate to

approximate the patch by a linear

combination of training patches:

Now, since gross misalignments and scalings

have already been compensated for, we may

assume that the corresponding point lies in a

small neighborhood of .

For each training subject i, we search for the

best match within this neighborhood, seeking a

point where the subspace spanned by the

training patches offers the best approximation to

:

Thus, for each feature denotes the

location of the best match to a linear

combination of training patches for subject i,

while denotes the corresponding best

coefficients.

B Classification Based On Nearest Subspace

Matching We compare two methods for determining the identity of the test image from the

correspondences obtained from (1). The first,

which we refer to as Nearest Subspace Patch Matching (NSPM), uses the approximation error

in (1) as a statistic for classification. The m-th

test patch is classified to the subject

i whose coefficients best approximate it.

The classifications of these patches are then

aggregated into a single classification of the test

image by voting. This process is summarized as

Algorithm 2. The algorithm is computationally efficient and scalable – the complexity is linear

in the number of subjects. A similar scheme has

achieved state-of-the-art performance for recognizing handwritten digits [10]. In this

Section, we will see that this simple approach

also performs quite well for face recognition with moderate variations in pose and

illumination.

C Classification Based On SIFT Matching

For purposes of comparison, we also outline a

second approach which uses the computed α to compensate for global illumination variations,

and then applies standard matching techniques.

This approach finds for each training subject i, a single illumination condition (expressed as

Input image1

Input imageN

Input image

Feature Extractor

Feature Extractor

Feature Extractor

Features of

Faces

(Various

expression,

Occlusion, Illuminatio

n and

poses) TESTING PHASE

TRAINING PHASE

Data base Comparison

using SRC

and NSPM algorithm

Correctly Identified

Found

21

coefficients ) that best reproduces the test

illumination. These coefficients are used to synthesize a set of “equivalent” frontal training

images . In the next paragraph, we describe

how the can be robustly computed from

noisy data. With illumination compensated for in

this manner, classical deformation-invariant

matching algorithms can then be applied. In practice, we find that once illumination has been

corrected, the SIFT (Scale-invariant feature

transform) algorithm[11] performs quite

successfully in matching the test image to the

synthetic image for the correct subject. The

number of SIFT correspondences between and

each of the provides a simple statistic for

classification. Equation (1) gives, for each training subject i, a set of putative

correspondences and

coefficient vectors . For each subject

i, we will synthesize a frontal training image

whose illumination best approximates the

illumination in the test image. That is, from each

set of coefficient vectors we

will compute a single and set

. If the correspondences

were perfect, and there were no deviations from

linearity, one could simply set to be the

average of . Here we use clustering to

remove outliers in fitting a single, global

illumination model. With illumination compensated for, conventional feature matching

techniques can be effectively applied to the test

image and the synthesized training image.

Algorithm 2.Nearest-Subspace Patch

Matching (NSPM)

1. Input: test image , frontal training

images taken at k

different illuminations, for each

subject i.

2. Select feature points in

3. for each feature point do

4. for each subject i, compute the

corresponding location and

coefficients from (1);

5. set Identity( ) to be:

6. end for

7. Output: class i that maximizes the

number of with Identity ( ) =

i.

IV SIMULATION RESULTS

In this work, we have used the ORL database, a

set of pictures taken at Olivetti Research

Laboratory. There are ten different images of each of 40 distinct subjects. For some subjects,

the images were taken at different times, varying

the lighting, facial expressions and facial details. All the images were taken against a dark

homogeneous background with the subjects in

an upright, frontal position. The images are 256

gray levels with a resolution of 92x112 pixels. This database includes the manually made

occluted images.

Output Screen

Fig 2: Form

The figure 2 represents the output screen of the

recognition system. This form contains training,

input image and testing buttons.

22

Training

Fig 3: Training Images

The figure 3 represents the training images of

the ORL database. By clicking training, the message “training is done” appears in the

command window.

Input Image

Fig 4: Input Image

The figure 4 represents the input image of the

recognition system. Clicking input image will display the ORL database, from this select the

input image.

Various Expressions

Fig 5: Various Expressions

The figure 5 represents the output of the various expression face image. Click testing, then

recognized image will display in the screen and

a message box also appeared that is “Correctly

Recognized”.

Occlusion Images

Fig 6: Occluted facial image

The figure 6 represents the output of an occluted

face image. Click testing, then recognized image will display in the screen and a message box also

appeared that is “Correctly Recognized”.

Illumination

Fig 7: Illumination

The figure 7 represents the output of an

illumination face image. Click testing, then

recognized image will display in the screen.

Pose Detection

Fig 8: Pose Detection

The figure 8 represents the output of a pose face image. Click testing, then recognized image will

display in the screen.

23

Incorrectly recognized image

Fig 9: Incorrectly recognized image

The figure 9 represents the output of the

incorrectly recognized image. Select an image

which is not in the database. Then image not

found will display in the screen. Images only

recognized in the database.

V CONCLUSION AND FUTURE WORK

We have presented here a method of computing

sparse representations of facial images that preserve the information required to estimate

expression, occlusion, pose and illumination

with SRC and NSPM. These both reduce the computation required to compute SRC, NSPM

and improves the accuracy of the results.

An intriguing question for future work is

whether this framework can be useful for object

detection, in addition to recognition. The

usefulness of sparsity in detection has been

noticed in the work in [12]. We believe that the

full potential of sparsity in robust object

detection and recognition together is yet to be

uncovered. From a practical standpoint, it would

also be useful to extend the algorithm to less

constrained conditions, especially variations in

object pose. Robustness to occlusion allows the

algorithm to tolerate small pose variation or

misalignment.

REFERENCES

[1] John Wright, Allen Y Yang, Arvind Ganesh, and S Shankar Sastry,”Robust Face

Recognition Via Sparse Representation “, IEEE

Transaction on Pattern Analysis And Machine

Intelligence, February 2009.

[2] K. Huang and S. Aviyente, “Sparse

Representation for Signal Classification,” Neural Information Processing Systems, 2006

[3] D.Donoho, “For Most Large

Underdetermined Systems of Linear Equations

the Minimal l1-Norm Solution Is Also the

Sparsest Solution,” Comm. Pure and Applied

Math., vol. 59, no. 6, pp. 797 829, 2006

[4] P. Sinha, B. Balas, Y. Ostrovsky, and R. Russell, “Face Recognition by Humans:

Nineteen Results All Computer Vision

Researchers Should Know about,” Proc. IEEE, vol. 94, no. 11, pp. 1948-1962,2006.

[5] W. Zhao, R. Chellappa, J. Phillips, and A.

Rosenfeld, “Face Recognition: A Literature

Survey,” ACM Computing Surveys, pp. 399-458, 2003.

[6] M. Turk and A. Pentland, “Eigenfaces for

Recognition,” Proc. IEEE Int’l Conf. Computer Vision and Pattern Recognition, 1991.

[7] X. He, S. Yan, Y. Hu, P. Niyogi, and H.

Zhang, “Face Recognition Using Laplacianfaces,” IEEE Trans. Pattern Analysis

and Machine Intelli, Mar.2005

[8] D. Lowe. Distinctive image features from

scale-invariant keypoints. IJCV, 2004. [9] Nearest-Subspace Patch Matching for Face

Recognition Under Varying Pose and

Illumination, Zihan Zhou, Arvind Ganesh and Yi Ma, 2007

[10] D. Keysers, T. Deselaers, C. Gollan, and H.

Ney. Deformation models for image recog.

PAMI, 2007. [11] J. Luo, Y. Ma, E. Takikawa, S. Lao, M.

Kawade, and B. Lu. Person-specific SIFT

features for face recognition. In proc. ICASSP, volume 2, 2007.

[12] D. Geiger, T. Liu, and M. Donahue,

“Sparse Representations for Image Decompositions,” Int’l J. Computer Vision, vol.

33, no. 2, 1999.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

24

ENHANCED ENSEMBLE KALMAN FILTER FOR TOMOGRAPHIC IMAGING

D.Dhanya Second year M.E( CSE), Mr.S.Selva Dhayanithy,Lecturer(CSE).

Jerusalem College of Engineering Chennaii.

Abstract-This addresses the problem of

dynamic tomography by formulating it

under the linear state-space model. The

main theoretical contribution is a new

theorem that addresses the convergence

of the EnKF.The main idea behind the

paper“ENHANCED ENSEMBLE KALMAN

FILTER FOR TOMOGRAPHIC IMAGING” is to

formulate the state estimation problem of

a dynamic image and thereby

reconstructing the original image. This can

be done by a filter known as Ensemble

Kalman Filter, a Monte Carlo algorithm

that is computationally tractable when the

state dimension is large. Since it is based

on mathematical procedure it is called

tomographic reconstruction. The Kalman

Filter, a time dependent construction

method is used in this application. In

medical tomography state may change by

a small amount during the measurement

interval and is considered to be static,

hence estimation becomes cumbersome

process. In Existing a high resolution with a

spatial resolution of 128x128 pixels are

used to get the clarity of the image. The

problem is solved by Ensemble Kalman

Filter which is used to get the result of

unfeasible or impossible task and is

tractable when the state dimension is

large. Frequency domain is used in this

application’s. A high resolution with a

spatial resolution of 256x256 pixels is used

in this application. The results of the

algorithm will be presented and

discussed.

Index Terms-Imageprocessing,Kalman

filtering,stateestimation,Ensemble

Kalman Filter,tomography.

I. Introduction

The Ensemble Kalman Filter has been

examined and applied in a number of

studies since it was first introduced by

Evensen(1994b)[3,5].It has gained

popularity Because of its simple conceptual

formulation and relative ease of

implementation, e.g., it requires no

derivation of a tangent linear operator or

adjoint equations, and no integrations

backward in time.The problem of

estimating the properties of a hidden

Markov random process is encountered in

many applications, including radar tracking

of multiple targets time-dependent

tomography and interferometric imaging,

geophysical data assimilation and economic

forecasting[2].Tomography is one of the

most important processes which involve

25

gathering projections of data from multiple

directions and feeding the data into a

tomographic reconstruction software

algorithm processed by a computer[8,6]. It

is a method of producing a image of the

internal structures of a solid object (as the

human body or the earth) by the

observation and recording the differences

in the effects on the passage of waves of

energy impinging on those structures

.Different types of signal acquisition can be

used in similar calculation algorithms in

order to create a tomographic

image[1].There are various types of

tomography such as

Neutron tomography, Optical

projection tomography, Optical coherence

tomography, Ultrasound transmission

tomography, X-ray tomography[6]etc. The

Dynamic tomography problem can be

solved by modeling the unknown object as

Markov random process and then

estimating its properties. The filters used in

this dynamic tomography is Kalman

Filter(KF) and Ensemble Kalman

Filter(EnKF).The Kalman Filter is a set of

mathematical equations that provides an

efficient computational recursive means to

estimate the state of a process, in a way

that minimizes the mean of the squared

error. The Kalman Filter operates on Hidden

Markov model(HMM).If the dynamic

tomography problem can be posed under

the linear state-space model,then the

kalman filter may be used to recursively

compute linear MMSE state estimates.

2 Methodology and System Architecture

This section represents the design

requirements and system architecture of

Ensemble Kalman Filter. The design phase

mainly depends on the detailed

specification in the feasibility study. The

design phase acts as a bridge between the

required specification and implementation

phase. It is the process or art of defining the

architecture, components, modules,

interfaces, and data for a system to satisfy

specified requirements.

2.1 System Architecture

This section presents an overview

of the system architecture. System

architecture is the conceptual design that

defines the structure or behavior of a

system. An architecture description is a

formal description of a system, organized in

a way that supports reasoning about the

structural properties of the system.

In this paper Ensemble Kalman

Filter is used for the reconstruction of

Dynamic images. It is a Monte-Carlo

algorithm which makes computation ease

when the state dimension of the image is

large. This framework explains the overall

process of the Ensemble Kalman Filter and

how the images are converged and how it is

been reconstructed.

The blurred images can be

reconstructed using Ensemble Kalman filter,

the images are been converged and these

images are given priority based on image

resolution, using Linear Minimum Mean

square error technique the noises and

errors are removed. The error freed images

are been sharpened using a tool known as

covariance tapering. The boundaries of the

26

images are identified using covariance

matrix, once these boundaries are

identified they are sharpened. Covariance

tapering improves sample size covariance

and introduces a systematic bias into the

estimator and reduces its variance.

The covariance matrix is otherwise

known as dispersion matrix which is a

matrix of covariances between elements of

a random vector. It is a generalization to

the higher dimensions of variance.

Frequency domain is mainly concerned. It is

a common image and signal processing

technique. The reason for doing filtering in

frequency domain is generally it is

computationally faster to perform and it

can smooth, sharpen, and restore the

images.

The Kalman Filter are used to

compute Minimum Mean Square Error and

is computationally intractable for large

dimension problems. The Ensemble Kalman

Filter is demonstrated to give dynamic

tomographic reconstruction of large

dimension image. A high resolution with a

spatial resolution of 256 * 256 pixels can be

used to get the high quality of large

dimension images.

III. Theoretical Development

A. Treat the problem of Dynamic

Tomography

This module covers the problem of dynamic

tomographic by developing it under linear

state space model. Under this model, the

unknown is the Markov random process

where the unknown state is a random

vector with real components and the time

index is a positive integer. Lowercase

boldfaced symbols denote vectors and

ordinary type symbols denote scalars. The

observable measurements are denoted by

the random process, each a random vector

with M real components. The state and

measurement processes are fully described

by the probability density functions (PDFs).

B. Designing the Kalman Filter

The Kalman filter is an efficient

recursive filter that estimates the state of a

linear dynamic system from a series of

Blurred

images

Convergence of

ensemble

kalman

filter

Prior estimation

Linear minimum mean

square error-LMMSE

Covariance tapering-

sharpening the image

Performance measures

using

EnKF

Reconstructed image

-

256X256 pixels

27

noisy measurements. It is used in a wide

range of engineering applications from a

radar to computer vision. The Kalman filter

exploits the trusted model of the dynamics

of the target, which describes the kind of

the movement possibly by the target to

remove the effects of the noise and get a

good estimate of the location of the target

at the present time (filtering),at a future

time(prediction),or at a time in the

past(interpolation or smoothing).

The Kalman filter is a minimum

mean square error estimator. The Kalman

filter in Dynamic Tomography works under

Hidden Markov Model, under this model

the unknown is the markov random

process. Observable measurements are

denoted by random process each random

vector with M real components. Using

Probability density function state and

measurements, processes are fully

described. Under hidden markov model,

dynamic tomography problems are solved

using minimum mean square error

standard. High dimensional state estimation

is computationally intractable under the

general HMM.Kalman Filter starts with

initial prior estimate and estimate error

covariance.

High-dimensionalstate estimation is

computationally intractable under the

general HMM and the remainder of the

paper focuses on the linear state-space

model. Linear state space model is matched

for many applications in remote sensing

and other fields. Kalman filter is recursive

algorithm for linear minimum mean square

error (LMMSE) calculates of state under

linear space model. The Kalman gain is a

function of the prior estimate error

covariance, a symmetric positive definite

matrix with unique components. In this

module we find out kalman filter.

C. Determine the Ensemble Kalman Filter

The EnKF is a Monte Carlo

approximation to the KF developed to find

approximate LMMSE estimates when the

state dimension N is large enough that the

storage of the KF estimate error covariance

becomes computationally intractable.

Initial state x1 (left) and the initial prior estimate µ1 (right)

given by the backprojected estimate at low resolution. The

error between x1

and µ1 is 0.49.

Log-lin comparison of the error for the three

filters

28

Initial state x1 (left) and the initial prior estimate µ1 (right)

given by the backprojected estimate at highresolution. The

error between x1

and µ1 is 0.49

The end result of EnKF is set of

approximate linear minimum mean square

error estimate for each time index i.The

advantage of EnKF results from the use of

sample error covariance where as Kalman

Filter requires the storage of N*N matrix to

calculate Kalman gain.

Theorem 1: The EnKF estimates xbi|i

converge in probability to the LKF estimates

xbi|i

1, i.e.,

in the limit as the ensemble size L→∞.

The LKF is initialized with

and

The LKF measurement update is given by

The LKF time update is given by

The general idea is to efficiently update an

ensemble of samples such that the

ensemble sample mean approximates the

LMMSE state estimate. The end result of

the EnKF is the set of approximate LMMSE

estimates for each time index i. The

computational advantage of the EnKF

results from the use of the sample error

covariance.To minimize storage and

processing requirements,a small ensemble

size L is desirable.

When no covariance taper is

applied,the LKF and KF are equivalent.The

EnKF estimates converge in probability to

the estimates given by the localized kalman

filter.LKF is computationally expensive as

the KF and is not intended for use in large

scale state estimation.The EnKF with the

taper matrix Ci,may process the

measurements sequentially and inherits the

same benefits as the KF with sequential

processing.

D. Test Cases

29

Test cases are used to develop the

test data, both input and the expected

output, based on the domain of the data

and the expected behaviors that must be

tested. Test data are the inputs which have

been devised to test the system.

3.4.1 Kalman Filter

1. Ensure that the two blurred

images of dynamic objects are

converged.

2. Frequency is applied for each

image.

Expected Result

Two Images are made to meet at

one point.

Frequency is filtered.

3.4.2 Ensemble Kalman Filter

Convergence

1. Ensure the two images of

dynamic object are converged.

Expected result

Two images are made to

meet at one point.

2. Ensure that better viewed images

are given priority respectively.

Expected result

Images are given priority.

3. Ensure that the noise is removed

form the image which is given fist

priority.

Expected result

Noise is removed and the

image is clear.

IV. Summary and Conclusion

This paper addresses the problem

of dynamic tomography by formulating it

under linear state space model.The Kalman

Filter, is implemented in frequency domain

rather than time domain. It is a set of

mathematical equations that provides an

efficient computational means to estimate

the state of a process. The results

demonstrate that the Kalman Filter is

computationally intractable for the large

dimension problems.The Ensemble Kalman

Filter was demonstrated to give dynamic

tomographic reconstructions of almost

equal quality as the optimal KF at a fraction

of computational expense. In Further

computational developments, Ensemble

Kalman Filter can be applied to full remote

sensing problems where the state may have

in excess of one million components.

30

REFERENCES

[1] R. A. Frazin,1 M. D. Butala,1 A. Kemball,2 and F.

Kamalabadi “Time-Dependent Reconstruction Of

Nonstationary Objects with Tomographic or Interferometric

Measurements” The Astrophysical Journal,Vol 635 pp.L197-

L200,2005.

[2] Geir Evensen ,“Sequential Data Assimilation with a

Nonlinear Quasi-Geostrophic model using Monte Carlo

methods to forecast error statistics” Nansen ,Environmental

and Remote Sensing Center, Bergen, Norway, “Vol .3,P.e

1994.

[3] P. L. Houtekamer, Herschel L. Mitchell, Gérard Pellerin,

Mark Buehner, Martin Charron,Lubos Spacek, and Bjarne

Hansen “Atmospheric Data Assimilation with an Ensemble

Kalman Filter: Results with Real Observations”Vol 133,2004 .

[4] Hussein El Dib, Andrew Tizzard and Richard Bayford,

Middlesex University, Department of Natural Sciences,

Hendon, London, UK” Dynamic Electrical Impedance

Tomography for Neonate Lung Function Monitoring using

Linear Kalman Filter”.

[5] L.Scherliess, R. W. Schunk, J. J. Sojka,andD. C.

Thompson“Development of a Physics-Based Kalman Filter

for the Ionosphere in GAIM” Schunk et al., 2002.

[6]Bonnet,A.Koenig,S.Roux,P.Hugonnard,R.Guillemaud,and

P.Grangeat,”Dynamic X-ray Computed

Tomography”Proc.IEEE,Vol 91,pp.1574-1587,2003.

[7] N. Gordon, D. Salmond, and C. Ewing, “Bayesian state

estimation for

tracking and guidance using the bootstrap filter,” J. Guid.

Control Dyn.,

vol. 18, pp. 1434–1443, 1995.

[8] J. Geweke, “Bayesian inference in econometric models

using Monte

Carlo integration,” Econometrica, vol. 57, pp. 1317–1339,

1989

[9] N. P. Willis and Y. Bresler, “Optimal scan for time

varying tomographic

imaging I: Theoretical analysis and fundamental limitations,”

IEEE Trans. Image Process., vol. 4, pp. 642–653, 1995.

[10] N. P. Willis and Y. Bresler, “Optimal scan for time

varying tomographic imaging II: Efficient design and

experimental validation,” IEEE

Trans. Image Process., vol. 4, pp. 654–666, 1995.

[11] N. Aggarwal and Y. Bresler, “Spatio-temporal modeling

and adaptive

acquisition for free-breathing cardiac magnetic resonance

imaging,”

presented at the 2nd SIAM Conf. Imag. Sci.,

2004.

[12] M. Vahkonen, P. A. Karjalainen, and J. P. Kaipio, “A

Kalman filterapproach to track fast impedance changes in

electrical impedance tomography,”

IEEE Trans. Biomed. Eng., vol. 45, pp. 486–493, 1998.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

32

Designing of Meta-Caching Based Adaptive

Transcoding for Multimedia Streaming

Mr.R.Siva Koumar, Mr.N.Durai Murugan, Lecturer Investigator, M.Tech Student,

Department of Information Technology,

Madras Institute of Technology, Anna University, Chennai 44.

Abstract - Internet access rates are heterogeneous and

many Multimedia content providers today are made up of

different versions of the videos, transcoding is a technique

employed by media proxies which is dynamically

personalized the multimedia objects for individual client

characteristics and respective device profile. The proposed

Meta-caching Based Adaptive Transcoding will always cache

certain selectively adaptive transcoded versions and identifies

some versions from which certain intermediate objects would

cached so that it can easily maximizes the CPU utilization

and minimizes the storage memory at media proxy. Both the

network and the media proxy CPU are potentially narrow

congested for streaming media delivery. The simulated

results show that the proposed system can rapidly improves

the system throughput over previous results.

Keywords – Streaming Media, CPU Intensive

Computing, Content Adaptation, Caching, Transcoding

1. INTRODUCTION

In expansion of multimedia applications and present

communication infrastructure comprised of different

underlying network bandwidth and there has been a

growing need for inter-network multimedia

communications over heterogeneous networks. With

aggregate increase of internet bandwidth and the rapid

development of wireless networks efficient delivery of

multimedia objects to all types of wireless devices, internet

accesses from portable devices, such as PDAs and cell

phones, are also growing rapidly. For example, in

transmitting a video bit stream over a heterogeneous

network, the connection from the video source to the end

user may established through links of different

characteristics and capacities. In this case, the bandwidth

required by the compressed video was usually adjusted by

the source in order to match the available capacity of the

most structural network link are used in the connection.

Media transcoder converts previously compressed

video signal into another one with different format, such as

different bit rate, frame rate, frame size, or even

compression standard. If real-time video is used, the

encoder will adjust its coding parameters. The PDAs and

mobile phones are having different screen sizes than the

personal computer and these devices requires significant

bandwidth as compared to the desktop computers, the

multimedia file formats which are supported by personal

computers (such as, movies, audio/video files, images,

texts, etc..) cannot be delivered directly on a mobile phone.

It must personalize appropriately before it is delivering to

the client or it has to done during the runtime. This can

supports the Quality of Services and is often refer to as

content adaptation. Two approaches are used for providing

this type of QoS support in the context of multimedia

content delivery.

I. Pre-encoding provides the multiple provisioned

translations and physically encodes the different media

objects. All the media object details are worked out before

they delivered to the client. For example, see [1], many

content hosts encodes their video clips at different bit rate

versions, 28/56 Kbps for dial-up clients and 100-plus Kbps

for broadband clients. If considering all possible

requirements of client devices (not limited to various

network speeds), pre-encoding demands a huge amount of

storage for different versions. Scalably coded content

requires less space than multiple individual versions, but it

is still not efficient in compression of this media content. In

addition, content created by this way can only satisfy

certain rough granular QoS requests. It is less adaptable

when QoS is required. More importantly, pre-encoding

does not scale to the vast variety of media adaptation

applications. It may be easy to show the possible bit rate

versions that are required for a streaming application, but it

would be difficult to pre-encoded content for more generic

adaptation tasks such as personalization. When pre-

33

encoded video need to distribute to users with different

connections, the implementation is difficult because the

target transmission channel conditions are generally

unknown when the video is originally encoded. Besides the

problem of characteristics and capacities, different client

devices used today’s communication also make some

problems.

II. Transcoding - However, gives us a second chance for

dynamically adjust the video format according to channel

bandwidth and users devices. Transcoding enables re-

compression of multimedia content, i.e. reduction in the

content's byte size, so that the delay for downloading

multimedia content can be reduced, here transcoding

provides the real-time content adaptation support. Media

transcoding is not mostly restricted for QoS support while

doing the personalization of media objects. This is

particularly useful when there are time variations in the

channel characteristics; it has more flexibility and scales

well with the variety of the applications. However,

transcoding is often computing intensive, especially for

multimedia content. Research on developing efficient real-

time transcoding algorithms has received much attention,

particularly on video transcoding.

The common motivation for applying adaptive

transcoding is to reduce the delay experienced when

downloading multimedia content over Internet access links

of limited bandwidth, the transcoding function has

typically placed within an HTTP proxy (media proxy

server) that resides between the content provider’s Web

server and the client Web browser. The transcoding proxy

reduces the size in bytes of the multimedia content via loss

compression techniques (e.g. video, audio and images are

more aggressively compressed). The media proxy then

sends the re-compressed multimedia content over the

modem/wireless access link down to the client. The

reduction in byte size over the access link, typically a

bottleneck, enables an often-significant reduction in the

perceived response time. Transcoding has also been

applied to continuous media to reduce a video’s bit rate and

thereby enable clients connected over bandwidth-

constrained links to receive the modified video streaming

2 ADAPTIVE TRANSCODING

Fig. 1 illustrates the processing flow of the adaptive

transcoding, media server gets request from the client and it

checks for the availability of the object, if the object is

available in the server then the object is serve to the client

and that object is cached in the media proxy. If the requested

object is not available in media server then it will checks in

media proxy, each requested object by client is cached in

media proxy, so that for frequently requesting objects can

accessed easily by the client. Those objects which are caching

from the media server are transcoded (full transcoder) and

this transcoded objects is also cached (regular-full or no

cache). These objects are even transcoded adaptively

(Adaptive Transcoder), when this media content is cached

(meta-cache). Since the media content are cached the

transcoding process can be reduced in future and the media

objects would deliver to client easily.

Fig1. Block Diagram for Adaptive Transcoding.

2.1. META-CACHING SCHEME

Caching is a feasible technique for achieve computing load

reduction. The results of transcoded versions for a client

request can be cached so that in future if same media content

is requested then it can be served without transcoding. The

design of adaptive transcoding for meta-caching scheme has

systematically examined. But now the focus is on efficient

utilization of different resources (Processing, Memory and

Bandwidth) for rapid development of throughput of the

transcoding in media proxy. The cached data may be the

input or output of the transcoder.

Specifically, if a transcoded version is fully cached full-

caching scheme, identical future requests can be directly

served without additional transcoding. However, to cache

each transcoded version may quickly exhaust the cache

space, if a transcoded version is not cached no-caching

scheme, identical requests will result in repetitive

transcoding, consuming extensive CPU cycles. If nothing is

cached, e.g., there is no cache at the transcoding proxy;

we call it the “no-caching” method. In this case, the

service proxy always starts a new session upon request

34

arrival and the per-session computing load is 100%. In

another extreme, if the system caches the final results,

we call it the “full-caching” method. In this case, no

computing is required for a future identical session

assuming unlimited cache space.

This paper revokes the meta-caching scheme; here

transcoding movements are deliberated and identified so that

appropriate intermediate results (called metadata) can be

cached. With the cached metadata, the fully transcoded object

can be easily produced with a small amount of CPU cycles.

The resource balancer makes the decision based on the cache

space and computing cycles available, since metadata only

cached, the required cache space is considerably reduced.

The saved cache space can be used to store metadata for

other transcoding sessions so that, the overall computing load

can be reduced. Thus the meta-caching scheme allows the

system to achieve a joint control of the CPU and storage

resources. It contributes a balanced position between the full-

caching and no-caching schemes. The contents are

transcoded according to client requests before transmitted

through the Internet. This setup is often found in many

server-side proxies. The cache space on the proxy is

partitioned into two sections. Considering a transcoding

proxy that performs transcoding on compressed MPEG

video objects, the transcoding flow may produce

metadata, Regular cache is used to store the final results of

transcoding sessions. Meta cache is used to store the

metadata which is defined as the intermediate result produced

during a transcoding session.

3. DESIGN ARCHITECTURE

The modules and features of a media transcoding system

are illustrated in Fig 2. In this system, pre-encoded, high

quality, high bit-rate videos are stored at the media source.

The media transcoding proxy has two components, a

transcoder and a control module. The transcoder is an actual

conversion engine of streaming media content. On the other

hand, different user clients maintain the client’s profile.

The media transcoding proxy can be implemented on the

media server where the pre-encoded media content are

stored, or on an intermediate node along the transmission

path. The basic function of the media transcoding proxy

includes, Frame Size: Change the frame size according to

the client profile. Such as most handheld devices can only

display small size video (176x144 pixel) When the pre-

encoded video is in big video size, it needs to be

transcoded to small frame size. Frame Rate: Change the

video frame rate. Since some handheld devices can only

play video at a low frame rate, such as 5, or 10 fps (frames

per second), the high quality video (30 fps) needs to be

transcoded to low frame rate one. Frame rate conversion

can also reduce the bit rate. Bit Rate Adaptation: Convert

the video bit rate according to the channel conditions.

Since the pre-encoded video is encoded at high quality and

bit rate. For low bandwidth connections, the video bit rate

needs to be converted to low bit rate. Frame format: The

control method determines frame format by estimating

both the display size and processing power of the client

device. First, the control module derives the frame format

for a video stream to be fit to the client display size, the

control module derives the largest frame format for which

both width and height are within the client display size, the

control module calculates the maximum frame rate at

which the client device can reproduce a video stream in

real time on the frame format. Under the condition of

constant encoding bit rate and the same frame format, the

more the quality value increases, the more the frame rate

decreases, that is to say there is a trade-off relationship

between quality and frame rate.

Basically transcoder will decodes media content and

then it will convert the size of media content and re-

encoding. Change the compression standards, such as from

MPEG1 to H.263, etc the transcoding process can be

synchronized with time, and it is possible to execute media

transcoding in real time. While consider the size

conversion it is need to specify the frame format. The

frame format indicates both width and height of the media

content, the encoding method, quality value, and encoding

bit rate must be specified. The encoding method is a

compression method such as MPEG-4. The quality value is

used for quantization of the video frame, the quality value

is dependent on applications that are a greater quality value

indicates high quality and the lesser the quality value, the

lower the quality. The encoding bit rate is the data rate at

which the transcoder looks for and the frame rate is

adjusted in the transcoder automatically so that the data

rate may become the specified rate. Then the control

module, based on profiles sent from the client proxy, the

client proxy relays the data between the media proxy and

the media client. It keeps data size relayed to the video

client; by adding up the data size in a specified interval, the

download throughput can be measured.

The client proxy notifies the media proxy of its result.

Further-more, the client proxy collects the responsibility of

transmission profile for monitoring the dynamic condition

35

of the transmission channel, such as effective channel

bandwidth, channel error rate, etc. Device profile describes

the capability of the device, such as display size,

processing power and decoder information, etc.

36

Fig2. Design Architecture of Transcoding Media Server .

User profile describes the user preference. Which enables

the mobile user to control the transcoding process, and

notifies the video? In implementation, the client proxy

may be integrated with the video client or separated from

it; say, the video client is located on a note-book PC or

PDA, and the client proxy on a mobile phone.

4. PERFOMANCE AND SIMULATIONS

On analyzing meta-cache, the simulated result shows

that the three type of caching schemes is achieved model-

driven analytical result, the proposed system is

reproducing a maximize throughput, thus adaptively use

different schemes upon dynamic client access patterns

and resource availability, on considering the performance

of the Adaptive transcoding, let we move on to the

calculation that are discussed in [1]. Based on the

available storage space the number transcoding sessions

are taken; the average storage requirement for full cache

will be Sf, here we consider K is most popular media

object requested by client v is transcoding versions N is

total no of access to object in transcoding proxy server.

- Eq. 4.1

For meta-caching Sm we consider -storage

requirement for meta-cache which relate to full cache,

- Eq. 4.2

For no caching Sn, it’s no need for the requirement of

transcoding versions v,

- Eq. 4.3

After considering the storage space, now we consider

the CPU load for full, meta and no caching. The

available CPU capacity for full caching Zfc, here we

consider C as total computing cycle

- Eq. 4.4

Where A= , M is total no of media objects in

media proxy server and is a skew factor and is

always set to be between 0.47 and 0.73. For meta-cache

the CPU capacity is Zfm

- Eq. 4.5

For no-caching Znc,

- Eq. 4.6

With reference to the parameter defined above the

Simulated Throughput can be conciseness on basis of

α and β values it may consider that α as 0.5 and β as

0.3. Fig 3 is simulated for 10000 objects; the

throughput is measured on an application layer at user

clients so that accurate throughput can be measured.

Fig 4. Consider the workload of 20000 objects. Each

one would contain around 1000 requests from different

clients. The framework made here would calculate the

ratio of the active storage size over the total Storage

37

size.

0 200 400 600 800 1000 1200 1400 1600 1800 2000200

300

400

500

600

700

800

Available CPU load

Thro

ughput

Full Cache

Meta Cache

No Cache

Fig3. Comparisons of CPU Load and Throughput for 2000 access

0 1000 2000 3000 4000 5000 6000 7000 8000 9000 100001000

2000

3000

4000

5000

6000

7000

8000

Available CPU load

Thro

ughput

Full Cache

Meta Cache

No Cache

Fig4. Comparisons of CPU Load and Throughput for 10000 accesses

The active storage size is the sum of the size of all active (requested) cached objects at present. The total

storage size is the total available cache size. As shown

in the figure, Adaptive Transcoding for meta-cache is

moreover performs well compared to other schemes. Meta-caching and no-caching are more dependent on the

available CPU resources, while full-caching demands

more cache space. Adaptive transcoding is least affected,

particularly when the cache space increases over 20%. It is demonstrate on various conditions, such as

available resources and client access patterns.

5. CONCLUSION

In this paper, we have investigated the performance

of the throughput on comparing with the available CPU

load; the media transcoding proxy server has gradually

minimize the storage space and the network traffic and

transcoding procedure had provide the reduction on the

media server load, the design of meta-caching scheme

is concentrates on reproducing the high throughput and

refers to computing and storage space of media server.

The proposed meta-caching based adaptive transcoding

access the analysis of the meta-caching scheme, which

would adaptively transcode the media objects based on

client accesses and available storage resources in the

system. Thus the simulated results show that the

proposed system can rapidly improves the system

throughput as it compared to the previously occurred

results.

ACKNOWLEDGMENT

I wish to express my sincere thanks to the Department of Information Technology for providing

the required resources for doing the research in the

field of multimedia streaming and communications.

REFFERENCE

[1] Dongyu Liu, Songqing Chen, Member, IEEE, and Bo Shen,

Senior Member, IEEE, “Modeling and Optimization of Meta-

Caching Assisted Transcoding”, IEEE Transactions on

Multimedia, vol. 10, no. 8, December 2008. [2] Bo Shen and Susie Wee, “Image/Video Transcoding with Hpl

Technology “, Mobile and Media Systems Lab Hewlett-Packard Laboratories, HPL-2007-145 August 27, 2007.

[3] B. Shen, “Meta-caching and meta-transcoding for server side

service proxy,” in Proc. IEEE Int. Conf. on Multimedia and

Expo (ICME03), Baltimore, MD, Jul. 2003, vol. I. [4] C.W. Lin, J. Xin, and M.-T. Sun, “Digital video transcoding,”

Proc. IEEE, vol. 93, no. 1, pp. 84–97, Jan. 2005.

[5] B. Shen, S. Lee, and S. Basu, “Caching strategies in

transcoding enabled proxy systems for streaming media

distribution networks,” IEEE Trans. Multimedia, vol. 6, no. 2,

pp. 375–386, Apr. 2004.

[6] Keqiu Li, Keishi Tajima, and Hong Shen, “Cache Replacement

for Transcoding Proxy Caching”, in Proceedings of the 2005

IEEE/WIC/ACM International Conference on Web

Intelligence. [7] R. Mohan, J. R. Smith, and C. S. Li, “Adapting multimedia

internet content for universal access,” IEEE Trans. Multimedia, vol. 1, no. 1, Mar. 1999.

[8] F. Hartanto, J. Kangasharju, M. Reisslein, and K. W. Ross,

“Caching video objects: layers vs versions?” in IEEE Int. Conf.

on Multimedia and Expo, Lausanne, Switzerland, Aug. 2002.

[9] A.Vetro, C. Christopoulos, and H. Sun, “Video transcoding

architectures and techniques: An overview,” IEEE Signal

Processing Mag., vol. 20, pp. 18–29, Mar. 2003. [10] E. Amir, S. McCanne, and H. Zhang, “An application level

video gateway,” in Proc. ACM Multimedia, San Francisco, CA, Nov. 1995.

[11] M. Chesire, A. Wolman, G. Voelker, and H. Levy, “Measurement and analysis of a streaming media workload,” in Proc. 3rd USENIX Symposium on Internet Technologies and Systems, San Francisco, CA, Mar. 2001.

[12] R. Rejaie and J. Kangasharju, “Mocha: A quality adaptive multimedia proxy cache for internet streaming,” in Proc. ACM Workshop on Network and Operating System Support for Digital Audio and Video (NOSSDAV), Port Jefferson, NY, Jun. 2001.

[13] W. Li. “Overview of Fine Granularity Scalability in MPEG-4 Video Standard”. IEEE Trans. on Circuits and Systems for Video Technology, 11(3):301317, March 2001.

[14] B. Shen, and S. Roy, "A Very Fast Video Special Resolution

Reduction Transcoder,” Proceedings of ICASSP2002. vol. 11,

Orlando, FL, May 2002.

38

[15] S. Ota et al., “Architecture of Multimedia Data Transcoding

System for Mobile Computing,” 2000 IEICE General Conf., B-

5-174, Mar. 2000, (in Japanese), pp. 559.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

39

DIGITAL IMAGE PROCESSING (DIP)

JPEG COMPRESSION WITH CLIENT BROWSER PLUG-IN APPROACH FOR

EFFICIENT NAVIGATION OVER INTERNET

K.Kavin Prabhu, Preethy Balakrishnan

Students, II-M.Sc Software Engineering

P.G. Department of Software Engineering

M.Kumarasamy College of Engineering

Thalavapalayam,Karur-639113,Tamilnadu,India

ABSTRACT

Image compression and

transmission has been steadily gaining in

importance with the increasing

prevalence of visual digital media. In this

work we present an application of the

latest image compression standard JPEG

in managing and browsing image

databases. Focusing on the image

transmission aspect, we combine the

technologies of JPEG image compression

with client-server socket connections and

client browser plug-in, as to provide with

an all-in-one package for remote browsing

of JPEG compressed image databases,

suitable for the effective dissemination of

cultural heritage. The main features of

JPEG, in which our implementation is

based upon, is progressiveness and

embedded bit stream syntax, which

implies the ability to extract from one

compressed file many different

resolutions or quality levels of an image,

depending on the user’s request.Our goal

was to achieve a more efficient way of

navigation. Particularly, in HTTP the

proposed system consists mainly of an

application specifically developed for

browsing JPEG images over the Internet

using JPIK (JPeg Interactive with Kakadu)

protocol rather than the World Wide Web

standard HTTP. On the other hand, in

wide spread browser, the system is based

on the implementation of a helper

application, which is a third-party piece of

software. The helper application has to be

distributed and is executed on the client

side whenever a request from a browser

activates it. In this case, the actual image

data are displayed in a separate window

and they cannot be manipulated from

within the browser. To overcome these

disadvantages, we decided to develop a

browser plug-in which is an approach

frequently used for displaying content

with in today’s internet browsers, which

increases the transmission rate.

1.INTRODUCTION

One of the major technical

problems that the Internet experiences

today is its data transmission rate

capability. The prohibitive transmission

rates for large images combined with

the need for enormous storage spaces

lead to the development of new coding

schemes to compress image

information. An example of such a

compression scheme with wide

40

acceptance on the Internet is JPEG . By

using JPEG on an image 10 * 15 cm at

600 dpi, we reduce the transmission

time through the modem line from 1 h

to about 30 s, while preserving a

moderate image quality (about 32 dB

PSNR).JPEG employs wavelet transform

to achieve energy compaction into a

small number of transform coefficients

and a highly sophisticated embedded

coding, combined with a binary

arithmetic encoder and rate control

mechanisms, packing data into a

meaningful, flexible and easily

parseable code stream. This new

standard shows great potential in aiding

the management and transmission of

images over the Internet and thus is one

of the most useful tools to adapt to a

cultural heritage database.

JPEG compression example:

JPEG compression example, (a) original image, 24 bpp, and transmission time about 1 h

and (b) compressed at 0.19 bpp, quality 32 dB PSNR and transmission time about 30 s.

The transmission of data is

done in a differential way, so that the

user accepts only the information

needed to move on to a next step in the

scenario. The interaction between the

user and the database server is

controlled by a socket connection, while

41

JPEG decoding and display is

accomplished by a client browser plug-

in. This means that the original image

code stream has to be structured in

such away as to include many different

resolutions and quality levels of itself.

At each request for one of these

resolutions or quality levels, the server

is responsible for extracting the

appropriate data from the original file

and for transmitting it to the client

where the final merging with the data

already existing from previous requests

would be accomplished.

2.JPEG CODING STEPS:

The key part for achieving

progressive image transmission in our

work is the use of JPEG, which provides

with a wide set of tools for image

compression. The main features of

JPEG, in which our implementation is

based upon, is progressiveness and

embedded bit stream syntax, which

implies the ability to extract from one

compressed file many different

resolutions or quality levels of an image,

depending on the user’s request. The

basic building blocks of the encoder are

shown.

3.SYNTAX OF JPEG CODESTREAM

The code stream syntax specifies such

fundamental quantities as image size, tile

size, number of components, and their

associated sub-sampling factors. It also

specifies all parameters related to

quantization and coding .most coding

parameters can be chosen on a tile by tile

basis. In the simplest case, a JPEG code

stream is structured as a main header

followed by a sequence of tile-streams.

The extension of such a JPEG file, as

proposed by the JPEG standard is noted

as “jpc” (JPEG Code stream).The main

header contains global information

Forward

Transform

Quantization Entrophy

Encoding

Compressed image data

Inverse

Transform

De-

Quantization

Entrophy

Decoding

Compressed image data

42

necessary for the decompression of the

entire code stream. Each tile-stream

consists of a tile header followed by the

compressed pack-stream data for a single

tile. Each tile header contains the

information necessary for decompressing

the pack-stream of its associated tile.

Finally, the pack-stream of a tile consists

of a sequence of packets

JPEG code stream

4 PROGRESSION IN JPEG

Progression enables increasing

the quality, resolution, spatial extent and

color components, as more bytes are

decoded sequentially from the beginning

of a compressed code stream. The type of

progression in a JPEG code stream is

governed by the order in which packets

appear within tile-streams. As such,

progression can be defined independently

on tile-by-tile basis. Actually, tile-streams

can be broken at any packet boundary to

form multiple tile-parts. Each tile-part has

its own header and the tile-parts from

different tiles can be interleaved within

the code stream. In Fig below,

represented graphically, are two of the

progression orders with high importance

for Internet image database applications:

by quality and by resolution.

43

5.CLIENT-SERVER TECHNOLOGY

Client-Server technology is based on

creating sockets to achieve the

communication between two or more

remote computers. A socket works much

like a telephone. It’s the end point of a

two way communication channel .By

connecting two sockets we can pass data

between processes, even processes

running on different computers, just as

we can talk over the telephone. Programs

written to use the TCP protocol are

developed using the client-server model.

When two programs use TCP to exchange

data, one of the programs must assume

the role of the client and the other the

role of the server. The client application

initiates what is called an active open. It

creates a socket and actively attempts to

connect to a server. On the other

hand,the server application creates a

socket and listens for incoming

connections from clients, performing

what is called a passive open . When the

client initiates a connection, the server is

notified that some process is attempting

to connect with it. By accepting the

connection, the server completes what is

called a virtual circuit, a logical

communication pathway between the

two programs

6.IMPLEMENTATION

For the purposes of our application

all of the aforementioned technologies

were combined in order to achieve a

progressive and dynamic way of accessing

the images in a database. Our

implementation includes a server and a

client application

6.1 PREPARATION OF DATABASE

IMAGES: To be able to extract from a

single image code stream multiple

segments that correspond to user

requests, a specific way of encoding

should be followed. Code stream data

must be arranged in resolution level

blocks, so that each sequential resolution

level added to the previous would

produce a greater resolution and

specifically of double width and height,

due to the nature of the wavelet

transform in JPEG. In order to have

images encoded in a way that the first

resolution level has maximum

dimensions, say 64*64 pixels, we have to

estimate the number of decomposition

levels needed for each image in order to

achieve the thumbnail resolution of

44

64*64 pixels, and to compress each image

using lossless JPEG coding.

nw ≤ log2 w/w′, nh ≤ log2 h/h′

n = [maxnw,nh]

where w and h are the width and height

of the original image, w’ and h’ are the

desired width and height, and ‘nw ‘and ‘nh

‘ stand for how many times the image

must be divided by 2 across the width and

height resolutions respectively, to

produce a desired resolution. In our case

the values of w’ and h’ are preset to 64

pixels. Thus, the number of wavelet

decomposition levels ‘n’ must be the

immediate greater integer of the

maximum value between ‘nw ‘and ‘nh ‘

Code stream structure of data base image

This way images will be encoded with

resolution levels (there will be n+1 SOT’

markers in the code stream), ensuring

that the image thumbnail will be of 64*64

pixels maximum resolution. A

representation of the code stream

structure of such an image is illustrated in

Fig above. Hereafter, the database

compressed images are ready to be used

by the server for code stream processing

and progressive transmission to any

remote user through the Internet.

6.2 THE SERVER

The server must be working in

multithreaded mode, so that it can

process many client connections at the

same time, something mandatory for our

application. The server has the ability to

send an image of any resolution by

parsing the image code stream and

selecting the required segments.

Additionally, it can automatically create

the respective HTML files through which

the clients can view the desired images in

45

desired resolutions. Each time a client

requests the thumbnail resolution of an

image, the server sends the respective

code stream segment, including the main

header. When the client has a specific

resolution already available and requests

for a greater one, the server sends the

appropriate differential information:

those tile-parts that correspond to the

requested resolution, not including the

main header and the already sent tile-

parts. The client should be able to merge

them into one single meaningful code

stream. For automation of the

process, the HTML pages in which the

images of each resolution are being

displayed are created on the server side,

always a step ahead of the next greater

resolution.

6.3 THE CLIENT:

The client consists of a browser

plug-in application. The plug-in is

responsible for receiving the coded JPEG

code stream from the server, decoding it,

and then displaying the resulting image

inside user’s browser. After the browser

locates, from the MIME type of the

EMBED tag, the suitable plug-in that must

be executed, it calls it and the whole

control of receiving and sending data is

passed to the plug-in. Particularly, when

the browser starts, all available plug-ins

are loaded into memory. Then the plug-in

follows the below logical flow chart.

46

Logical diagram of plug-in structure

Server’s response to client’s request

47

[Note :All the diagrams in this paper is drawn using AUTOCAD software ]

CONCLUSION:

Usage of JPEG2000 leads to image

databases where images are compressed

and each stored only in a single file. In this

work we developed an infrastructure for

progressive by resolution and quality

transmission of images that were

losslessly encoded and stored in

databases. To this end, the new standard

in image compression, JPEG, has been

employed, which is characterized by its

feature of embedded, progressive and

high-performance coding. A server

application that prepares and handles the

image database, parses JPEG code

streams and handles client requests has

been developed. Additionally, a client

browser plug-in was implemented that is

responsible for handling user requests,

receiving and merging JPEG2000 code

streams, and, finally, decoding and

displaying the database images at four

preset resolutions, optimized for

navigation and network utilization.

REFERENCE:

1. Digital Image Processing –

R.C.Gonzalez & Addison Wesley.

2. Fundamentals of Digital Image

Processing – A.K.Jain.

48

3. www.sigmedia.com

4.D.Artz. Digital Steganography; hiding

Data within Data, IEEE Internet

Computing, Vol.5

5.N.F. Johnson, S. Jajodia, Steganalysis :

The Investigation of Hidden Information

Proc. IEEE Information Technology Conf.

Abstract — In this paper, we propose the Multichannel

blind image deconvolution technique to restore the input image

from blurred and noisy images. The direct image restoration

methods are: Mutually Referenced Equalizer (MRE) and

Regularized Mutually Referenced Equalizer (RMRE). The MRE

method based on inverse filtering works only in noiseless case for

perfect image restoration. In noisy environment regularization

method (R-MRE) is used. The objective function of R-MRE

method is to avoid noise amplification. Another restoration

method is identification of PSF for Multichannel blind

deconvolution method. We determine the degradation filter.

Next, using the degradation filter, original image is restored.

Finally, the performance of the two restoration method can be

analyzed using quality of the image.

I. INTRODUCTION

OWADAYS, Multichannel(MC) image processing is

relatively active field of research. Normally MC method

is used in signal case, now it is extended to the image.

Because of increasing number of application where several

directions of the captured image are available, In Multichannel

framework, in a single scene observed by several images and

it passes through different channels. The multispectral images

are used in frequency band channel. In different time slots, we

captured different images are provided at different resolution.

This is also treated as Multichannel representation. The

advantages of MC processing is to exploit the diversity and

redundancy of information in the different images. Hence, the

set of observed images are considered as one entity based on

it‘s worth.

Image deconvolution/restoration solutions can be divided

into two classes: stochastic and deterministic. Stochastic

methods consider observed images as random fields and

estimate the original image as the most probable realization of

a certain random process. These methods are, mainly, the

linear minimum mean squares error (LMMSE) [13], the

maximum likelihood (ML) [11], and the maximum a

posteriori (MAP) [3]. These methods have two major

drawbacks: i) they are very sensitive to perturbations and

modeling errors, ii) strong statistical hypothesis are made on

the image and the noise which are considered as uncorrelated

homogeneous random processes. On the other hand,

deterministic methods do not rely on such hypothesis and

estimate the original image by minimizing a norm of a certain

residuum. These methods include among others: constrained

least square (CLS) methods which incorporate a regularization

term [12], the iterative blind deconvolution technique [5] and

the non-negativity and support constraint—recursive inverse

filtering (NAS-RIF) algorithm [10]. The latter methods are

based on minimizing a certain criterion under some constraints

like non-negativity of the original image, finite support of the

convolution masks, smoothness of the estimate, etc.

Blind MC image deconvolution can be performed in two

restoration method. The original image is directly using

equalization or inverse filtering technique called Mutually

Referenced Equalizers (MRE). The MRE method is not a

perfect restoration method at the noise environment. Then the

regularization term is incorporated in order to perfect from

noise amplification. In PSF Identification method, first

identify the filter then it will restore the original image using

the identified filter. We compared these techniques in terms of

estimation accuracy (restored image quality). More precisely,

our contributions consist of i) a new blind restoration method

using the mutually referenced equalizers technique in

conjunction with a regularization method that truncates the

greatest singular values of the inverse filter matrix; ii) a new

robust Multichannel restoration method based on a total

variation technique; and iii) a performance evaluation and

comparative study of the different blind restoration methods

using some objective image quality metrics.

II. MOTIVATION FOR MC PROCESSING

The motivations behind the use of MC framework as

compared to the standard mono-channel one are first the

increasing number of applications in this field, but more

importantly, the potential diversity gain that may lead to

significant improvements in the restored image quality. And

another method based on the iterative Richardson–Lucy

scheme. Clearly, this example highlights the performance

gain that can be obtained by MC processing. This gain is due

to the inherent diversity of Multichannel systems where

multiple replica of the same image is observed through

different (independent) channels. In that case, if a part of the

original information is lost (degraded) in one of the observed

images, it can be retrieved from the other observed images

upon certain diversity conditions. More precisely, this is

possible when the same image distortion does not occur on all

the observed images simultaneously. Mathematically, this is

expressed in the condition that the spectral transforms of the

PSF’s do not share common zeros. Indeed, a PSF’s zero

represents roughly a “fading” around a particular frequency

point, and, hence, common zeros represent the situation where

the same fading occurs on all channels simultaneously. In the

presence of noise, perfect reconstruction is not possible but the

Image Restoration Using Regularized

Multichannel Blind Deconvolution Technique

Munia Selvan L, Vinoth Kumar C

Department of ECE, SSN College of Engineering Mail id: [email protected], [email protected]

N

Proceedings of the third National conference on RTICT 2010Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

49

diversity gain provides significant improvement in the restored

image quality. For the MC case, we use the restoration method

with the PSF identification approach (Section VI). One can

observe that the gain increases in terms of image restoration

quality when using the MC processing. Furthermore, this gain

is obtained with a relatively small number of independent

PSFs (3 to 4 depending on the SNR level), which limits the

extra computational burden of MC processing.

III. PROBLEM STATEMENT

A. Notations

It is assumed that a single image passes through K>2

independent channels and, hence, K different noisy blurred

images are observed. Each channel corresponds to a degrading

filter. The PSF functions are assumed to have no common

zeros, i.e., the polynomials are strongly co-prime.

KkzzjiHkzzHj

m

i

n

j

i

k

h h

,,,,1,).(),( 2

1

0

1

0

121 == −−

=

=

−∑∑ ….(1)

Then the system model can be written as

),(),(),(),(1

0

21

1

0

21 nmNlnlmFllHnmG k

m

i

n

j

kk

h h

+−−×= ∑∑−

=

=

….(2)

In the above eqn, F (m, n) is the original image of size (mf,

nf). Hk is the point spread function (PSF) of Kth image and its

size is (mh, nh). The Kth channel degraded images Gk of its size

is (mg, ng). In the sequel, the images and impulse responses

will be processed in a vectorized, windowed form of size (mw,

nw). Hence, we denote Gk (m, n) by the data vector

corresponding to a window of the Kth image where the last

pixel is indexed by (m,n), i.e., the right bottom pixel.

T

wwk

wkkk

nnmmG

nnmGnmGnmg

)]1,1(,,,,,

)1,(,,,),,([),(

+−+−

+−= ….(3)

In order to exploit the diversity and the redundancy offered

by the multiple observations of F, we deal simultaneously with

all observed images by merging them into a single observation

vector

[ ]TT

k

Tgnmgnmg ,,,,,,,),,(),( 1=

),(),( nmnnmHf +=

….(4)

B. Objectives

The ultimate goal is to restore the original image in a

satisfactory way even in “severe” observation conditions. In

practice, the original image as well as the degrading filters is

totally unknown, so that our restoration procedure is totally

blind. To tackle this problem two types of solutions are

proposed. The direct image restoration technique [2] and the

indirect one, that first estimates the unknown PSFs and then

restores the original image in a nonblind way (i.e., using the

previously estimated PSFs) [1]. 1) Direct Restoration:

Our objective here is to directly restore the original image

using only its degraded observed versions. More precisely, we

search for a unique equalizer or inverse filter which, applied to

the set of observations, allows us to restore the original image.

We pay a particular attention to the robustness against additive

noise by including in the estimation criterion of the inverse

filter an additional term that controls and limits the noise

effect.

2) Restoration via PSF Identification:

Our objective here is to, first identify the PSF function and

then inverse it in order to restore the original image. In the

noisy case, the filter response inversion leads to noise

amplification, and, hence, we propose to add a regularization

term in order to reduce this undesired effect.

IV. MUTUALLY REFERENCED EQUALIZERS (MRE)

In this section, we introduce our first image restoration

method using the direct estimation of the inverse filters. These

filters are estimated by MRE method in [9]. The proposed

method builds on the concept of the mutually referenced

equalizers (MRE), which were first introduced in which a set

of K (K>1) filters are considered, the outputs of which act as

training signals for each other. A multidimensional mean-

square error MRE criterion for blind MC equalization is

derived, and certain minimization procedures are discussed.

The obtained algorithm is shown to meet several conditions of

important practical concern:

1) The MRE criterion is unimodal; hence, the algorithm

exhibits global convergence.

2) The MRE criterion is a mean-square error. Full flexibility is

gained for the implementation. In contrast, flexibility of

design is difficult to obtain with subspace decomposition

methods as well as with adaptive BE methods relying on a

high-order nonlinear cost.

3) The method directly provides channel inverses with all

possible delays, hence, yielding robustness with respect to the

noise amplification problems that may be related to a specific

delay (typically the minimum and maximum delays yield poor

results if the channel has coefficients that taper off at the

ends).

4) Finally, the MRE method shows empirical robustness in the

presence of channel length mismatch. We investigate the

theoretical properties of the MRE criterion in the noise-free

case and show that its minimization is necessary and sufficient

for linear MC equalization. We show that the MRE criterion is

a multidimensional MSE that has to be constrained in order to

avoid the convergence toward undesirable solutions.

We search, here, for the restoration filter denoted η which,

applied to the Κ observed images, provides us with an

estimate of the original image. This η is a Multichannel 2-D

filter of size (Kme, Kne) filter exists under the following

assumptions: i) the PSFs have no common zeros [2] and ii) the

filter matrix has full column-rank [9].

)1)(1( −+−+≥Κ ehehee nnmmnm ….(5)

When both conditions are satisfied, there exists a set of

equalizers η1,…ηd each of them allowing us to restore the

original image with a specific spatial shift (md,nd). Note that

there exists an infinity of equalizers of different sizes (an

infinity of couple (me,ne) that satisfy the condition. Hence,

when is fixed, the original image is estimated up to a certain

constant factor and a certain spatial shift. The principle of

50

mutually referenced equalizers is as follows: suppose we have

computed two equalizers ηi and ηj inducing the spatial shift

(mi , mj) and (ni , nj) respectively

),(),,(),(1

, nmnnmmFnmG iik

k

ki ∀−−=∗∑Κ

=

αη ….(6)

),(,),(),(1

, nmnnmmFnmGk

jjkkj ∀−−=∗∑Κ

=

αη ….(7)

where αis a given positive scalar

kiiii ,,1, ],......,[ ηηηη Κ∆ ….(8)

being the 2-D filter of size (me, ne) applied to the Kth observed

image.

In the noiseless case, if we apply the equalizer ηi to G(m-mj ,

n-nj) and the equalizer ηj to G(m-mi , n-ni), we obtain exactly

the same windowed area of the original image.

∑∑Κ

=

Κ

=

−−∗=−−∗1

,

1

, ),(),(k

jjkki

k

iikkj nnmmGnnmmG ηη

….(9)

This solution presents a major drawback: it requires the

computation of a large number of equalizers and, hence, it is

computationally expensive. We can reduce the number of

equalizers to be estimated it is shown in [9] and [7],

consequently, the computational cost of this solution. More

precisely, it was proved that only 2 extremal equalizers taken

from the Kth observed image. The two equalizers are in

boundary of observed image are like (0,0) and (me+mh-1,

ne+nh-1) are sufficient perfect image restoration In practice,

these shifts do not appear to perform good restoration quality

due to the artifacts appearing in the boundaries of the restored

image. Therefore, we propose to use a third equalizer

corresponding to the median shift

))2/()2/(()3,3( 22 nmnm ⋅= ….(10)

where . the integer part. Solving the equalizers η1, η2 and

η3 consists in solving the following set of linear equations

∑∑

∑∑

∑∑

Κ

=

Κ

=

Κ

=

Κ

=

Κ

=

Κ

=

−−∗=−−∗

−−∗=−−∗

−−∗=−−∗

1

11,3

1

33,1

1

33,2

1

22,3

1

22,1

1

11,2

),(),(

),(),(

),(),(

k

kk

k

kk

k

kk

k

kk

k

kk

k

kk

nnmmGnnmmG

nnmmGnnmmG

nnmmGnnmmG

ηη

ηη

ηη

….(11)

V. REGULARIZED MUTUALLY REFERENCED

EQULAIZER

In the noisy case, the MRE algorithm fails in restoring

efficiently the image. This is due to the ill-conditioned filter

matrix whose inversion leads to noise amplification. In order

to come through this difficulty, we propose, here, to combine

the MRE criterion with a regularization technique [8] and

adapted to the multichannel framework. In fact, the noise

amplification is due to the largest singular values of the

Inverse Filter Matrix. Therefore, our regularization technique

simply consists in the truncation of the largest singular values

of the IFM. This truncation is realized through an adaptive

thresholding technique, which is explained below. Now, to

reduce the computational cost of the desired eigenvalues, we

exploit the Toeplitz structure and the large dimension of the

inverse filter matrix in such a way to approximate it by a block

circulant matrix [14] whose eigenvalues can be computed by

means of Fourier transform [4].

In this paper, we choose η3 among the MRE equalizers to

restore the original image as it provides the best restoration

performance compared to external shift equalizers η1 and η2

as mentioned previously. Let us write this equalizer as

]......,[ ,31,33 Κ∆ ηηη ….(12)

Let Ek be the IFM associated with γ3,K . Since Ek is a

large block Toeplitz matrix with Toeplitz blocks, a large block

circulant matrix with Ek circulant blocks can approximate it.

This approximation leads to the following estimation

=−

132

11

21

:

EEE

EEE

EEE

EEKK

K

k

L

M

L

L

….(13)

Then the above equation becomes,

∗≈

31 f

g

g

E

k

M

….(14)

We propose to truncate the largest eigenvalues of E to avoid

noise amplification when restoring the original image. A well-

known property of circulant matrices is that their eigenvalues

can be expressed as a function of the elements of the first

column. This property can be extended to block circulant

matrices by considering the first column of each column

block. Since E is a block circulant matrix with circulant blocks

Ek [14]

MeuuuT

n == Κ ][ 1L . ….(15)

where u is a vector containing the eigenvalues of E, M is a

known sparse matrix and n is the number of pixels in the

restored image.

VI. IMAGE RESTORATION USING PSF

IDENTIFICATION METHOD

a) Image restoration algorithm using Wiener filter

If the image and the noise are assumed to be generalized

stationary process, image may be restored through Wiener

filter. When the discrete Fourier transform (DFT) method is

51

used to estimate the restored image, the Wiener filter may be

expressed as follows:

+

=

xx

nn

SS

H

YHZ

2

2

….(16)

where, X, Y and H are the DFT of the real image (x), the

blurred image (y) and the blur function (h) respectively; *

denotes the conjugate operation; Snn and Sxx denote the

power spectrum of the noise and the real image. As it is

usually very difficult to estimate Snn and Sxx, the Wiener

filter is usually approximated by the following formula:

( )Γ+=

2

2

H

YHZ ….(17)

where, Γ is a positive constant. The best value of Γ is the

reciprocal of the SNR of the observed image.

b) SNR estimation of image

The estimated signal-to-noise ratio (SNR) may be a reference

to select the regularized parameters in the Wiener filtering

restoration algorithm. The SNR of a blurred image is usually

defined as follows:

= 2

2

10log10n

xSNRδ

δ ….(18)

where 2

xδ is the variance of the blurred image, and 2

nδ is the

variance of the noise.

The local variances between the flat region and the edge are

different from each other in an image. Places with large local

variance shows that there are much detail information, and

places with small local variance means that it is relatively flat

in this region. If the quality of the observed image is good (for

example, the SNR is above 30dB), it is reasonable to take the

region with the maximum local variance as the edge, and to

take the area with the minimum local variance as the flat

region. Thus, the variance of the image (δ) is taken as the

maximum local variance, and the variance of the noise (δn) is

approximated as the minimum local variance. The local

variance of the observed image (y) at position (i,j) is defined

as follows:

[ ]

( ))12)(12(

),(),(2

2

++

−++

=

∑ ∑−= −=

qp

jiljkiyp

pk

q

ql

y

yl

µ

δ ….(19)

where, p and q are the sizes of the local area; µ y denotes

the local mean value which is defined as follows:

[ ]

( ))12)(12(

),(

++

++

=

∑ ∑−= −=

qp

ljkiyp

pk

q

ql

yµ ….(20)

Generally, the local variance is taken as p=q=2, and the

template may be expressed as:

++=

11111

11111

11111

11111

11111

)12)(12[(

1

qpe

….(21)

So, the local mean value may be denoted as:

[ ]

( )

[ ]

cy

lksljkiy

qp

ljkiy

p

pk

q

ql

p

pk

q

ql

y

*

),(.),(

)12)(12(

),(

=

++=

++

++

=

∑ ∑

∑ ∑

−= −=

−= −=µ

….(22)

The local variance may be written as:

( )[ ]

( )( )( )

( )[ ] [ ][ ]

( ) cy

lkejiljkiy

qp

jiljkiy

ji

y

p

pk

q

ql

y

p

pk

q

ql

y

yk

*

),(),(,

1212

),(,

),(

2

2

2

2

µ

µ

µ

δ

==

−−++=

++

−++

=

∑∑

∑∑

−= −=

−= −=

….(23)

Thus, the SNR of image y may be estimated as the ratio of

the maximum local variance and the minimum local variance,

namely:

( ))min()max(log10 22

10 ylylSNR δδ= ….(24)

c) Estimation of Gaussian PSF

Gaussian point spread function (PSF) is the most common

blur function of many optical measurements and imaging

systems. Generally, the Gaussian PSF may be expressed as

follows:

( ) ( )

+−

=

others

Rnmnmnmh

,0

,2

1exp

2

1

),(

22

2δπδ ….(25)

where, δ is the standard deviation; R is a supporting region.

Commonly, R is denoted by a matrix with size of K×K, and K

is often an odd number. Thus, two parameters that are the size

(K) and the standard deviation (δ) need to be identified for the

Gaussian PSF. Because the Fourier transformation of a

Gaussian function is still a Gaussian function, it is impossible

to identify the parameters by the zero- crossing point in the

frequency domain. But in many cases, the isolated point and

the intensity edges in the observed image may provide the

necessary information to identify the blurring function.

In Wiener filtering algorithm, in order to restrain the parasitic

ripple induced by the boundary cutoff, the image needs to

have circular boundary. For the observed image (y) with size

of M × N, reflection symmetric extension is performed on it,

52

and the size of the extended image becomes 2M×2N. Then,

calculate the Fourier transformation of the extended image

(Y). Given a size (K) of the PSF, the error-parameter curve is

generated at different standard deviations (δ). According to

error-parameter the curves at different sizes, we can estimate

these two parameters approximately. The identification

process is expressed as follows:

Step 1: Select a standard deviation range given by the

minimum value (δmin) and the maximum value (δmax)

Step 2: Set a searching number (S), and we will get:

∆δ=(δmax-δmin) /S

Step 3: Set different sizes (K) of the PSF, repeating Step 4

Step 4: For i=1:S, repeat Step 4.1~ Step 4

Step 4.1: Compute the current standard deviation:

si ∆−+= )1(minδδ

Step 4.2: Generate the Gaussian blurring function h according

to K and δ

Step 4.3: Add zeroes to h and make it to be the size of

2M×2N, and compute its DFT (H)

Step 4.4: According to equation (24), estimate the Fourier

transformation of the real image (X)

Step 4.5: Compute the estimation error: K=||Y-ZH||2, and

normalize it

Step 5: Plot the error-parameter curves at different sizes

Step 6: Give an estimation error (e), calculate the distance

between curves, the curve once the distance (d) is smaller than

threshold T1 represent the real parameters, from which the

size (K) of the PSF is estimated

Step 7: Calculate the increment at different standard

deviations on this curve, once the increment is greater than

threshold T2, the deviation (δ) of the PSF is estimated

VII. SIMULATION RESULTS

In the following, we test the image restoration performance

using regularized MRE and identification of PSF based

restoration. Experiments were carried out for an image: the

cameraman one which has a homogeneous background with a

man in the middle. These images are adequate to measure the

ability of the developed algorithms to restore the image details

and edges as well as the homogeneous area. In all

experiments, the degraded images are altered by a set of PSFs,

which simulate a camera motion, a motion, average and a

gaussian filtering. The number of observed images

corresponding to the number of independent PSFs is K=4 and

the PSFs’ size is 3x3. We first propose to the regularized

version of the MRE algorithm and the nonregularized one for

a set of degraded images.

(a) (b)

Fig.1 (a) Original image. (b) Blurred image with motion filter

(a) (b)

Fig.2 (a) Blurred image with average filter (b) Gaussian filter

(a) (b)

(c)

Fig.3. Restoration using MRE method, the equalizer obtained

by different spatial shifts

(a) (b)

(c) (d)

Fig.4.Restored image using PSF identification of Multichannel

(a) Channel 1 (b) Channel 1 &2 (c) Channel 1,2 & 3 (d)

Channel 1,2,3 & 4.

53

Fig.5 Peak signal to noise Vs number of channels.

Fig.6.Mean square error Vs number of channels

Fig.7.Figure of merit Vs number of channels

The degraded images are shown in Fig. 1(b), 2(a) and 2(b).

The image restoration results are obtained to particular spatial

shift and depicted in Fig. 3(a), 3(b) and 3(c). This experiment

confirms the inefficiency of MRE algorithm in the noisy case.

It demonstrates the importance of the regularization to avoid

the noise amplification phenomenon associated with the image

deconvolution problem. The result of PSF identification

method, we obtained different PSF from different channel. In

this method, we have to use 4 channels. The restoration

images are shown in Fig. 4. The PSNR and MSE are

calculated and then it compared to number of channels in Fig.

5 and Fig. 6. The quality of image calculated from figure of

merit and it compared to number of channels in Fig. 7

VIII. CONCLUSION

Blind image restoration has always been one of the

challenges in image processing, as the blur function of the

imaging system is unknown in practice. This paper introduces

two Multichannel restoration techniques with regularization.

The first one is a direct restoration technique based on

regularized MRE algorithm. The MRE algorithm ensures a

perfect restoration of the original image in the noiseless case,

but is inefficient in presence of noise. Hence, in the noisy

case, the MRE algorithm was used jointly with an appropriate

regularization technique that improves significantly its

performance even at 17 dB SNRs. The second method is

proposed to estimate the parameters of Blur function. Utilizing

Wiener filter image restoration algorithm, multiple error-

parameter are generated. Utilizing the estimated PSF, image is

restored through Wiener filter. In addition, in order to show

the influence of PSF estimation on the quality of the restored

image, image restoration is performed at PSF with different

parameters.

REFERENCES

[1] W. Souidene, K. Abed-Meraim, and A. Beghdadi,

“Deterministic techniques for multichannel blind image

deconvolution,” presented at the Proc. ISSPA, Aug. 2005.

[2] W. Souidene, K. Abed-Meraim, and A. Beghdadi, “Blind

multichannel image deconvolution using regularized

MRE algorithm: Performance evaluation,” presented at

the ISSPIT, Dec. 2004.

[3] F. Sroubek and J. Flusser, “Multichannel blind

deconvolution of spatially misaligned images,” IEEE

Trans. Image Process., vol. 14, no. 7, pp. 874–883, Jul.

2005.

[4] G. J. Tee, “Eigenvectors of block circulant and alternating

circulant matrices,” New Zealand J. Math., vol. 8, pp.

123–142, 2005.

[5] F. Sroubek and J. Flusser, “Multichannel blind iterative

image restoration,” IEEE Trans. Image Process., vol. 12,

no. 9, pp. 1094–1106, Sep.2003.

[6] A. Beghdadi and B. Pesquet-Popescu, “A new image

distortion measure based on wavelet decomposition,”

presented at the ISSPA, Jul. 2003.

[7] Wirawan, “Multichannel Image Blind Restoration: Second

Order Methods”, Ph.D. Dissertation, 2002.

[8] M. K. Ng, R. J. Plemmons, and S. Qiao, “Regularization of

RIF blind image deconvolution,” IEEE Trans. Image

Process., vol. 9, no. 6, pp.1130–1134, Jun. 2000.

[9] G. B. Giannakis and R. W. Heath, “Blind identification of

multichannel fir blurs and perfect image restoration,”

IEEE Trans. Image Process., vol. 9, no. 11, Nov. 2000

[10] C. A. Ong and J. A. Chambers, “An enhanced NAS-RIF

algorithm for blind image deconvolution,” IEEE Trans.

Image Process., vol. 8, no. 7, pp. 988–992, Jul. 1999.

[11] U. A. Al Suwailem and J. Keller, “Multichannel image

identification and restoration using continuous spatial

domain modeling,” presented at the Int. Conf. Image

Processing, Oct. 1997.

[12] N. P. Galatsanos, A. K. Katsaggelos, R. T. Chin, and A.

D. Hillery, “Least squares restoration of multichannel

images,” IEEE Trans. Acoust., Speech, Signal Process.,

vol. 39, no. 10, pp. 2222–2236, Oct. 1991.

[13] H. J. Trussel, M. I. Sezan, and D. Tran, “Sensitivity of

color LMMSE restoration of images to the spectral

estimation,” IEEE Trans. Signal Process., vol. 39, no. 1,

pp. 248–252, Jan. 1991.

[14] A. Mayer, A. Castiaux, and J. P. Vigneron, “Electronic

green scattering with n-fold symmetry axis from block

circulant matrices,” Comput.Phys. Commun., vol. 109,

pp. 81–89, Mar. 1998.

54

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

55

TEXT EXTRACTION IN VIDEO

J.Dewakhar

Student, II- M.Sc- information Technology

P.G. Department of Information Technology

M.KumarasamyCollege of Engineering,

ThalavapalayamKarur-639113,Tamilnadu,India

E-Mail: [email protected], Contact No: 9789258879

K.Prabhakaran

Student, II- M.Sc- information Technology

P.G. Department of Information Technology

M.KumarasamyCollege of Engineering,

ThalavapalayamKarur-639113,Tamilnadu,India

E-Mail: mit816prabu@yahoo., Contact No: 978696921

Abstract

Text data present in images and video contain useful information for automatic

annotation, indexing, and structuring of

images. Extraction of this information involves detection, localization, tracking,

extraction, enhancement, and recognition of

the text from a given image. However,

variations of text due to differences in size, style, orientation, and alignment, as well as

low image contrast and complex background

make the problem of automatic text

extraction extremely challenging. Most work so far has made restrictive

assumptions about the nature of text

occurring in video. Such work is therefore not applicable to unconstrained, general-

purpose video.

In this paper we present methods for

automatic segmentation of caption text in digital videos. The output is directly passed

56

to a standard OCR software package in

order to translate the segmented text into ASCII. The algorithms we propose make

use of typical characteristics of text in

videos in order to enable and enhance

segmentation performance. Especially the inter-frame dependencies of the characters

provide new possibilities for their

refinement. Keywords-segmentation, caption text, OCR,

text extraction, general-purpose video

1 Introduction Nowadays the volume of data contained in

video format makes necessary to create useful tools, which allow extracting

information from these video sequences in

order to classify (e.g. indexation) or to analyse (e.g. CBIR) them without human

supervision. Contents can be perceptual,

such as colour, shapes, textures, etc., or semantic, such as objects, text or events and

its relationships. The perceptual ones are

easier to analyse automatically while the

semantic are easier to handle linguistically. Caption or superimposed text is a semantic

content whose computational cost is lower

than the cost of others semantic contents. Due to the fact that usually text is

synchronized and related to the scene, its

extraction becomes a useful feature

providing very relevant information for the semantic analysis.

Although text is easily detectable for

humans, even in the case of a foreign language, up to our knowledge, there are no

methods allowing its extraction in any kind

of video sequence. This is due to the fact that there is a wide range of text formats

(size, style, orientation), the low resolution

of the images (quality) and the complexity

of the background. Despite these facts, text lines present some

homogeneity features, which make it

detectable such as contrast, spatial cohesion, textured appearance, colour homogeneity,

stroke thickness, temporal uniformity,

movement on the sequences, position on the frame, etc.

Therefore, the aim of this paper is to

introduce our methods for automatic text segmentation in digital videos. Text features

are presented in Section 2,followed by

description of our segmentation algorithms

in Section 3, which are based on the features stated in Section 2. Information about the

text recognition step is given in Section 4.

Section 5 reviews the related work. Section 6concludes the paper.

2 Text Features Text may appear anywhere in the video and in different contexts. We discriminate

between two kinds of text: scene text and

artificial text. Scene text appears as a part of

and was recorded with the scene, whereas artificial text was produced separately from

the video shooting and is laid over the scene

in a post-processing stage, e.g. by video title machines.The following features

characterize the mainstream of artificial text

appearances: Characters are in the foreground. They are

never partially occluded. Characters are

monochrome. Characters are rigid. They do

not change their shapeframe. Characters have size restrictions. A letter is not as large

as the whole screen, nor are letters smaller

than a certain number of pixels as they would otherwise be illegible to viewers.

Character are mostly upright. Characters are

either stationary or linearly moving. Moving

characters also have a dominant translation direction: horizontally from right to left or

vertically from bottom to top. Characters

contrast with their background since artificial text is designed to be read easily.

The same characters appear in multiple

consecutive frames. Characters appear in clusters at a limited distance aligned to a

horizontal line, since that is the natural

method of writing down words and word

groups. Our text segmentation algorithms are based on these features.

3 Text Segmentation The objective of text segmentation is to produce a binary image that depicts the text

appearing in the video. Hence, standard

OCR software packages can be used to recognize the segmented text. All the

processing steps are performed on color

57

images in the RGB color space and not on

grayscale images.

3.1 Color Segmentation First each frame is segmented into suitable

objects. The monochromaticity character feature is taken as grouping criterion for

pixels and contrast with the local

environment is taken as the separation

criterion for pixels. Together with a segmentation procedure, which is capable of

extracting monochrome regions that contrast

highly to their environment under significant noise, suitable objects can be constructed.

Such a segmentation procedure preserves

the characters of artificial text occurrences. Its effect on multicolored objects and/or

objects lacking contrast to their local

environment is insignificant here.

Subsequent segmentation steps identify the regions of such objects as non-character

regions and thus eliminate them.

As a starting point we over-segment each frame by a region-growing algorithm

[15](see Figure 2). The threshold value for

the color distance is selected by the criterion to preclude that occurring characters merge

with their surroundings.

Figure 1. Original video frame

Hence, the objective of the region growing is to strictly avoid any under-segmentation

of characters. The output after the split and

merge algorithm is shown in Figure 3. Then, regions are merged to remove the over-

segmentation of characters while at the same

time avoiding their under-segmentation. Given a monochrome object in the frame

under high additive noise, these

segmentation algorithms would always split

up the object randomly into different regions. It is the objective of the merger

process to detect and merge such random

split-ups of objects. We identify random split-ups via a frame’s edge and orientation

map. If the border between two regions does

not coincident with a roughly perpendicular edge or local orientation in the close

neighborhood, the separation of the regions

is regarded as incidentally due to noise, and

they are merged. Let f be an image for which

regions are to be grown

Define a set of regions, R1,R2,….Rn, each consisting

repeat

for i=1 to n do for each

pixel,p,at the

border of Ri do

for all neighbours of p

do

Let x,y be the neighbour’s

coordinates

Let μi be the

mean grey level of pixels in Ri

if the neighbour

is unassigned and

| f(x,y) - μi | ≤ Δ

then

Add neighbour

to Ri Update μi

endif endfor

endfor

endfor until no more pixels are being

assigned to regions.

Figure 2. Region Growing Algorithm, size or

orientation from frame to

58

Orientation and Localization together allow

to detect most random split-ups of objects. Edges are localized by means of the Canny

edge detector extended to color images, i.e.

the standard Canny edge detector is applied

to each image band. Then, the results are integrated by vector addition. Edge detection

is completed by non-maximum suppression

and contrast enhancement. Figure 3.

Applying split and merge algorithm to

Figure 1. Merging regions of similar colors completes

the color segmentation. This segmentation

algorithm yields an excellent segmentation of a video with respect to the artificial

characters. Usually most of them will now

consist of one region.

3.2 Contrast Segmentation Video frame can also be segmented properly

by means of the high contrast of the

character contours to their surroundings and by the fact that the strength of the stroke of a

character is considerably less than the

maximum character size. For each video frame a binary contrast image is derived in

which set pixels mark locations of

sufficiently high absolute local contrast.

Figure 4:

Image Result after contrast analysis

The absolute local color contrast at position

I(x,y) is measured by where || || color metric employed

G k,l Gaussian filter mask

r size of the local neighborhood Then, each set pixel is dilated by half the

maximum expected strength of the stroke of

a character.

As a result, all character pixels as well as some non-character pixels, which also show

high local color contrast, are registered in

the binary contrast image. Likewise for color segmentation, the contrast threshold is

selected in such a way that, under normal

conditions, all character-pixels are captured by binary contrast image. Finally, all regions

that overlap by less than 80% with the set

pixels in the binary contrast image are

discarded. Figure 4 gives the result after applying contrast analysis to Figure 3.

3.3 Geometry Analysis

59

Characters are subjected to certain

geometric restrictions. Their height, width, width-to-height ratio and compactness do

not take on any value, but usually fall into

specific ranges of values. If a region’s

geometric features do not fall into these ranges of values the region does not meet

the requirements of a character region and is

thus discarded. The precise values of these restrictions depend on the range of the

character sizes selected for segmentation. In

our work, the geometric restrictions have been determined empirically based on the

bold and bold italic versions of the four

TrueType fonts Arial, Courier, Courier New

and Times New Roman at the sizes of 12pt,

24pt, and 36 pt (4*3*4 = 24 fonts in total).

The measured ranges of width, height, width-to-height ratio and compactness are

listed in Table 1 below.

Since we have assumed that each character consists of exactly one region after the color

segmentation step, the empirical values can

be used directly to rule out non-character regions. All regions, which do not comply

with the measured geometric restrictions, are

discarded. The resulting image after

applying size restriction is given in Figure 5.

3.4 Motion Analysis Another feature of artificial text occurrences

is that they either appear statically at a fixed

position on the screen or move linearly

across the screen. More complicated motion

paths are extremely improbable between the on- and disassembly of text on the screen.

Figure 5:

Applying size restrictions to Figure 3

Any other, more complex motion would

make it much harder to track and thus read

the text, and this would contradict the intention of artificial text occurrences.

This feature applies both to individual

characters and whole words. It is the

objective of motion analysis to identify regions which cannot be tracked or which do

not move linearly, in order to reject them as

non-character regions. The object here is to track the characters not

only over a short period of time but also

over the entire duration of their appearance

Geometri

c

Restriction

M

in

Max

Width 1 31

Height 4 29

Width-to-

height-ratio

0.

56

7.00

Compactn

ess

0.

2

1

1.00

60

in the video sequence. This enables us to

extract exactly one bitmap of every text occurring in the video.

Motion analysis can also be used to

summarize the multiple recognition results

for each character to improve the overall recognition performance. In addition, a

secondary objective of motion analysis is

that the output should be suitable for standard OCR software packages.

Essentially this means that a binary image

must be created.

Character Objects Formation A central term in motion analysis is the

character object C. It gradually collects from contiguous frames all those regions,

which belong to one individual character.

Since we assume that a character consists of exactly one region per image after the color

segmentation step, at most one region per

image can be contained in a character object. A character object C is described formally

by the triple (A, [a, e],v) where, A feature

values of the regions, which were assigned

to the character object and which are employed for comparison with ther regions,

[a,e] the frame number interval of the

regions’ ppearance, v the estimated and constant speed of the

character in pixels/ frame.

In a first step, each region ri in frame n is

compared against each character object Cj,j

∈ 1,....,J constructed from the frames 1 to

n-1. To this are compared the mean color, size and position of the region ri and the

character objects Cj,j ∈ 1,....,J.

.

If a region is sufficiently similar to the best-

matching character object, a copy of that character object will be created, and the

region added to the initial character object.

We need to copy the character object before

assigning a region to it since; at most, one region ought to be assigned to each

character object per frame. Due to necessary

tolerances in the matching procedure, however, it is easy to assign the wrong

region to a character object. The falsely

assigned region would block that character

object for the correct region. By means of

the copy, however, the correct region can still be matched to its character object. It is

decided at a later stage in motion analysis,

whether the original character object or one

of its copies is to be eliminated If a region does not match to any character

object existing so far, a new character object

will be created and initialized with the region. Also, if a region best fits to a

character object that consists of fewer than

three regions, a new character object is created and initialized with the region. This

prevents a possible starting region of a new

character object from being sucked up by a

still shorter and thus unstable character object. Finally upon the processing of frame

n, all character objects, which offend against

the features of characters, are eliminated. In detail all copies of a character object are

discarded which were created in frame n-1

but not continued in frame n as well as all character objects, which could not be

continued during the last 6 frames or whose

forecasted location lies outside the frame

and whose regions do not fit well to each other, which are shorter than 5 frames,

which consist of fewer than 4 regions or

whose regions move faster than 9 pixels/frame.Figure 6 gives the resulting

image after applying motion analysis to two

consecutive frames of Figure 5.

Text Objects Formation In order to, firstly, eliminate character

objects standing alone which either represents no character or a character of

doubtful importance, and secondly, to group

character objects into words and lines of text, character objects are merged into so-

called text objects. A valid text object Ti =

Ci1,…,C

i n (i) is formed by at least three

character objects which approximately

1. occur in the same frames, 2. show the same (linear) motion,

3. are the same mean color,

4. lie on a straight line and 5. are neighbors.

These grouping conditions result directly

from the features of Roman letters. We use a

61

fast heuristics to construct text objects: At

the beginning all character objects belong to the set of the character objects to be

considered. Then, combinations of three

character objects are built until they

represent a valid text object. These character objects are moved from the set of the

character objects into the new text object.

Next, all character objects remaining in the set, which fit well to the new text object, are

moved from the set to the text object. This

process of finding the next valid text object and adding all fitting character objects is

carried out until no more valid text objects

can be formed or until all character objects

are grouped to text objects. To avoid splintering multi-line horizontal

text into vertical groups, this basic grouping

algorithm must be altered slightly. In a first run, only text objects are constructed whose

characters lie roughly on a horizontal line.

The magnitude of the gradient of the line must be less than 0.25. In a second run,

character groups are allowed to run into any

direction.

The text objects constructed so far are still incomplete. The precise temporal range

[ai,e

i] of occurrence of each character object

C i of a text object are likely to differ

somewhat. In addition, some character

objects have gaps at frames in which, for various reasons, no appropriate region was

found. The missing characters are now

interpolated.

Figure 6:

After applying motion analysis to Figure 5

At first, all character objects are extended to

the maximum length over all character

objects of a text object, represented by [mina

i1,….,a

im(1),maxb

i1,…,b

in(i)]. The

missing regions are interpolated in two passes: a forward and a backward pass. The

backward pass is necessary in order to

predict the regions missing at the beginning of a character object.

The procedure is given in Figure 7.

4 Text Recognition

For text recognition, the OCR-Software

Development Kit Recognita V3.0 for Windows can be used. Two recognition

modules are offered: one for typed and one

for handwritten text. Since most artificial text occurrences appear in block letters, the

OCR module for typed text was used to

translate the rearranged binary text images

into ASCII text. The recognized ASCII text in each frame was written out into a

database file. The recognition result can be

improved by taking advantage of the multiple instances of the same text over

consecutive frames, because each character

in the text often appears somewhat altered from frame to frame due to noise, and

changes in background and/or position.

Combining their recognition results into one

final character result might improve the overall recognition performance.

Character Object

62

region region no region no in frame in frame region in frame region

n+1 n+2 n+4

Character Object after extension

no region region no region region no

region frame frame region frame frame region

n+1 n+2 n+4

Character Object after forward interpolation

no region region inter- region region inter- region frame

frame polated frame frame polated. n+1 n+2 n+4 n+5

63

inter- region region inter- region region inter-

polated frame frame polated frame framepolated.

5 Related Works

Numerous reports have been published about text extraction in digital video

sequences, each concentrating on different

aspects. Some employ manual annotation

[4][1], others compute indices automatically. Automatic video indexing generally uses

indices based on the color, texture, motion,

or shape of objects or whole images [2][8][13]. Sometimes the audio track is

analyzed, too, or external information such

as storyboards and closed captions is used[5]. Other systems are restricted to

specific domains such as newscasts [13],

football, or soccer [3]. None of them tries to

extract and recognize automatically the text appearing in digital videos.

Existing work on text recognition has

focused primarily on optical recognition of characters in printed and handwritten

documents in answer to the great demand

and market for document readers for office

automation systems. These systems have attained a high degree of maturity [6].

Further text recognition work can be found

in industrial applications, most of which focus on a very narrow application field. An

example is the automatic recognition of car

license plates [10]. The proposed system works only for characters/ numbers whose

background is mainly monochrome and

whose position is restricted.

There exist some proposals regarding text detection in and text extraction drom

complex images and video. In [9], Smith

and Kanade briefly propose a method to detect text in video frames and cut it out.

However, they do not deal with the

preparation of the detected text for standard

optical character recognition software. In

particular, they do not try to determine

character outlines or segment the individual characters. They keep the bitmaps

containing text as they are. Human beings

have to parse them. They characterize text as

a “horizontal rectangular structure of clustered sharp edges” [9] and use this

feature to identify text segments. Their

approach is completely intra-frame and does not utilize the multiple instances of the same

text over successive frames to enhance

segmentation and recognition performance. Yeo and Liu propose a scheme of caption

detection and extraction based on a

generalization of their shot boundary

detection technique for abrupt and gradual transitions to locally restricted areas in the

video [12]. According to them, the

appearance and disappearance of captions are defined as a localized cut or dissolve.

Thus, their approach is inherently inter-

frame. It is also very cheap computationally

since it operates on compressed MPEG videos. However, captions are only a small

subset of text appearances in video.

Yeo and Liu’s approach seems to fail when confronted with general text appearance

64

produced by video title machine, such as

scroll titles, since these text appearances cannot just be classified by their sudden

appearance and disappearance. In addition,

Yeo and Liu do not try to determine the

characters’ outline, segment the individual characters and translate these bitmaps into

text.

Zhong et. al. propose a simple method to locate text in complex images [14]. Their

first approach is mainly based on finding

connected monochrome color regions of certain size, while the second locates text

based on its specific spatial variance. Both

approaches are combined into a single

hybrid approach. Wu et. al. propose a four-step system that

automatically detects text in and extracts it

from images such as photographs [11]. First, text is treated as a distinctive texture.

Potential text locations are found by using 3

second-order derivatives of Gaussians on three different scales. Second, vertical

strokes coming from horizontally aligned

text regions are extracted. Based on several

heuristics, strokes are grouped into tight rectangular bounding boxes.

These steps are then applied to a pyramid of

images generated from the input images in order to detect text over a wide range of font

sizes. The boxes are then fused at the

original resolution. In a third step, the

background is cleaned up and binarized. In the fourth and final step, the text boxes are

refined by repeating steps 2 and 3 with the

text boxes detected thus far. The final output produces two binary images for each text

box and can be passed by any standard OCR

software. Another interesting approach to text

recognition in scene images is that of Ohya,

Shio, and Akamatsu [7]. Text in scene

images exists in 3-D space, so it can be rotated, tilted, slanted, partially hidden,

partially shadowed, and it can appear under

uncontrolled illumination. In view of the many possible degrees of freedom of text

characters, Ohya et al. restricted characters

to being almost upright, monochrome and not connected, in order to facilitate their

detection.

6 Conclusion

We have presented our new approach to text

segmentation and text recognition in digital

video. The text segmentation algorithms operate on uncompressed frames and make

use of intra- and inter-frame features of text

appearances in digital video. The algorithm can be been tested on title sequences of

feature films, newscasts and commercials

References

[1] Marc Davis. Media Streams:

Representing Video for Retrieval and Repurposing. Proc. ACM Multimedia 94,

pp. 478-479, San Francisco, CA, USA,

October 15-20, 1994. [2] M. Flickner, H. Sawney, et al. Query by

Image and Video Content: The QBIC

System. IEEE Computer, Vol. 28, No. 9, pp.

23-32, September 1995. [3] Y. Gong, L. T. Sin, C. H. Chuan, H. J.

Zhang, and M. Sakauchi. Automatic Parsing

of TV Soccer Programs. In Proc. International Conference of Multimedia

Computing and Systems, pp. 167-174, May

1995. [4] R. Hjelsvold, S. Langørgen, R.

Midstraum, and O. Sandstå. Integrated

Video Archive Tools. Proc. ACM

Multimedia 95, San Francisco, CA, Nov. 1995, pp. 283-293.

65

[5] C. J. Lindblad, D. J. Wetherall, and W.

Stasior. View- Station Applications: Implications for Network Traffic.IEEE

Journal on Selected Areas in

Communications, Vol. 13, 1995.

[6] S. Mori, C. Y. Suen, and K. Yamamoto. Historical Review of OCR Research and

Development. Proceedings of the IEEE,

Vol. 80, No. 7, pp. 1029-1058, July 1992. [7] J. Ohya, A. Shio, and S. Akamatsu.

Recognizing Characters in Scene Images.

IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 16, No. 2, pp.

214-220, 1994.

[8] J. R. Smith and S. F. Chang. Visualseek:

A Fully Automated Content-based Image Query System. In ACM Multimedia, pp. 87-

98, November 1996.

[9] M. A. Smith and T. Kanade. Video Skimming for Quick Browsing Based on

Audio and Image Characterization. Carnegie

Mellon University, Technical Report CMU-CS-95-186, July 1995.

[10] M. Takatoo et al., Gray Scale Image

Processing Technology Applied to Vehicle

License Number Recognition System. In Proc. Int. Workshop Industrial Applications

of Machine Vision and Machine

Intelligence, pp. 76-79, 1987. [11] V. Wu, R. Manmatha and E. M.

Riseman. Finding Text in Images. In

Proceedings of Second ACM International

conference on Digital Libraries, Philadelphia, PA, S. 23-26, July 1997.

[12] B.-L. Yeo and B. Liu. Visual Content

Highlighting via Automatic Extraction of Embedded Captions on MPEG compressed

Video. In Digital Video Compression:

Algorithms and Technologies, Proc. SPIE 2668-07 (1996).

[13] H.J. Zhang, Y. Gong, S. W. Smoliar,

and S. Y. Tan. Automatic Parsing of News

Video. Proc. IEEE Conf.on Multimedia Computing and Systems, pp. 45-54, 1994.

[14] Y. Zhong, K. Karu and A. K. Jain.

Locating Text in Complex Color Images. Pattern Recognition, Vol. 28, Nr. 10, S.

1523-1535, October 1995.

[15] S. Zucker. Region Growing: Childhood

and Adolescence. Computer Graphics and

Image Processing, Vol. 5, pp. 382-399, 1976.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

66

Information Retrieval using Moment techniques

N.Anilkumar1,Dr Amitabh wahi2

1ME (software engineering)

2Professor and Head of the Department

Department of Information Technology

Bannari Amman Institute of Technology, Sathyamangalam

E-mail id: [email protected]

Abstract:

Moment technique provides a valuable

method for detecting various information

from the images. The image features we

investigate in this study are a set of

combinations of geometric image moments

which are invariant to translation, scale,

rotation and contrast. Moments and

functions of moments have been

extensively employed as invariant global

features of images in pattern recognition.

While earlier the representation of objects

using two dimensional images that are

taken from different angles for objective. In

this study, after preprocessing divide the

image in to four similar non overlapping

datasets and apply any moment technique

to obtain the accurate value. This will

increase the strength of the value and find

the exact image from the database or pool

of images.

Keywords: Thermal medical images,

medical infrared images, content- based

image retrieval, moment invariants.

1 Introduction

While image analysis and pattern

recognition techniques have been applied

to infrared (thermal) images for many years

in astronomy and military applications,

relatively little work has been conducted on

the automatic processing of thermal

medical images. Furthermore, those few

approaches that have been presented in

67

the literature are all specific to a certain

application or disease such as the detection

of breast cancer as in [6]. In this paper we

consider the application of content-based

image retrieval (CBIR) for thermal medical

images as a more generic approach for the

analysis and interpretation of medical

infrared images. CBIR allows the retrieval of

visually similar and hence usually relevant

images based on a pre-defined similarity

measure between image features derived

directly from the image data. In terms of

medical infrared imaging, images that are

similar to a sample exhibiting symptoms of

a certain disease or other disorder will be

likely to show the same or similar

manifestations of the disease. These known

cases together with their medical reports

should then provide a valuable asset for the

diagnosis of the unknown case.

1.1 Thermal medical imaging

Advances in camera technologies

and reduced equipment costs have lead to

an increased interest in the application of

infrared imaging in the medical fields [2].

Medical infrared imaging uses a camera

with sensitivities in the infrared to provide a

picture of the temperature distribution of

the human body or parts thereof. It is a

non-invasive, radiation-free technique that

is often being used in combination with

anatomical investigations based on x-rays

and three-dimensional scanning techniques

such as CT and MRI and often reveals

problems when the anatomy is otherwise

normal. It is well known that the radiance

from human skin is an exponential function

of the surface temperature which in term is

incensed by the level of blood perfusion in

the skin. Thermal imaging is hence well

suited to pick up changes in blood perfusion

which might occur due to animation,

angiogenesis or other causes. Asymmetrical

temperature distributions as well as the

presence of hot and cold are known to be

strong indicators of an underlying

dysfunction [8]. Computerized image

processing and pattern recognition

techniques have been used in acquiring and

evaluating medical thermal images [5, 9]

and proved to be important tools for clinical

diagnostics.

1.2 Content-based image retrieval

Content-based image retrieval has

been an active research area for more than

a decade. The principal aim is to retrieve

digital images based not on textual

annotations but on features derived directly

from image data. These features are then

stored alongside the image and serve as an

index. Retrieval is often performed in a

query by example fashion where a query

image is provided by the user. The retrieval

system is then searching through all images

in order to

Find those with the most similar indices

which are returned as the candidates most

alike to the query. A large variety of

features have been proposed in the CBIR

literature [7]. In general, they can be

grouped into several categories: color

features, texture features, shape features,

sketch features, and spatial features. Often

68

one or more feature types are combined in

order to improve retrieval performance.

2 Existing systems

2.1 Retrieving thermal medical

images

In this paper we report on an initial

investigation on the use of CBIR for thermal

medical images. One main advantage of

using this concept is that it represents a

generic approach to the automatic

processing of such images. Rather than

employing specialized techniques which will

capture only one kind of disease or defect,

image retrieval when supported by a

sufficiently large medical image database of

both 'healthy' and 'sick' examples will

provide those cases that are

Most similar to a given one, the query by

example method is perfectly suited for this

task with the thermal image of an

'unknown' case as the query image. The

features we propose to store as an index for

each thermal image are invariant

combinations of moments of an image. Two

dimensional geometric moments Mpq of

order p + q of a density distribution function

f(x; y) are defined as

In terms of a digital image g(x; y) of size N

_M the calculation of Mpq becomes

discredited and the integrals are hence

replaced by sums leading to

Rather than Mpq often central moments

With

are used, i.e. moments where the centre of

gravity has been moved to the origin

(i.e. _10 = _01 = 0). Central moments have

the advantage of being invariant to

translation.

It is well known that a small number of

moments can characterize an image fairly

well; it is equally known that moments can

69

be used to reconstruct the original image

[1]. In order to achieve invariance to

common factors and operations such as

scale, rotation and contrast, rather than

using the moments themselves algebraic

combinations thereof known as moment

invariants are used that are independent of

these transformations. It is a set of such

moment invariants that we use for the

retrieval of thermal medical images. In

particular the descriptors we use are based

on Hu's original moment invariants given by

[1]

M1=µ20+µ02

M2= (µ20-µ02)2+4µ211

M3= (µ30-3µ12)2+3(µ21+µ03)2

M4= (µ30+µ12)2+ (µ21+µ03)2

M5= (µ30+µ12) (µ30-µ12) [(µ30+µ12)2-

3(µ21+µ03)2] +

(3µ21-µ03) (µ21+µ03) [3(µ30+µ12)2-

3(µ21+µ03)2]

M6= (µ20-µ02) [(µ30+µ12)2- (µ21+µ03)2]

+ 4µ11 (µ30+µ12) (µ21+µ03)

M7= (3µ21-µ03) (µ30-µ12+) [(µ30+µ12)2-

3(µ21+µ03)2] + (µ30+µ12)

[3(µ30+µ12)2-3(µ21+µ03)2]

Combinations of Hu's invariants can be

found to achieve invariance not only to

translation and rotation but also to scale

and contrast [4]

The distance between two invariant vectors

computed from two thermal images I1 and

I2 is defined as

where C is the covariance matrix of the

distribution.

2.2 Experimental results

70

The moment invariant descriptors

described above were used to index an

image database of 530 thermal medical

images provided by the University of

Glamorgan [3]. An example of an image of

an arm was used to perform image retrieval

on the whole dataset. The result of this

query is given in Figure 1 which shows those

20 images that were found to be closest to

the query (sorted according to descending

similarity from left to right, top to bottom).

It can be seen that all retrieved images

contain an arm of a subject. Unfortunately,

due to the lack of enough samples of cases

of known diseases, such retrieval as

outlined earlier cannot be performed at the

moment. We are in the processes of

collecting a large number of thermo grams,

both images of 'normal' people [3] and

cases of known symptoms of diseases. This

documented dataset will then provide a

tested for the evaluating of our method

proposed in this paper as well as future

approaches.

3 Proposed systems:

In this proposal we are doing the

invariant feature extraction for object

classification. After finding edge of the

image then calculating the moment using

Hu moment technique. It is similar to the

existing approach; here we are dividing the

image in to similar parts and applying the

Hu moments technique. By this method we

can get the accurate moment value and

with that we can compare the image with

database. The database had related images

of specific application. Which image we

want to compare take it as input image and

compare that with database. Main aim of

the proposal is giving single input and

getting single output. For this we divide the

input image in to similar parts and apply the

Hu moments to all separately. We will get

the fractional value after finding the

moment invariants. Related work of the

proposed system is shown below.

3.1 Related work

Dividing the image in to similar

none overlapping blocks and apply the Hu

moment technique.

71

Fig2. Divide image in to similar block .

4 Conclusions:

We have investigated the

application of content-based image

retrieval to the domain of medical infrared

images. Each image is characterized by a set

of moment invariants which are

independent to translation, scale, rotation

and contrast. Retrieval is performed by

returning those images whose moments are

most similar to the ones of a given query

image. The aim is that to give single input

and obtain single output from the database.

5 References:

1. M.K. Hu. Visual pattern recognition by

moment invariants. IRE Transactions on

Information Theory, 2002

2. B.F. Jones. A re-appraisal of infrared

thermal image analysis for medicine. IEEE

Trans. Medical Imaging, 2000

3. B.F. Jones. EPSRC Grant GR/R50134/01

Report, 2001.

4. S. Maitra. Moment invariants.

Proceedings of the IEEE, 1979.

5. P. Plassmann and B.F. Jones. An open

system for the acquisition and evaluation of

medical thermo logical images. European

Journal on Thermology,1997

6. H. Qi and J. F. Head. Asymmetry analysis

using automatic segmentation and

Classification for breast cancer detection in

thermo grams. In 23rd Int. Conference IEEE

Engineering in Medicine and Biology, 2001.

7. A.W.M. Smeulders, M. Worring, S.

Santini, A. Gupta, and R.C. Jain. Content-

based image retrieval at the end of the

early years. IEEE Trans. Pattern Analysis and

Machine Intelligence, December 2000.

8. S. Uematsu. Symmetry of skin

temperature comparing one side of the

body to the Other. Thermology, 1985.

9. B. Wiecek, S. Zwolenik, A. Jung, and J.

Zuber. Advanced thermal, visual and

72

Radiological image processing for clinical

diagnostics. In 21st Int. Conference IEEE

Engineering in Medicine and Biology, 1999.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

73

INTELLIGENT FACE PARTS GENERATION SYSTEM FROM

FINGER PRINTS

A.Gowri, II ME

Department of Information Technology

Bannari Amman Institute of Technology

[email protected]

Mr.S.Daniel Madhan Raja

Senior lecture in Department of Information Technology

Bannari Amman Institute of Technology

Sathyamangalm

[email protected],

ABSTRACT

Biometric dashed person identification systems are used to provide alternative solutions for security.

Although many approaches and algorithm for biometric recognition techniques have been developed

and proposed in the literature. I have tried to generate the face of a person using only fingerprint

biometric feature of the same person without any information about his or her face. Biometric based

person identification systems are used to provide alternative solutions for security. Although many

approaches and algorithms for biometric recognition techniques among biometric features have not

been studied in the field so far. In this study, we have analyzed the existence of any relationship

between biometric features and we have tried to obtain a biometric feature of a person from another

biometric feature of the same person .consequently, we have designed and introduced a new and

intelligent system using a novel approach based on artificial neural network for generating face masks

including eyes, nose and mouth from finger print. In addition it is shown that fingerprints and faces are

related to each other closely. In spite of the proposed system is initial study and it is still under

development, the results are very encouraging and promising. Also proposed work is very important

from view of the point that it is a new research area in biometrics.

74

1. INTRODUCTION

Biometrics is automated methods of

recognizing a person based on a

physiological or behavioral characteristic.

Among the features measured are; face

fingerprints, hand geometry, handwriting,

iris, retinal, vein, and voice. Biometric

technologies are becoming the foundation

of an extensive array of highly secure

identification and personal verification

solutions. As the level of security breaches

and transaction fraud increases, the need

for highly secure identification and personal

verification technologies is becoming

apparent.

Biometric-based solutions are able to

provide for confidential financial

transactions and personal data privacy. The

need for biometrics can be found in federal,

state and local governments, in the military,

and in commercial applications. Enterprise-

wide network security infrastructures,

government IDs, secure electronic banking,

investing and other financial transactions,

retail sales, law enforcement, and health

and social services are already benefiting

from these technologies. Figure 1 illustrates

a general biometric system

Figure 1. The basic block diagram of a

biometric system

2. RELATED WORKS

2.1 Neural network

A biological neural network is composed of

a group or groups of chemically connected

or functionally associated neurons. A single

neuron may be connected to many other

neurons and the total number of neurons

and connections in a network may be

extensive. Connections, called synapses, are

usually formed from axons to dendrites,

though dendrodendritic microcircuits and

other connections are possible. Apart from

the electrical signaling, there are other

forms of signaling that arise from

neurotransmitter diffusion, which have an

effect on electrical signaling. As such, neural

networks are extremely complex.

Artificial intelligence and cognitive

modeling try to simulate some properties of

neural networks. While similar in their

techniques, the former has the aim of

75

solving particular tasks, while the latter

aims to build mathematical models of

biological neural systems.

The artificial intelligence field, artificial

neural networks have been applied

successfully to speech recognition, image

analysis and adaptive control, in order to

construct software agents (in computer and

video games) or autonomous robots. Most

of the currently employed artificial neural

networks for artificial intelligence are based

on statistical estimation, optimization and

control theory.

The cognitive modeling field involves the

physical or mathematical modeling of the

behavior of neural systems; ranging from

the individual neural level (e.g. modelling

the spike response curves of neurons to a

stimulus), through the neural cluster level

(e.g. modelling the release and effects of

dopamine in the basal ganglia) to the

complete organism (e.g. behavioral

modelling of the organism's response to

stimuli).

Automatic fingerprint recognition system

Automated fingerprint verification is a

closely-related technique used in

applications such as attendance and access

control systems. On a technical level,

verification systems verify a claimed

identity (a user might claim to be John by

presenting his PIN or ID card and verify his

identity using his fingerprint), whereas

identification systems determine identity

based solely on fingerprints.

With greater frequency in recent years,

automated fingerprint identification

systems have been used in large scale civil

identification projects. The chief purpose of

a civil fingerprint identifications system is to

prevent multiple enrollments in an

electoral, welfare, driver licensing, or

similar system. Another benefit of a civil

fingerprint identifications system is its use

in background checks for job applicants for

highly sensitive posts and educational

personnel who have close contact with

children.

The fingerprint recognition problem can

be grouped into two sub-domains: one is

fingerprint verification and the other is

fingerprint identification (Figure 2). In

addition, different from the manual

approach for fingerprint recognition by

experts, the fingerprint recognition here is

referred as AFRS (Automatic Fingerprint

Recognition System), which is program-

based.

Figure 2 Verification vs. Identification

Fingerprint verification is to verify the

authenticity of one person by his

fingerprint. The user provides his fingerprint

together with his identity information like

his ID number. The fingerprint verification

system retrieves the fingerprint template

76

according to the ID number and matches

the template with the real-time acquired

fingerprint from the user. Usually it is the

underlying design principle of AFAS

(Automatic Fingerprint Authentication

System).

Fingerprint identification is to specify one

person’s identity by his fingerprint(s).

Without knowledge of the person’s identity,

the fingerprint identification system tries to

match his fingerprint(s) with those in the

whole fingerprint database. It is especially

useful for criminal investigation cases. And

it is the design principle of AFIS (Automatic

Fingerprint Identification System).

However, all fingerprint recognition

problems, either verification or

identification, are ultimately based on a

well-defined representation of a fingerprint.

As long as the representation of fingerprints

remains the uniqueness and keeps simple,

the fingerprint matching, either for the 1-

to-1 verification case or 1-to-m

identification case, is straightforward and

easy.

Face recognition system

Face recognition process is defined as the

identification or verification of individuals

from still images or video images of their

face using a stored database of faces. This

recognition process is really complex and

difficult due to numerous factors affecting

the appearance of an individual’s facial

features such as 3D pose, facial expression,

hair style, make up, etc.[9]. In addition to

these varying factors, lighting, background,

scale, noise and face occlusion and many

other possible factors make these tasks

even more challenging [9]. In general, a face

recognition system (FRS) consists of three

main steps. These steps cover detection of

the faces in a complicated background,

localization of the faces followed by

extraction of the features from the faces

regions and finally identification or

verification tasks.

Many effective and robust methods have

been proposed for the face recognition

[10].Knowledge-based methods encode

human knowledge of what constitutes a

typical face. In general the rules capture the

relationships between facial features.

Feature invariant methods aim to find

structural features that exit even when the

pose, viewpoint or lighting conditions vary

to locate faces. Theses two methods are

designed mainly for face localization.

Template matching based methods are

from the face recognition techniques which

in several standard patterns of a face were

stored to describe the face as a whole or

the facial features separately. The

correlations between an input image and

the stored patterns are computed for

detection and recognition. Theses methods

have been used for both face localization

and detection. The last main group of

techniques is appearance –based methods

that operate directly on images or

appearances of face objects and process the

images as tow dimensional holistic patterns.

In theses approaches, models or templates

are learned from a set of training images

capturing the representative variability of

facial appearance. Theses learned models

are used for detection and recognition.

Appearance- based methods are designed

mainly for face detection [11]. Face part

generating system from

fingerprint:

77

As briefly expressed in the previous

sections, fingerprint verification and face

recognition topics have been received

significantly more attention. The aims of

this study are to establish a relationship

among fingerprints and faces (Fs&Fs), to

analyze this relationship and to generate

the face features from fingerprints,

requiring no priori knowledge about faces,

using a system equipped with the best

parameter settings. The majority of these

aims were achieved in this work.

Our motivation in this study arises from

biological and physiological conditions, as

briefly reviewed below. It is known that the

phenotype of the biological organism is

uniquely determined by the interaction of a

specific genotype and a specific

environment [12]. Physical appearances of

faces and fingerprints are also a part of an

individual’s phenotype. In the case of

fingerprints, the genes determine the

general characteristics of the pattern [12].

In dermatoglyphics studies, the maximum

generic difference between fingerprints has

been found among individuals of different

races. Unrelated persons of the same race

have very little generic similarity in their

fingerprints, parent and child have some

generic similarity as they share half of the

genes, siblings have more similarity and the

maximum generic similarity is observed in

the identical twins, which have the closest

genetic relationship [13]. Some scientists in

biometrics field have focused on analyzing

the similarities in fingerprint minutiae

patterns in identical twin fingers [13], and

have confirmed the claim that the

fingerprints of identical twins have a large

class correlation. In addition to this class

correlation, other correlations based on

generic attributes of the fingerprints such

as ridge count, ridge width, ridge separation

and ridge depth were also found to be

significant in identical twins. In the case of

faces, the situation is very similar with the

fingerprints. The general characteristics of

the face patterns were determined by the

genes and the maximum generic difference

between faces has been found among

individuals of different races. Very little

generic similarity was found in the faces of

unrelated persons of the same race. Parent

and child have some generic similarity as

they share half of the genes, siblings have

more similarity and the maximum generic

similarity is observed in the identical twins,

which bear the closest

genetic relationship. A number of studies

have especially focused on analyzing the

significant correlation among faces and

fingerprints of the identical twins. The large

correlation among biometrics of identical

twins was repeatedly indicated in the

literature by declaring that identical twins

would cause vulnerability problems in

biometrics based security applications. For

example, the similarity measure of identical

twin fingerprints is reported as much as

95%. In the case of faces of identical twins,

the situation is very similar. The reason of

this high degree similarity measure was

explained in some studies as: identical twins

have identical DNA except for the generally

undetectable micro mutations that begin as

soon as the cell starts dividing. Fingerprints

and faces of identical twins start their

development from the same DNA, so they

show considerable generic similarity.

Generally, it is a simple process for an

individual to distinguish the fingerprints or

faces of different people. However,

78

distinguishing the fingerprints or faces of

identical twins is a very difficult and

complicated process, not only for the eyes

and brain of a human being but also for

biometric based recognition systems. The

high degree of similarity in fingerprints and

faces of identical twins, of examples are

shown in Figure 4, converts this simple

recognition process to a hard task.

Face part generating system

The intelligent face parts generation system

from fingerprints (IFPGSF) system was

designed to generate face border of a

person using only one fingerprint of the

same person without having any

information about his or her face. The

architecture of the IFPGSF system is given in

figure 4.1

The IFPGSF system to establish a

relationship among Fs&Fs can be

summarized as follows: A multimodal

database having Fs&Fs of 120 people was

established. Fingerprint feature sets were

obtained from the fingerprints.

Enhancements, binarization, thinning, and

feature extraction processes were applied

to the acquired fingerprint to achieve this

process. Face feature sets were obtained

from the faces. 22 points were used to

represent for a face border. Training and

test data sets were established to train and

test the ANN predicators in the IFPGSF

system.

Feature invariant methods aim to find

structural features that exist even when the

pose, viewpoint or lighting conditions vary

to locate faces. The features sets of the

faces a feature based approach has been

preferred from the face recognition. A

minutiae-based approach to get the feature

sets of the fingerprints. Actually minutiae-

based approaches rely on the physical

features of the fingerprint. Therefore it is

reasonable that the feature sets of both

face and fingerprint should be obtained in

the same way. The template that was

shaped from manually extracted 35 points

to represent a face. In comparison to the

approach proposed, increasing the number

Data Enrollment:

(Fs&Fs)

Feature Extraction:

(Fs&Fs)

ANN module:

(Train &test)

Test & Evaluation:

(Minimize the error in

training evaluate the results

in test)

79

of reference points helped to represent the

faces more accurately and sensitively

Experimentation Results

A fingerprint database from the FVC2000

(Fingerprint Verification Competition 2000)

is used to test the experiment performance.

My program tests all the images without

any fine-tuning for the database. The

experiments show my program can

differentiate imposturous minutia pairs

from genuine minutia pairs in a certain

confidence level. Furthermore, good

experiment designs can surely improve the

accuracy as declared by [10]. Further

studies on good designs of training and

testing are expected to improve the result.

Here is the diagram for Correct Score and

Incorrect Score distribution:

Figure6 Distribution of Correct Scores and

Incorrect Scores

Red line: Incorrect Score

Green line: Correct Scores

It can be seen from the above figure that

there exist two partially overlapped

distributions. The Red curve whose peaks

are mainly located at the left part means

the average incorrect match score is 25. The

green curve whose peaks are mainly

located on the right side of red curve means

the average correct match score is 35. This

indicates the algorithm is capable of

differentiate fingerprints at a good correct

rate by setting an appropriate threshold

value.

The above diagram shows the FRR

and FAR curves. At the equal error rate

25%, the separating score 33 will falsely

reject 25% genuine minutia pairs and falsely

accept 25% imposturous minutia pairs and

has 75% verification rate. The high incorrect

acceptance and false rejection are due to

some fingerprint images with bad quality

and the vulnerable minutia match algorithm

80

Figure 7 FAR and FRR curve

Blue dot line: FRR curve

Red dot line: FAR curve

Conclusion

In this study generate the face parts from

fingerprint is successfully achieved and

introduced. In addition the relationship

among biometrics are also experimentally

shown in order to do the experiments

easily and effectively, an IFPGSF is designed,

implemented and introduced for

generating the face parts from fingerprint

without having any information about his or

her face .

REFERENCE

[1] Zhang, Q., Yan, H., Fingerprint

classification based on extraction and

analysis of singularities and pseudo

ridges, Pattern Recognition, no. 11, pp.

2233-2243 (2004)

[2] Ozkaya, N., Sagiroglu, S., Wani, A., An

intelligent automatic fingerprint recognition

system design, 5th International Conference

on Machine Learning and Applications, pp:

231 – 238 (2006).

[3] Haykin, S., Neural Networks: A

Comprehensive Foundation, Macmillan

College Publishing Company, New

York,(1994).

[4] Maio, D., Maltoni D., Neural network

based minutiae filtering in fingerprints, 14th

International Conference on Pattern

Recognition, pp. 1654 -1658 (1998)

[5] A. Jain and L. Hong, On-line Fingerprint

Verification, Proc. 13th ICPR, Vienna, pp.

596-600, 1996.

[6] N.Ozkaya, S.Sagiroglu, Intelligent face

border generation system from fingerprint.

2008 IEEE International Conference on fuzzy

systems (FUZZY2008)

[7] Jain, A., Prabhakar, S., Pankanti, S., on

the similarity of identical twin fingerprints,

Pattern Recognition 35 (11), 2653–2663

(2002).

[8] Anil Jain, Lin Hong and Yatin Kulkarni

F2ID: A Personal Identification System Using

Faces and Fingerprints, proc. 14th

International Conference on Pattern

Recognition, (1998), Brisbane

pp.1373_1375.

[9] D.Boucha and A. Amira, structural

hidden Markov Models for Biometrics:

Fusion of face and fingerprint “, In special

issue of pattern recognition, Journal,

feature Extraction and Machine Learning for

81

Robust Multimodal Biometrics(2007),Article

in press, available.

[10] N.Ozkaya, S.Sagiroglu, “Intelligent

Mask prediction system”. 2008

International Joint Conference on Neural

Networks (IJCNN 2008)

[11] Jain, A., Prabhakar, S., Pankanti, S., on

the similarity of identical twin fingerprints,

Pattern Recognition 35 (11), 2653–2663

(2002).

[12] Cummins, H., Midlo, C., Fingerprints,

Palms and Soles: An Introduction to

Dermatoglyphics, Dover Publications Inc.,

New York, 1961.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

83

Abstract— The Semantic Web brings structure and meaningful

content to the Web; The Semantic Web is specifically a web of

machine-readable information whose meaning is well defined by

standards. Ontologies are at the heart of semantic web, which

define concepts and relationships and make global semantic

interoperability possible. This is assisting the ontology engineers

and domain experts in operating ontology integration. One of

the main hurdles towards a wide endorsement of ontologies is

the high cost of constructing them. Reuse of existing ontologies

offers a much cheaper alternative than building new ones from

scratch. The main advantage of Ontology integration helps

reduce end user effort and produce better quality ontology. The

above issues pertaining to ontology integration has become the

subject matter for this survey and has been exhaustively

presented.

Index Terms — Semantic Web, Ontology Integration, Ontology

Reuse.

I. INTRODUCTION

The proliferation of information on the World Wide Web

(WWW) has made it necessary to make all this information

not only available to people, but also to machines. The

Semantic Web is a web of data. The Semantic Web can bring

structure and meaningful content to the Web. This involves

moving the Web from a repository of data without logic to a

level where it is possible to express logic through knowledge

representation systems. The vision for the Semantic Web is to

augment the existing Web with resources more easily

interpreted by programs and intelligent agents.

The term “ontology” has been introduced to computer

science as a means to formalize the kinds of things that can

be talked about in a system or a context. With a long-standing

tradition in philosophy, where “Ontology” denotes “the study

of being or existence”, ontologies provide knowledge

engineering and artificial intelligence support for modeling

some domain of the world in terms of labeled concepts,

attributes and relationships, usually classified in

specialization/generalization hierarchies.

Ontologies are widely being used to enrich the semantics of

the web, and the corresponding technology is being developed

to take advantage of them [1]. An ontology is defined as “a

formal, explicit specification of a shared conceptualization”

[2], where formal refers to the meaning of the specification

which is encoded in a logic-based language, explicit means

concepts, properties, and axioms are explicitly defined,

shared indicates that the specification is machine readable,

and conceptualization models how people think about things

of a particular subject area.

Ontologies are likely to be everywhere, and constitute the

core of many emerging applications in database integration,

peer-to-peer systems, e-commerce, semantic web services, and

social networks [3]. With the infrastructure of the semantics

web, we witness a continuous growth in both the number and

size of available ontologies developed to annotate knowledge

on the web through semantics markups to facilitate sharing

and reuse by machines. This, on the other hand, has resulted

in an increased heterogeneity in the available information as

different parties adopt different ontologies. The ontologies are

developed with different purpose in mind; therefore the

ontologies are modeled in different ways. For example, the

same entity could be given different names in different

ontologies or it could be modeled or described in different

ways.

Despite advances in physical and syntactic connectivity, the

goal of having networks of seamlessly connected people,

software agents and IT systems remains elusive and

ontologies together with semantics based technologies could

be the key in resolving semantic issues [4]. Semantic

interoperability may be achieved through semantic processing

to provide means to address the heterogeneity gap between

ontologies.

Nowadays, the increasing amount of semantic data

available on the Web leads to a new stage in the potential of

Semantic Web applications. However, it also introduces new

issues due to the heterogeneity of the available semantic

resources. One of the most remarkable is redundancy, that is,

the excess of different semantic descriptions, coming from

different sources, to describe the same intended meaning.

Ontology integration is the process of creating a new

ontology from two or more existing ontologies with

overlapping parts. Ontology matching is the process of

discovering similarities between two source ontologies.

Ontology matching is carried out through the application of

match operator, similarity measures, In ontology integration,

a new ontology is created which is the union of source

ontologies in order to capture all the knowledge from the

original ontologies.

The rest of the paper is structured as follows. Section 2

presents existing approaches of ontology integration. A

A Survey on Ontology Integration

S.Varadharajan

Department of Computer Science,

Pondicherry University, Puducherry, India.

[email protected]

R. Sunitha, Lecturer,

Department of Computer Science,

Pondicherry University, Puducherry, India.

[email protected]

84

section 3 presents basic concepts of ontology integration,

ontology integration approaches and ontology matching

techniques. Section 4 discusses the ontology integration tools.

Section 5 depicts the challenges in ontology integration.

Section 6 presents the research issues in ontology integration.

Finally Section 7 concludes the work.

II. EXISTING APPROACHES

The hybrid approach [5] of ontology integration is

combination of materialized (data warehouse) and virtual

views. The author‟s claim that, while much work is still

ahead, their experiments so far indicate that the ideas used in

this work are promising which may result in significant

theoretical as well as practical contributions. The

materialized and virtual views are needed to improve the

performance and complexity of reasoning.

The HCONE [6] project is the design and implementation

of a human-centered and collaborative workbench for the

engineering of ontologies. The workbench aims to support the

HCOME methodology for ontology engineering developed by

members of the AI-Lab and to follow up the prototype version

of Human Centered Ontology Engineering Environment

(HCONE). In addition to the implementation of the proposed

tool, It is further to develop and conduct research on subjects

such as the support and integration of new semantic web

standards, the integration of an ontology library, ontology

versioning and discussion facilities, and effective ontology

mapping/merging techniques. This approach is semi-

automated and requires human involvement for defining

semantics for information and for analysis of relationships

between concepts, in case there are some conflicts.

Ontology composition algebra [7] is a set of basic operators

to manipulate ontologies. The algebraic operators can be used

to declaratively specify how to compose new, derived

ontologies from the existing source ontologies. The algebraic

operators use the articulation rules to compose the source

ontologies. Articulation rule generation functions can be

implemented as semi-automatic subroutines that deploy

heuristic algorithms to articulate ontologies. The ontology

composition algebra has unary and binary operations that

enable an ontology composer to select interesting portions of

ontologies and compose them.

Schema-based alignment algorithm [8] compares each pair

of ontology terms by, firstly, extracting their ontological

contexts up to a certain depth (enriched by using transitive

entailment) and, secondly, combining different elementary

ontology matching techniques. Benchmark results show a

very good behavior in terms of precision, while preserving an

acceptable recall. The Benchmark test can only consider the

positive matching. The human behavior is requiring the

mapping the ontology terms.

Another approach for ontology mapping is Quick Ontology

Mapping [9]. The Quick Ontology Mapping first calculates

various similarities based on expert encoded rules, and then it

use neural network to integrate all these similarity measures.

In the contrast, the features are not limited to the variety of

similarities. It will not produce good result in large size

ontologies. It requires the evaluation steps to check the

syntactic errors.

III. ONTOLOGY INTEGRATION

Ontologies are increasingly used to represent the intended

real-world semantics of data and services in information

systems. One of the main hurdles towards a wide

endorsement of ontologies is the high cost of constructing

them. Reuse of existing ontologies offers a much cheaper

alternative than building new ones from scratch, yet tools to

support such reuse are still in their infancy.

Fig.1. Graphical representations of Pinto et al (1999)‟s

definitions on ontology integration D=Domain O=Ontology.

Ontology integration is the process of creating a new

ontology from two or more existing ontologies with

overlapping parts. Ontology Integration is building a new

ontology reusing other available ontologies and evaluate.

Ontology Integration is the process of finding

commonalities between two different ontologies A and B

and deriving a new ontology C. The Figure 1 shows the

Pinto et al approach of ontology definitions of ontology

Integration. Ontology Integration requires the ontology

matching for matching the concepts. Ontology matching can

be defined as the process of discovering similarities between

two ontologies, and it can be processed by exploiting a

number of different techniques

The task of integrating heterogeneous information sources

put ontologies in context. They cannot be perceived as

standalone models of the world. Consequently, the relation of

ontology to its environment plays an essential role in

information integration. The term mappings are referring to

the connection of ontology to other parts of the application

system.

A. ONTOLOGY INTEGRATION APPROACHES

The main approaches to ontology integration are including

ontology reusing, merging, and mapping. The term “ontology

integration” designates the operations and the process of

building ontologies from other ontologies, available in some

ontology development environments. This involves following

O D

O1 D1

O2

D2

O3 D3

O1

O2

O4 O3

integrate

85

methodologies that specify how to build ontologies using

other, publicly available, ontologies [10].

Ontology integration is motivated by the following three

factors. First, the use of multiple ontologies. For example,

suppose we want to build ontology about tourism in Montreal

that contains information about transportation, hotels,

restaurants, etc. This requires a lot of effort, especially since

ontologies are huge and complex. A more reasonable

approach is to reuse available ontologies on the topics, such

as transportation, restaurants, and hotels in Montreal, to build

a desired “integrated” ontology.

The second motivation is the use of an integrated view.

Suppose the company has branches, dealers, etc, distributed

around the world. The main branch needs information from

the other, such as customers, sellers, and some statistics about

the employees, sales, etc. In this case, user can query the

ontologies at various branches through proper mappings and

wrappers, thus providing a unified view in the main branch.

The third motivation for ontology integration is the merge

of source ontologies. Suppose there is much ontology on the

same topic, such as medicine, covering different aspects of

the field, which may contain overlapping information. We

might want to build a new, single ontology about the medical

field, which “unifies” the various concepts, terminologies,

definitions, constraints, etc., from the existing ontologies. For

instance, among many existing medical ontologies.

1) Ontology Reuse

The use of existing ontologies can be considered as a

„lower‟ level integration, because it does not modify the

ontologies, but merely uses the existing concepts. Since the

survey in [10], there have been some developments in

using/reusing ontologies, such as the On-To-Knowledge

project [11]. This project resulted in a software toolkit for

ontology development, maintenance, and (re)use. In [14],

they proposed to combine ontology reuse and merging,

consisting of merging local (federated) ontologies at some

stage. These “federated ontologies” are analogous to federated

databases. Another interpretation of reusing existing

ontologies, in conjunction with formal integration, is the

architecture of Fisheries ontology [12].

2) Ontology Mapping

Ontology mapping methods [13] into three categories:

mapping ontology, mapping revisions, and intersection

ontology. In mapping ontologies, a created ontology OM

contains the rules that map concepts between ontologies O1

and O2. In the mapping revisions method contains rules that

map objects in O2 to terminologies in O1 and vice versa. In

an intersection ontology, where the created ontology ON

includes the intersection of concepts common to O1 and O2,

and renames terms where necessary.

3) Ontology Merging

Ontology Merging defines the act of bringing together two

conceptually divergent ontologies or the instance data

associated to two ontologies. This is similar to work in

database merging. The ontology merging [10] defines

merging as combining different ontologies with the same

subject domain to create a unified ontology. Synonymous with

this definition, [11] defines the unification process. Moreover,

the proposal in [12] defines the merger of two ontologies as

their intersection, and the knowledge engineer is in charge of

making merger decisions. Their intention is to create a

massive governmental knowledge base. While the process of

ontology merging defined in [12] yields a merged ontology

from input ontologies, it is not clear how the performance is

affected by various assumptions about the input ontologies

when their subjects are the same, similar, or complementary.

B. ONTOLOGY MATCHING TECHNIQUES

The ontology matching process [15] aims at finding a

similarity between two ontologies which express

correspondences between their entities. This section reviews

techniques currently used for ontology matching. These

techniques are classified into element level and structure level

techniques.

1) Element-level techniques

These techniques view ontology entities or their instances

as isolated from other entities or instances. This technique is

classified into string-based, language-based, and constraints-

based.

a) String-based techniques

These techniques are used to match names of the entities in

ontologies. Such techniques are based on the similarity of the

names of entities, considered strings. The more similar the

strings, the more likely denote the same concepts. There are

numerous methods introduced for string similarity matching.

The most frequently used methods are:

Edit distance: In this method of matching two entities,

a minimal cost of operations to be applied on one entity

in order to obtain the other entity is considered.

Examples of such well-known measures are Levenshtein

distance, Needleman-Wunch distance, Smith-

Waterman, Gotoh, Monge-Elkan, Jaro measure, and

Smoa .

Normalization: To improve the matching results

between strings, a normalization operation is performed

in before matching. In particular, these operations are

case normalization, diacritics suppression, blank

normalization, link stripping, digital suppression, and

punctuation elimination.

String equality: The string equality method basically

returns 0 if the input strings compared are not identical,

and 1 if they are. An example of such a method is the

Hamming distance.

Substring test: This identifies the ratio of common

subparts between two strings. Also, it is used to compute

if a string is a substring of another string. i.e., a prefix

or suffix.

86

Token-based distances: This method considers a string

as a set of words. These methods are used to split long

strings (strings that are composed of many words) into

independent tokens.

b) Language-based techniques

These techniques measure the relatedness of concepts, for

which consider names as words in some natural language,

e.g. English. They use Natural Language Processing (NLP)

techniques to extract meaningful terms from the text. Usually,

they are applied to words (names) of entities. The matching

similarity is determined based on linguistic relations between

words, such as synonyms and hyponyms. Many language-

based methods have been implemented in the WordNet [4].

c) Constraints-based techniques

In order to calculate the similarity between entities, these

techniques are mainly applied to the definitions of entities,

such as their types, attributes, cardinality and ranges, and the

transitivity or symmetry of their properties.

There are different methods proposed based on constraints,

which compare the properties, data types, and domains of

entities.

Property comparison: When the properties of two

classes are similar (similar names and types), it is more

likely that these two classes are similar.

Data type comparison: This compares the way in

which the values are represented, e.g. integer, float,

string.

Domain comparison: Depending on the entities to be

considered, what can be reached from a property can be

different: in classes, these are domains, while in

individuals, these are values.

2) Structure-level techniques

In contrast to element-based techniques, structure-based

techniques compare the two entities from two ontologies with

regards to the relations of these entities with other entities in

the ontologies: the more similar the two entities are the more

alike their relation would be. Mainly, there are two well-

known structure level techniques: graph-based techniques and

taxonomy-based techniques.

a) Graph-based techniques

This technique considers the ontologies to be

matched as labeled graphs. The basic idea here is that, if two

nodes from two ontologies are similar, their neighbors should

also somehow be similar.

b) Taxonomy-based techniques

These techniques are basically graph-based techniques

which consider only the specialization relation. The basic

idea they focus on is that an is-a relationship links terms that

are already similar; therefore their neighbors may also be

similar. Matching ontologies using their structure

information is important as it allows all the relations between

entities to be taken into account.

The most common techniques used for ontology matching

are taxonomy-based, since taxonomies play a pivotal role in

describing ontologies.

IV. ONTOLOGY INTEGRATION TOOLS

A. ONION

ONION (ONtology compositION) [7] is a framework

for ontology integration that uses a graph-oriented model for

the representation of ontologies. In ONION, there are two

types of ontologies, individual ontologies, referred as source

ontologies and articulation ontologies, which contain the

concepts and relationships expressed as articulation rules.

The mapping between ontologies is executed by ontology

algebra. It consists of three operations namely, intersection,

union and difference.

B. MAFRA

MAFRA [16] is part of a multi-ontology system, and

it aims to automatically detect similarities of entities

contained in two different ontologies. Both ontologies must be

normalized to a uniform representation; it is eliminating

syntax differences and making semantic differences between

the source and the target ontology more apparent. This

normalization process is done by a tool, LIFT, which brings

DTDs, XML Schema and relational databases to the

structural level of the ontology. Another interesting

contribution of the MAFRA framework is the definition of a

semantic bridge. This is a module that establishes

correspondences between entities from the source and target

ontology based on similarities found between them. All the

information regarding the mapping process is accumulated,

and populates ontology of mapping constructs; this technique

is so called Semantic Bridge Ontology (SBO).

C. PROMPT

Prompt [17] is an algorithm for ontology merging and

alignment embedded in Protege. It starts with the

identification of matching class names. Based on this initial

step an iterative approach is carried out for performing

automatic updates, finding resulting conflicts, and making

suggestions to remove these conflicts. The tools are offer

extensive merging functionalities, most of them based on

syntactic and semantic matching heuristics, which is derived

from the behavior of ontology engineers when confronted

with the task of merging ontologies. It is not a complete

automation tool. Prompt Algorithm is matching only concept

names.

D. CHIMAERA

Chimaera [18] is an ontology integration merging

and diagnosis tool developed by the Stanford University

Knowledge System Laboratory (KSL). It deals with

lightweight ontologies. The two major tasks in merging

ontologies that Chimaera support are (1) coalesce two

87

semantically identical terms from different ontologies so that

they are referred to by the same name in the resulting

ontology, and (2) identify terms that should be related by

subsumption, disjointness or instance relationships and

provide support for introducing those relationships. Chimaera

also has diagnostic support for verifying, validating and

critiquing ontologies. Chimaera can be used to solve

mismatches at the terminological level. It is also able to find

some similar concepts that have a different description at the

model level. Further, Chimaera does a great job in helping

the user to find possible edit point. The diagnostic functions

are difficult to evaluate, because their description is very

brief.

V. CHALLEGES IN ONTOLOGY INTEGRATION

The Ontology integration system is reusing existing

ontologies; the quality of those ontologies will certainly affect

the quality of the output ontology. Users might want to

restrict the system to only those ontologies that pass certain

quality tests, or are provided by specific organizations or

authors.

Some of the specific challenges in ontology integration that

we must address in the near future are:

• Finding similarities and differences between ontologies in

automatic and semi-automatic way

• Defining mappings between ontologies

• Developing an ontology-integration architecture

• Composing mappings across different ontologies

• Representing uncertainty and imprecision in mappings

One of the crucial factors when building new ontologies

from existing ones is obviously the availability of ontologies

to reuse, in terms of numbers and domain variety. Many of

the ontologies constructed by semantic web researchers and

developers are never put on the web.

This will hopefully change once ontology search engines

become more popular, and the benefits of making ontologies

available for others become more apparent. The system may

make some changes to ontologies they are reusing, declare

equivalence between their terms and terms in other

ontologies, and so on.

There is a challenge to find similarities and differences

between the ontologies in automatic or semi-automatic way.

Differences could be as simple as the use of synonyms for the

same concept.

VI. RESEARCH ISSUES

Ontology Integration application provides a partially

automated solution to a specific aspect of ontology integration

within their chosen implementation language. Automation on

the system and syntactic level is relatively straightforward

and achieved. No such tools are currently available for

rendering ontology from web [19].

Ontology Matching is a challenging problem in many

applications, and is a major issue for interoperability in

information systems. It is major issue of finding semantic

correspondences between a pair of input ontologies.

These existing approaches are not performing ontology

evaluation and human intervention is there.

VII. ONTOLOGY ALIGNMENT EVALUATION

INITIATIVE

Ontology Alignment Evaluation Initiative [20] is popular

organization to contact the ontology engineering events. The

increasing number of methods available for schema

matching/ontology integration suggests the need to establish a

consensus for evaluation of these methods. The Ontology

Alignment Evaluation Initiative is a coordinated international

initiative to forge this consensus.

The goals of the Ontology Alignment Evaluation Initiative

are:

Assessing strength and weakness of alignment/matching

systems;

Comparing performance of techniques;

Increase communication among algorithm developers;

Improve evaluation techniques;

Most of all, helping improving the work on ontology

alignment/matching.

The organization of a yearly evaluation event.

The publication of the tests and results of the event for

further analysis.

The names "Ontology Alignment Evaluation Initiative",

"OAEI", or the name of the tool and/or method that was used

to generate the results must not be used to endorse or promote

products derived from the use of these data without prior

permission of their respective owners.

VIII. CONCLUSION

The Semantic Web is a promising research topic that will

allow the construction of intelligent applications capable of

understanding the contents on the web. The power of such

applications will rely on metadata expressed as ontologies.

This ontology integration helps to reduce end-user effort

and produce good quality ontology. The ontology integration

is an effective approach rather than building a new one.

Future works of ontology integration may include

developing a prototype for efficient and effective ontology

integration with less human intervention, less conflicts and

less suggestions and try to contribute in automating ontology

integration.

REFERENCES

[1] John Davies, Dieter Fensel , Frank Van Harmelen , Towards The Semantic

Web , John Wiley & Sons LTD , 2003

[2] Gruber, T. A Translation Approach to Portable Ontology Specification,

Knowledge Acquisition, 5(2): 199-220,1993.

[3] T. Berners-Lee, J.Hendler, and O. Lassila, “Semantic Web,” Scientific

American, May 2001.

[4] Dr.S.A.M.Rizvi, Nadia Imdadi, Jamia Millia, Ontology Integration

Paradigms for Automatic Semantic Integration Incorporating Semantic

Repositories, Proceedings of second National conference on challenges

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

89

An Efficient Multistage Motion Estimation Scheme for

Video Compression Using Content Based Adaptive

Search Technique First Author Second Author

P.DINESH KHANNA T.SARATH BABU P.G. Student (C&SP) Associate Professor

ECE Dept., G.Pulla Reddy Engineering College, ECE Dept., G.Pulla Reddy Engineering College ,

Kurnool (dt), Andhra Pradesh. Kurnool (dt), Andhra Pradesh. [email protected] [email protected]

Abstract—This paper describes about a multistage motion

estimation scheme that encompasses technique such as motion

vector prediction, determination of search range and search

patterns, and identification of termination criteria. Each of

these techniques has several diversions that may suit a

particular set of video characteristics. The proposed algorithm,

called content adaptive search technique (CAST), is a multistage

approach that first performs an online motion analysis. Motion

characteristics are extracted and passed on to the next stage that

includes a mathematical model of block distortion surface to

estimate the distance from the current search point to the global

optimal position. The MV search is performed on a precise search

area adaptive to the statistical properties of the MV prediction

error, separately for foreground and background regions. The

proposed algorithm, including the pre-stage, is fast and, therefore, is

suitable for online and real-time encoding. Due to its self-tuning

property, not only does the proposed scheme adapt to scenes

by yielding better visual quality but it also yields a lower

computational complexity, compared with the other predictive

motion estimation algorithms.

Keywords-video compression, motion vector prediction, adaptive,

block matching motion estimation.

I. INTRODUCTION

Video Compression makes it possible to use, transmit, or

manipulate videos easier and faster. Many applications

benefits from video compression. Compression is the process of converting data to a format that requires fewer bits. If there

exists a loss of information along this process, the

compression is lossy. Lossless compression is a manner in

which no data is lost. The compression ratio obtained is

usually not sufficient and lossy compression is required.

Video encoder includes a number of components, out of

which motion estimation (ME) carries a greater significance

because it can consume more than 80% of the overall

processing power and also because it has a high impact on video

quality. The goal of an ME algorithm is to find the most accurate

motion vectors (MVs) with the minimum processing cost. This paper proposes an ME algorithm that performs

self-adjustment to suit the video sequence in order to achieve the best possible video quality while minimizing the complexity. we introduce a ME scheme for a new paradigm of video compression that we call content adaptive video compression. The notion of content adaptive is that the

encoder is aware of the content (e.g., objectivity and scene complexity) that it is encoding and the context in which it is being used (e.g., a certain performance measure). The aim is to provide adaptability and self-adjustment to the environment that the compression is being used for. The proposed algorithm, called content adaptive search technique (CAST), is a multistage approach that first performs an on-line motion analysis. Motion characteristics are extracted and passed on to the next stage that includes a mathematical model of block distortion surface to estimate the distance from the cur-rent search point to the global optimal position. The MV search is performed on a precise search area adaptive to the statistical properties of the MV prediction error, separately for foreground and background regions. Fast ME algorithms have been proposed as alterna-

tives to the exhaustive full search technique. Some of these al-

gorithms have been widely utilized in video coding applica-

tions. Examples include the three step search (TSS) [4], the new

three step search (NTSS) [5], the four step search (4SS) [9],

the diamond search (DS) [13], the gradient search [6], [7], hi-

erarchical search [10], and the cross search (CS) [2]. Due to

their uniform search strategy without differentiating the types of

motion, these algorithms perform well only for certain video

sequences. A representative algorithm is the motion vector field

adaptive search technique (MVFAST) [3], which has been

adopted by MPEG-4 Part 2 as the core technology for fast ME

algorithm. MVFAST selects the MVs from the region of

support (ROS) as the predictors. It classifies motion types based

on local motion activity. The predictor that yields the

minimum distortion is selected as the initial search center for

high motion activity MB. MVFAST uses different search

patterns (small diamond and large diamond) based on the local

motion activity. FAME [1] is another ME algorithm based on

MV prediction. In addition to the spatial predictors, it includes a temporal predictor obtained from motion inertia prediction.

FAME adaptively selects different motion search patters: large

diamond pattern and elastic diamond pattern are used for a

coarse search with a range, while small diamond pattern is

used for fine tuning an MV. However, the performance of

these algorithms highly depends on the characteristics of the

video contents. In contrast, the proposed algorithm aims to per-

form well for diverse types of videos by flexibly adjusting its

search strategies. The rest of this paper is organized as follows

90

Fig.1. Block diagram for the content adaptive search technique Section II describes the CAST algorithm. Section III presents

the evaluation results and performance comparisons with other

MV prediction ME algorithms. Section IV provides some

concluding remarks.

II.PROPOSED TECHNIQUE

MV prediction, determination of search range and

search patterns, and identification of termination criteria are

basic techniques used in ME. Each of these techniques has several diversions that are suitable for different video

characteristics. As opposed to pre-deciding these strategies, the

proposed MV algorithm analyzes and extracts the

characteristics from the video contents and adapts its

strategies to the video contents. Fig. 1 shows the block

diagram of the proposed scheme. The CAST algorithm

consists of three stages: MV field prediction, motion analysis

and motion estimation. They are described as follows:

Stage 1) Motion Vector Field (MVF) Prediction: This stage

predicts the MVF from the previous coded frame using the

proposed Weighted Mean Inertia prediction technique.

Stage 2) Motion Analysis: This stage clusters MBs into back-

ground and foreground regions based on the predicted MVF and

extract motion characteristics of each region. Parameters

indicating motion characteristics are delivered to the next

stage.

Stage 3) Motion Estimation: This stage performs motion es-

timation which is fine tuned by the regional motion

parameters. The subsequent subsections introduce the details

of these three stages.

A. Motion Vector Field Prediction

A MVF is a rich source for motion analysis. It can be

predicted from the coded MBs by exploiting the spatial and

temporal correlations between MVs. Several MV prediction

techniques have been proposed [3], [11], [12], [9]. However, due

to the dependency, it is difficult to construct the whole MVF

using the spatial prediction before the encoding of a frame.

Temporal prediction works well only for slow motion videos due to the small correlation in fast motion. To deal with these

problems, a few methods exploiting the motion inertia property

have been proposed to increase the temporal prediction

accuracy [1], [8]. Because a natural object has the inertia to stay

in the state of rest or uniform motion unless acted upon by an

external force, the underlying assumption of inertia prediction

is that the motion of an object tends to stay unchanged during a

short period between frames. The inertia prediction can be

formulated as follows:

arg min x v x y v yc rx r c ry rinertiar R

Where (xc, yc) is the coordinate of the current MB, R is the set of MBs in the reference frame, (xr, yr) is the coordinate of a

MB r , rR, vrx, and vry are the MVs of MB r in horizontal

and vertical directions.

Although the inertia prediction increases temporal

prediction accuracy in fast motion, the MVF obtained by

inertia prediction is not smooth due to its block based

inherence. To solve this problem, we propose to average the

surrounding inertia MVs with weighting. This method, called

weighted mean inertia (WMI) prediction, can increase the

smoothness as well as the accuracy of the predicted MVF. The

WMI prediction is described in the following and illustrated in

Fig. 2.

Fig. 2. Example of WMI MV prediction.

Let MBi,t-1 denote MB i in frame t -1, as shown in white

blocks. Denote the MV of MBi ,t-1as MV i,t-1.Due to the mo-

tion inertia, MB i ,t-i intends to maintain its motion and

move to Pi,t in frame t , as shown in gray blocks.

TABLE I

CORRELATION BETWEEN THE PREDICTED MVF AND THE TRUE MVF

The prediction accuracy can be evaluated by the correlation

coefficients between the predicted MVF and the true MVF

obtained by exhaustive search. Table I shows the correlation

coefficients comparison between WMI and the inertia method

proposed in [1]. The experimental results verify that WMI has

higher prediction accuracy.

91

B. Motion Analysis

In this stage, the predicted MVF is decomposed into

foreground and background MVFs. Based on this

decomposition, a frame is segmented into different regions. For

each type of region, motion parameters characterizing the

regional motion are computed. 1) Motion Based Region Segmentation:

Motion in video contents usually consists of foreground

motion and background motion. Generally, background motion

is dominated by camera motion, which may be still, panning,

tilting, zooming or their combinations. Motion of foreground

MBs is typically homogeneous because foreground objects are

often correlated.

Motions in a local region generally exhibit similar

properties. The second stage of the CAST algorithm clusters

the MBs into three regions based on the MVF predicted from

the reference frame. Fig. 3 illustrates three regions: foreground,

background, and uncovered background. The uncovered background is a region initially covered by an object in the

reference frame but uncovered in the current frame due to the

object movement.

Fig. 3. Demonstration of foreground, background, and uncovered

background regions.

The foreground and background region can be

segmented as follows:

1) reconstruct the background MVF

1

1

u

v

, with the estimated

affine parameters '

1a to '

6a ;

2) compute the error (E) between the predicted MVF and the

reconstructed MVF ' '

, , , , ,x y x y x y x y x yE u u v v (2)

3) threshold T is set to be

,

0 ,0

1x y

x N y m

T E EN

(3)

Where E is the average of E in a frame;

4) for each MB, if the sum of its overlaps is less than 128, we mark it as uncovered background. Otherwise, its type is

determined as follows:

foreground ,MBtype=

background ,

E x y T

E x y T

(4)

An example of the motion-based region segmentation is

presented in Fig. 4. Fig. 4(a) is the original MVF of a frame

in "Foreman." Fig. 4(b) and (c) are the background MVF

and foreground MVF obtained with the above approach. The

background MVF consists of affine motion. The foreground

MVF captures the object motions.

Fig. 4.(a) Original MVF.(b) Background MVF.(c) Foreground MVF.

2) Motion Parameters: We define two parameters to model

the regional motion characteristics as follows.

Motion Velocity: This is the average length of MVs in a

region. MBvel and MFvel denote the background motion

velocity and foreground motion velocity, respectively.

Motion Complexity: This is the standard derivation of MVs in a region, which indicates the degree of the motion disorder. The

motion complexity of background and foreground are denoted

by MBcomp and MFcomp. R e s p e c t i v e l y

C. Motion Estimation

The motion parameters charactering the regional

motions are employed to optimize the search technique. In this

92

section, we present a search technique that adapts to the

motion parameters obtained through the first two stages. The

CAST technique encompasses several techniques, which are

described in the following.

Fig. 5. Average block distortion surface of test sequence Stefan.

1) Modeling the Block Distortion Surface: Block distortion

surface (BDS) is a 2-D array of distortion values at all points in

the search window. Although almost all fast ME algorithms are

based on the assumption that the BDS is unimodal, no mathe-

matical model for the BDS has been reported. Here, we present

a model of block distortion that is a function of distance, mo-

tion and texture. Based on this function, we estimate the dis-

tance from the current search point to the global minimum point

(GMP) given the motion parameters, texture complexity and the

block distortion at the search point. We start by considering a

BDS. Fig. 5 illustrates the average BDS of the test sequence

Stefan. Note that we center the (x, y) plane at the GMP. The

block distortion increases as the search position moving

further to the GMP.

To simplify the problem, we transform this 2-D

function to 1-D. r is defined as the chess-board distance of

vector (x, y)

r x y (5) D(r) denotes the block distortion of the search point having a

distance r to the GMP. D(O) is the global minimum distortion

value. Generally, the distortion in a block depends on the

complexity of the texture. It is because when a block contains

complex textures, it is less likely to find a matching block with

similar texture structure. Therefore, the value of D(O) will be

large if the texture is complex, and vice versa. We use D(O) to

represent the texture complexity.

By exploring the relation between D(r) and , we

observe that is ((D(r) - D(0)/D(0)))2 closely linear related to

the distance , as shown in Fig. 6. The solid lines in Fig. 6 are the

fitting curves of the measured data. 2.

D r Dg r

D

(6)

Where g is a factor.

2) Tight Search Area: To avoid checking the unnecessary

area, a tight search area derived from the statistics of MV

Fig. 6. Linear relation between ((D(r)-D(0)/D(0)))2 and r

prediction errors are proposed. It is well known that the

distribution of the MV prediction error is highly concentrated.

Here, we define the MV prediction error as the distance

between the predicted MV and the true MV. Shown in Fig. 7 is

the MV prediction error distribution using MV prediction

technique proposed in [3], that is, using best MV of the left,

upper, upper-right coded MB in the current frame and the

colocated MB in the reference frame.

Fig. 7. MV prediction error distribution of “Stefan” using MV

prediction technique in [3].

TABLE II

TIGHT SEARCH AREA FOR VARIOUS MOTION COMPLEXITY LEVELS

Table II includes the search area radius for various

motion complexity levels. Note that for high complexity

motion, we do not restrict the search area in order to increase

search accuracy. 3) Initial Search Point: The intention is to select an initial

search point as close to the GMP as possible. Therefore, a set of MV predictors (denoted by PSET) is evaluated, and the one

producing the minimum SAD is set to be the initial MV. Often,

increasing the number of MV predictors obtains more accurate

initial MVs and accordingly reduces the complexity in the sub-

sequent search. However, in a chaotic MVF where the predicted

MVs are disordered and inaccurate, increasing the number of

predictors may not help. Moreover, using more MV predictors

creases the computational cost in the initial step. Thus, the

choice of the MV predictors is critical.

93

TABLE III PERFORMANCE COMPARISON (RESOLUTION 176 x 144)

TABLE IV PERFORMANCE COMPARISON (RESOLUTION 720 x 480)

4) Search Strategy: We propose an exponentially expanding search process starting from the initial search point and restricted

by the search area. Two search modes are used in the search

process: cross search mode and exponential search mode.

The search process has the following steps.

Step 1) Let the current search point P be the initial search

point. If SAD(P) < T , stop search.

Step 2) Switch to cross search mode. Check the cross pattern

centered at P. Mark the point with minimum

SAD as the current best point (CBP). If P=CBP go

to Step 3). Otherwise, go to Step 4).

Step 3) If s = 1 stop the search process. Otherwise, set s =

1 and go to Step 2).

Step 4) Switch to exponential mode. Compute the expo-

nential step size V = CBP — P = (dx,dy),

where dx and dy are the displacement in horizontal and

vertical direction. Compute expanded search

ESP=P+2V=P+2 x (dx,dy).If

SAD(ESP)<SAD(CBP),mark ESP as the

new CBP and repeat Step 4).Otherwise, mark

the CBP as new P and go to Step 2).

T is an early stop threshold, which is defined as follows.

1) If all the MV predictors in PSET are identical, T is

set to the maximum SAD of the spatial adjacent

MBs.

2) Otherwise, T is set to the minimum SAD of the

spatial adjacent MBs.

III. PERFORMANCE EVALUATION

A. Complexity Analysis: The computational complexity of the

motion analysis is dominated by the computation of the affine

model parameter and motion segmentation. The complexity of

94

obtaining affine parameters and motion segmentation is

negligible. The SAD is dominant computation in the ME

process.

B. Experimental Results

This section includes the performance of CAST as

well as a comparison with MVFAST and FAME. We used the

Microsoft MPEG-4 Verification Model software encoder in the

simulation with 17 most popular test sequences containing low

and fast motion. We encoded 150 frames of each sequence with

constant bit rate (384 kbps for the 176 x 144 format and 1

Mbps for 720 x 480 format).

Tables III -IV show the objective quality and

complexity comparisons among MVFAST, FAME, and CAST.

The objective quality is measured by the peak signal-to-noise

ratio (PSNR), which is commonly used in the objective quality

comparison. Since the SAD operation dominates the

complexity of the ME algorithm, we can measure the

algorithm complexity by the number of SAD operations,

which is identical to the number of search points (NSP). This

metric is also commonly used for fair inter-platform

comparison.

The speedup column in these tables shows the

reduction in complexity as compared to MVFAST. It can be

observed that CAST yields the lowest complexity among the

algorithms being compared. CAST exhibits considerable

speedup for sequences with ordered background motion field,

irrespective of whether the background motion is still, slow or

fast. For example, for the sequence containing still or slow

background motion, such as News, Hall monitor and Container,

CAST achieved speedups of 2.86,4.31, and 12.68 in CIF format

and 7.22, 3.49, and 11.79 in QCIF format. An advantage of CAST compared to other ME

algorithms is that it drastically reduces the complexity in the

background. When the background motion is very slow and

smooth, the algorithm terminates without SAD computation

and the WMI predicted MV is the final MV. We observe that

"News" sequence shows a large difference in speedup in

different resolutions (2.86 at 176 x 144). In higher resolution,

MVBevel usually is larger. Thus, fewer MBs satisfy this

condition in high resolution than in low resolution, especially

for the sequences where most of the MBs are located in slow

motion background, such as "News." Further improvement can

be obtained by fine tuning the early termination threshold for

different resolutions.

The average complexity comparison is shown in

Table V. CAST achieved 3.83, 3.05, and 2.44 times speedup

compared to the MVFAST for the 176 x 144 and 720 x 480 for-

mats, respectively. On the other hand, FAME gained only 1.44,

1.44, and 1.37 speedups for the same test. CAST outperformed

both MVFAST and FAME.

TABLE V

AVERAGE SPEED UP OF PSNR GAINS

IV CONCLUSION

This paper presented a multistage algorithm for

block matching motion estimation, encompassing a motion

analysis stage to assist the motion vector search. The

analysis stage helps the search technique to adapt to the

motion characteristics. The experimental results show that the proposed algorithm outscores the other predictive ME

algorithms in terms of computational cost and visual quality,

while showing the adaptability to various types of scenes. The

proposed scheme has the best overall performance among the

compared algorithms. Motion characteristics and their

utilization on motion estimation can be further studied.

REFERENCES

[1] I. Ahmad, W. Zheng, J. Luo, and M. Liu, "A fast adaptive motion estimation algorithm," IEEE Trans. Circuits Syst. Video Technol., vol. 16, no. 3, pp. 439-446, Mar. 2006. [2] M. Ghanbari, "The cross-search algorithm for motion estimation," IEEE Trans. Commun., vol. 38, no. 7, pp. 950-953, Jul. 1990. [3] P. I. Hosur and K. K. Ma, "Motion vector field adaptive fast motion estimation," presented at the 2nd Int. Conf. Info., Commun. Signal Process. (ICICS), Singapore, Dec. 1999. [4] J. R. Jain and A. K. Jain, "Displacement measurement and its application in interframe image coding," IEEE Trans. Commun., vol. 29, no. 12, pp. 1799-1808, Dec. 1981. [5] R. Li, B. Zeng, and M. Liou, "A new three-step search algorithm for block motion estimation," IEEE Trans. Circuits Syst. Video Technol., vol. 4, no. 4, pp. 438-442, Aug. 1994. [6] B. Liu and A. Zaccartin, "New fast algorithms for estimation of block motion vectors," IEEE Trans. Circuits Syst. Video Technol., vol. 3, no2,pp. 148-157, Apr. 1993. [7] L. K. Liu and E. Feig, "A block-based gradient descent search algorithm for block motion estimation in video coding," IEEE Trans. Circuits Syst. Video Technol., vol. 6, no. 4, pp. 419^22, Aug. 1996. [8] T. Liu, K.-T. Lo, J. Feng, and X. Zhang, "Frame interpolation scheme using inertia motion prediction," Signal Process.: Image Commun., vol. 18, pp. 221-229, 2003. [9] J. Luo, I. Ahmad, Y. Liang, and Y. Sun, "Motion estimation for content adaptive video compression," in Proc. ICME, Taiwan, Jun. 2004, pp. 1427-1430. [10] X. Song, T. Chiang, and Y.-Q. Zhang, "A scalable hierarchical motion estimation algorithm for MPEG-2," in Proc. IEEE Int. Conf. Image Process. (ICIP), 1998, pp. IV126-IV129. [11] A. M. Tourapis, O. C. Au, and M. L. Liou, "Fast block-matching motion estimation using predictive motion vector field adaptive search technique (PMVFAST)," presented at the ISO/IEC JTC1/SC29/WG11 MPEG99/m5866, Noordwijkerhout, The Netherlands, Mar. 2000. [12] A. Tourapis, G. Shen, M. Liou, O. Au, and I. Ahmad, "New predictive diamond search algorithm for block-based motion estimation," in Proc. VCIP, 2000, pp. 1365-1373. [13] S. Zhu and K. K. Ma, "A new diamond search algorithm for fast block-matching motion estimation," IEEE Trans. Image Process., vol. 9, no. 2, pp. 287-290, Feb. 2000.

SPEED UP PSNR GAIN

176X144 720X480 176X144 720X480

MVFAST 1.00 1.00 0 0

FAME 1.48 1.37 -0.010 +0.070

CAST 4.27 2.23 +0.015 +0.046

88

& opportunities in information technology(COIT-2008),

MandiGobindgarh, March 29, 2008.

[5] Alasoud, A., Haarslev, V., And Shiri, N. A Hybrid Approach for Ontology

Integration. In VLDB Workshop on Ontologies-based techniques for Data-

Bases and Information Systems (ODBIS) Trondheim, Norway, 2005.

[6] The HCONE Approach to Ontology Merging Konstantinos Kotis1, George

A. Vouros1, In Proceedings of the 21st International Symposium on

Computer Architecture (ESWS), Vol. 3053,Springer (2004), p. 137-151.

[7] Mitra, P. and Wiederhold, G. (2001) “An Algebra for Semantic

Interoperability of Information Sources,” 2nd. IEEE Symposium on

BioInformatics and Bioengineering (BIBE), Bethesda, MD, IEEE

Computer Society Press, 174-182.

[8] J. Gracia and E. Mena. Ontology Matching with CIDER: Evaluation

report for the OAEI 2008. In Proc. of 3rd Ontology Matching Workshop

(OM'08) at ISWC'08, Karlsruhe, Germany, October 2008

[9] M. Ehrig, S. Staab. QOM - Quick Ontology Mapping. In: Proceedings of

the 3rd International Semantic Web Conference (ISWC2004), November,

2004, Hiroshima, Japan. LNCS, Springer, 2004

[10] Pinto, H. Some issues on ontology integration. In: Proceedings of the

IJCAI-99 Workshop on Ontologies and Problem-Solving methods

(KRR5), Stockholm, Sweden, pp. 7-1 – 7-12, 1999.

[11] Fensel, D., Harmelen, F., Ding, Y., Klein, M., Akkermans, H., Broekstra,

J., Kampman, A., Meer, J., Sure, Y., Studer, R., Krohn, U., Davies, J.,

Engels, R., Iosif, V., Kiryakov, A., Lau, T., Reimer, U., and Horrocks, I.

On-To-Knowledge in a Nutshell. Special Issue of IEEE Computer on Web

Intelligence (WI). 2002

[12] Gangemi, A., Fisseha, F., Pettman, I., Pisanelli, D., Taconet, M.,Keizer, J.

A Formal Ontological Framework for Semantic Interoperability in the

Fishery Domain. Proceedings of the ECAI-02 Workshop on Ontologies

and Semantic Interoperability, pp. 16-30, Lyon, France. 2002.

[13] Heflin, J. and Hendler, J. Dynamic ontologies on the Web. Proceedings of

17th National Conference on Artificial Intelligence (AAAI-2000).

[14] Stumme, G. and A. Maedche, FCA-Merge: Bottom-up Merging of

Ontologies, In seventh international Conference on Artificial Intelligence

(IJCAI’01), seattle, WA, USA, 200, pp: 225-230

[15] Ahmed Khalifa Alasoud, A Multi-Matching Technique for Combining

Similarity Measures in Ontology Integration, Department of Computer

Science and Software Engineering, Concordia University,

Montréal,Canada,February 2009.

[16] Silva, N. and Rocha, J. MAFRA - An Ontology MApping FRAmework for

the Semantic Web, Proceedings of the 6th International Conference on

Business Information Systems, Colorado Springs (CO), USA, 2003.

[17] Noy N.F. and M.A. Muser, The PROMPT suite: Interactive tools for

Ontology merging and mapping, Int. J. Human Comput. Stud, 59: 983–

1024,2003

[18] D.McGuinnes, R.Fikes,J. Rice, and S.Wilder, An Environment for

Merging and Testing Large Ontologies, In Proceedings of the 17th

International conference on Principles of Knowledge Representation

and Reasoning (KR-2000),USA,April 2000.

[19] H. Alani. Position paper: Ontology construction from online ontologies. In

Proc. of 15th International World Wide Web Conference (WWW2006),

Edinburgh, UK, May 2006.

[20] http://oaei.ontologymatching.org/, Ontology Alignment Evaluation

Initiative.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

89

An Efficient Multistage Motion Estimation Scheme for

Video Compression Using Content Based Adaptive

Search Technique First Author Second Author

P.DINESH KHANNA T.SARATH BABU P.G. Student (C&SP) Associate Professor

ECE Dept., G.Pulla Reddy Engineering College, ECE Dept., G.Pulla Reddy Engineering College ,

Kurnool (dt), Andhra Pradesh. Kurnool (dt), Andhra Pradesh. [email protected] [email protected]

Abstract—This paper describes about a multistage motion

estimation scheme that encompasses technique such as motion

vector prediction, determination of search range and search

patterns, and identification of termination criteria. Each of

these techniques has several diversions that may suit a

particular set of video characteristics. The proposed algorithm,

called content adaptive search technique (CAST), is a multistage

approach that first performs an online motion analysis. Motion

characteristics are extracted and passed on to the next stage that

includes a mathematical model of block distortion surface to

estimate the distance from the current search point to the global

optimal position. The MV search is performed on a precise search

area adaptive to the statistical properties of the MV prediction

error, separately for foreground and background regions. The

proposed algorithm, including the pre-stage, is fast and, therefore, is

suitable for online and real-time encoding. Due to its self-tuning

property, not only does the proposed scheme adapt to scenes

by yielding better visual quality but it also yields a lower

computational complexity, compared with the other predictive

motion estimation algorithms.

Keywords-video compression, motion vector prediction, adaptive,

block matching motion estimation.

I. INTRODUCTION

Video Compression makes it possible to use, transmit, or

manipulate videos easier and faster. Many applications

benefits from video compression. Compression is the process of converting data to a format that requires fewer bits. If there

exists a loss of information along this process, the

compression is lossy. Lossless compression is a manner in

which no data is lost. The compression ratio obtained is

usually not sufficient and lossy compression is required.

Video encoder includes a number of components, out of

which motion estimation (ME) carries a greater significance

because it can consume more than 80% of the overall

processing power and also because it has a high impact on video

quality. The goal of an ME algorithm is to find the most accurate

motion vectors (MVs) with the minimum processing cost. This paper proposes an ME algorithm that performs

self-adjustment to suit the video sequence in order to achieve the best possible video quality while minimizing the complexity. we introduce a ME scheme for a new paradigm of video compression that we call content adaptive video compression. The notion of content adaptive is that the

encoder is aware of the content (e.g., objectivity and scene complexity) that it is encoding and the context in which it is being used (e.g., a certain performance measure). The aim is to provide adaptability and self-adjustment to the environment that the compression is being used for. The proposed algorithm, called content adaptive search technique (CAST), is a multistage approach that first performs an on-line motion analysis. Motion characteristics are extracted and passed on to the next stage that includes a mathematical model of block distortion surface to estimate the distance from the cur-rent search point to the global optimal position. The MV search is performed on a precise search area adaptive to the statistical properties of the MV prediction error, separately for foreground and background regions. Fast ME algorithms have been proposed as alterna-

tives to the exhaustive full search technique. Some of these al-

gorithms have been widely utilized in video coding applica-

tions. Examples include the three step search (TSS) [4], the new

three step search (NTSS) [5], the four step search (4SS) [9],

the diamond search (DS) [13], the gradient search [6], [7], hi-

erarchical search [10], and the cross search (CS) [2]. Due to

their uniform search strategy without differentiating the types of

motion, these algorithms perform well only for certain video

sequences. A representative algorithm is the motion vector field

adaptive search technique (MVFAST) [3], which has been

adopted by MPEG-4 Part 2 as the core technology for fast ME

algorithm. MVFAST selects the MVs from the region of

support (ROS) as the predictors. It classifies motion types based

on local motion activity. The predictor that yields the

minimum distortion is selected as the initial search center for

high motion activity MB. MVFAST uses different search

patterns (small diamond and large diamond) based on the local

motion activity. FAME [1] is another ME algorithm based on

MV prediction. In addition to the spatial predictors, it includes a temporal predictor obtained from motion inertia prediction.

FAME adaptively selects different motion search patters: large

diamond pattern and elastic diamond pattern are used for a

coarse search with a range, while small diamond pattern is

used for fine tuning an MV. However, the performance of

these algorithms highly depends on the characteristics of the

video contents. In contrast, the proposed algorithm aims to per-

form well for diverse types of videos by flexibly adjusting its

search strategies. The rest of this paper is organized as follows

90

Fig.1. Block diagram for the content adaptive search technique Section II describes the CAST algorithm. Section III presents

the evaluation results and performance comparisons with other

MV prediction ME algorithms. Section IV provides some

concluding remarks.

II.PROPOSED TECHNIQUE

MV prediction, determination of search range and

search patterns, and identification of termination criteria are

basic techniques used in ME. Each of these techniques has several diversions that are suitable for different video

characteristics. As opposed to pre-deciding these strategies, the

proposed MV algorithm analyzes and extracts the

characteristics from the video contents and adapts its

strategies to the video contents. Fig. 1 shows the block

diagram of the proposed scheme. The CAST algorithm

consists of three stages: MV field prediction, motion analysis

and motion estimation. They are described as follows:

Stage 1) Motion Vector Field (MVF) Prediction: This stage

predicts the MVF from the previous coded frame using the

proposed Weighted Mean Inertia prediction technique.

Stage 2) Motion Analysis: This stage clusters MBs into back-

ground and foreground regions based on the predicted MVF and

extract motion characteristics of each region. Parameters

indicating motion characteristics are delivered to the next

stage.

Stage 3) Motion Estimation: This stage performs motion es-

timation which is fine tuned by the regional motion

parameters. The subsequent subsections introduce the details

of these three stages.

A. Motion Vector Field Prediction

A MVF is a rich source for motion analysis. It can be

predicted from the coded MBs by exploiting the spatial and

temporal correlations between MVs. Several MV prediction

techniques have been proposed [3], [11], [12], [9]. However, due

to the dependency, it is difficult to construct the whole MVF

using the spatial prediction before the encoding of a frame.

Temporal prediction works well only for slow motion videos due to the small correlation in fast motion. To deal with these

problems, a few methods exploiting the motion inertia property

have been proposed to increase the temporal prediction

accuracy [1], [8]. Because a natural object has the inertia to stay

in the state of rest or uniform motion unless acted upon by an

external force, the underlying assumption of inertia prediction

is that the motion of an object tends to stay unchanged during a

short period between frames. The inertia prediction can be

formulated as follows:

arg min x v x y v yc rx r c ry rinertiar R

Where (xc, yc) is the coordinate of the current MB, R is the set of MBs in the reference frame, (xr, yr) is the coordinate of a

MB r , rR, vrx, and vry are the MVs of MB r in horizontal

and vertical directions.

Although the inertia prediction increases temporal

prediction accuracy in fast motion, the MVF obtained by

inertia prediction is not smooth due to its block based

inherence. To solve this problem, we propose to average the

surrounding inertia MVs with weighting. This method, called

weighted mean inertia (WMI) prediction, can increase the

smoothness as well as the accuracy of the predicted MVF. The

WMI prediction is described in the following and illustrated in

Fig. 2.

Fig. 2. Example of WMI MV prediction.

Let MBi,t-1 denote MB i in frame t -1, as shown in white

blocks. Denote the MV of MBi ,t-1as MV i,t-1.Due to the mo-

tion inertia, MB i ,t-i intends to maintain its motion and

move to Pi,t in frame t , as shown in gray blocks.

TABLE I

CORRELATION BETWEEN THE PREDICTED MVF AND THE TRUE MVF

The prediction accuracy can be evaluated by the correlation

coefficients between the predicted MVF and the true MVF

obtained by exhaustive search. Table I shows the correlation

coefficients comparison between WMI and the inertia method

proposed in [1]. The experimental results verify that WMI has

higher prediction accuracy.

91

B. Motion Analysis

In this stage, the predicted MVF is decomposed into

foreground and background MVFs. Based on this

decomposition, a frame is segmented into different regions. For

each type of region, motion parameters characterizing the

regional motion are computed. 1) Motion Based Region Segmentation:

Motion in video contents usually consists of foreground

motion and background motion. Generally, background motion

is dominated by camera motion, which may be still, panning,

tilting, zooming or their combinations. Motion of foreground

MBs is typically homogeneous because foreground objects are

often correlated.

Motions in a local region generally exhibit similar

properties. The second stage of the CAST algorithm clusters

the MBs into three regions based on the MVF predicted from

the reference frame. Fig. 3 illustrates three regions: foreground,

background, and uncovered background. The uncovered background is a region initially covered by an object in the

reference frame but uncovered in the current frame due to the

object movement.

Fig. 3. Demonstration of foreground, background, and uncovered

background regions.

The foreground and background region can be

segmented as follows:

1) reconstruct the background MVF

1

1

u

v

, with the estimated

affine parameters '

1a to '

6a ;

2) compute the error (E) between the predicted MVF and the

reconstructed MVF ' '

, , , , ,x y x y x y x y x yE u u v v (2)

3) threshold T is set to be

,

0 ,0

1x y

x N y m

T E EN

(3)

Where E is the average of E in a frame;

4) for each MB, if the sum of its overlaps is less than 128, we mark it as uncovered background. Otherwise, its type is

determined as follows:

foreground ,MBtype=

background ,

E x y T

E x y T

(4)

An example of the motion-based region segmentation is

presented in Fig. 4. Fig. 4(a) is the original MVF of a frame

in "Foreman." Fig. 4(b) and (c) are the background MVF

and foreground MVF obtained with the above approach. The

background MVF consists of affine motion. The foreground

MVF captures the object motions.

Fig. 4.(a) Original MVF.(b) Background MVF.(c) Foreground MVF.

2) Motion Parameters: We define two parameters to model

the regional motion characteristics as follows.

Motion Velocity: This is the average length of MVs in a

region. MBvel and MFvel denote the background motion

velocity and foreground motion velocity, respectively.

Motion Complexity: This is the standard derivation of MVs in a region, which indicates the degree of the motion disorder. The

motion complexity of background and foreground are denoted

by MBcomp and MFcomp. R e s p e c t i v e l y

C. Motion Estimation

The motion parameters charactering the regional

motions are employed to optimize the search technique. In this

92

section, we present a search technique that adapts to the

motion parameters obtained through the first two stages. The

CAST technique encompasses several techniques, which are

described in the following.

Fig. 5. Average block distortion surface of test sequence Stefan.

1) Modeling the Block Distortion Surface: Block distortion

surface (BDS) is a 2-D array of distortion values at all points in

the search window. Although almost all fast ME algorithms are

based on the assumption that the BDS is unimodal, no mathe-

matical model for the BDS has been reported. Here, we present

a model of block distortion that is a function of distance, mo-

tion and texture. Based on this function, we estimate the dis-

tance from the current search point to the global minimum point

(GMP) given the motion parameters, texture complexity and the

block distortion at the search point. We start by considering a

BDS. Fig. 5 illustrates the average BDS of the test sequence

Stefan. Note that we center the (x, y) plane at the GMP. The

block distortion increases as the search position moving

further to the GMP.

To simplify the problem, we transform this 2-D

function to 1-D. r is defined as the chess-board distance of

vector (x, y)

r x y (5) D(r) denotes the block distortion of the search point having a

distance r to the GMP. D(O) is the global minimum distortion

value. Generally, the distortion in a block depends on the

complexity of the texture. It is because when a block contains

complex textures, it is less likely to find a matching block with

similar texture structure. Therefore, the value of D(O) will be

large if the texture is complex, and vice versa. We use D(O) to

represent the texture complexity.

By exploring the relation between D(r) and , we

observe that is ((D(r) - D(0)/D(0)))2 closely linear related to

the distance , as shown in Fig. 6. The solid lines in Fig. 6 are the

fitting curves of the measured data. 2.

D r Dg r

D

(6)

Where g is a factor.

2) Tight Search Area: To avoid checking the unnecessary

area, a tight search area derived from the statistics of MV

Fig. 6. Linear relation between ((D(r)-D(0)/D(0)))2 and r

prediction errors are proposed. It is well known that the

distribution of the MV prediction error is highly concentrated.

Here, we define the MV prediction error as the distance

between the predicted MV and the true MV. Shown in Fig. 7 is

the MV prediction error distribution using MV prediction

technique proposed in [3], that is, using best MV of the left,

upper, upper-right coded MB in the current frame and the

colocated MB in the reference frame.

Fig. 7. MV prediction error distribution of “Stefan” using MV

prediction technique in [3].

TABLE II

TIGHT SEARCH AREA FOR VARIOUS MOTION COMPLEXITY LEVELS

Table II includes the search area radius for various

motion complexity levels. Note that for high complexity

motion, we do not restrict the search area in order to increase

search accuracy. 3) Initial Search Point: The intention is to select an initial

search point as close to the GMP as possible. Therefore, a set of MV predictors (denoted by PSET) is evaluated, and the one

producing the minimum SAD is set to be the initial MV. Often,

increasing the number of MV predictors obtains more accurate

initial MVs and accordingly reduces the complexity in the sub-

sequent search. However, in a chaotic MVF where the predicted

MVs are disordered and inaccurate, increasing the number of

predictors may not help. Moreover, using more MV predictors

creases the computational cost in the initial step. Thus, the

choice of the MV predictors is critical.

93

TABLE III PERFORMANCE COMPARISON (RESOLUTION 176 x 144)

TABLE IV PERFORMANCE COMPARISON (RESOLUTION 720 x 480)

4) Search Strategy: We propose an exponentially expanding search process starting from the initial search point and restricted

by the search area. Two search modes are used in the search

process: cross search mode and exponential search mode.

The search process has the following steps.

Step 1) Let the current search point P be the initial search

point. If SAD(P) < T , stop search.

Step 2) Switch to cross search mode. Check the cross pattern

centered at P. Mark the point with minimum

SAD as the current best point (CBP). If P=CBP go

to Step 3). Otherwise, go to Step 4).

Step 3) If s = 1 stop the search process. Otherwise, set s =

1 and go to Step 2).

Step 4) Switch to exponential mode. Compute the expo-

nential step size V = CBP — P = (dx,dy),

where dx and dy are the displacement in horizontal and

vertical direction. Compute expanded search

ESP=P+2V=P+2 x (dx,dy).If

SAD(ESP)<SAD(CBP),mark ESP as the

new CBP and repeat Step 4).Otherwise, mark

the CBP as new P and go to Step 2).

T is an early stop threshold, which is defined as follows.

1) If all the MV predictors in PSET are identical, T is

set to the maximum SAD of the spatial adjacent

MBs.

2) Otherwise, T is set to the minimum SAD of the

spatial adjacent MBs.

III. PERFORMANCE EVALUATION

A. Complexity Analysis: The computational complexity of the

motion analysis is dominated by the computation of the affine

model parameter and motion segmentation. The complexity of

94

obtaining affine parameters and motion segmentation is

negligible. The SAD is dominant computation in the ME

process.

B. Experimental Results

This section includes the performance of CAST as

well as a comparison with MVFAST and FAME. We used the

Microsoft MPEG-4 Verification Model software encoder in the

simulation with 17 most popular test sequences containing low

and fast motion. We encoded 150 frames of each sequence with

constant bit rate (384 kbps for the 176 x 144 format and 1

Mbps for 720 x 480 format).

Tables III -IV show the objective quality and

complexity comparisons among MVFAST, FAME, and CAST.

The objective quality is measured by the peak signal-to-noise

ratio (PSNR), which is commonly used in the objective quality

comparison. Since the SAD operation dominates the

complexity of the ME algorithm, we can measure the

algorithm complexity by the number of SAD operations,

which is identical to the number of search points (NSP). This

metric is also commonly used for fair inter-platform

comparison.

The speedup column in these tables shows the

reduction in complexity as compared to MVFAST. It can be

observed that CAST yields the lowest complexity among the

algorithms being compared. CAST exhibits considerable

speedup for sequences with ordered background motion field,

irrespective of whether the background motion is still, slow or

fast. For example, for the sequence containing still or slow

background motion, such as News, Hall monitor and Container,

CAST achieved speedups of 2.86,4.31, and 12.68 in CIF format

and 7.22, 3.49, and 11.79 in QCIF format. An advantage of CAST compared to other ME

algorithms is that it drastically reduces the complexity in the

background. When the background motion is very slow and

smooth, the algorithm terminates without SAD computation

and the WMI predicted MV is the final MV. We observe that

"News" sequence shows a large difference in speedup in

different resolutions (2.86 at 176 x 144). In higher resolution,

MVBevel usually is larger. Thus, fewer MBs satisfy this

condition in high resolution than in low resolution, especially

for the sequences where most of the MBs are located in slow

motion background, such as "News." Further improvement can

be obtained by fine tuning the early termination threshold for

different resolutions.

The average complexity comparison is shown in

Table V. CAST achieved 3.83, 3.05, and 2.44 times speedup

compared to the MVFAST for the 176 x 144 and 720 x 480 for-

mats, respectively. On the other hand, FAME gained only 1.44,

1.44, and 1.37 speedups for the same test. CAST outperformed

both MVFAST and FAME.

TABLE V

AVERAGE SPEED UP OF PSNR GAINS

IV CONCLUSION

This paper presented a multistage algorithm for

block matching motion estimation, encompassing a motion

analysis stage to assist the motion vector search. The

analysis stage helps the search technique to adapt to the

motion characteristics. The experimental results show that the proposed algorithm outscores the other predictive ME

algorithms in terms of computational cost and visual quality,

while showing the adaptability to various types of scenes. The

proposed scheme has the best overall performance among the

compared algorithms. Motion characteristics and their

utilization on motion estimation can be further studied.

REFERENCES

[1] I. Ahmad, W. Zheng, J. Luo, and M. Liu, "A fast adaptive motion estimation algorithm," IEEE Trans. Circuits Syst. Video Technol., vol. 16, no. 3, pp. 439-446, Mar. 2006. [2] M. Ghanbari, "The cross-search algorithm for motion estimation," IEEE Trans. Commun., vol. 38, no. 7, pp. 950-953, Jul. 1990. [3] P. I. Hosur and K. K. Ma, "Motion vector field adaptive fast motion estimation," presented at the 2nd Int. Conf. Info., Commun. Signal Process. (ICICS), Singapore, Dec. 1999. [4] J. R. Jain and A. K. Jain, "Displacement measurement and its application in interframe image coding," IEEE Trans. Commun., vol. 29, no. 12, pp. 1799-1808, Dec. 1981. [5] R. Li, B. Zeng, and M. Liou, "A new three-step search algorithm for block motion estimation," IEEE Trans. Circuits Syst. Video Technol., vol. 4, no. 4, pp. 438-442, Aug. 1994. [6] B. Liu and A. Zaccartin, "New fast algorithms for estimation of block motion vectors," IEEE Trans. Circuits Syst. Video Technol., vol. 3, no2,pp. 148-157, Apr. 1993. [7] L. K. Liu and E. Feig, "A block-based gradient descent search algorithm for block motion estimation in video coding," IEEE Trans. Circuits Syst. Video Technol., vol. 6, no. 4, pp. 419^22, Aug. 1996. [8] T. Liu, K.-T. Lo, J. Feng, and X. Zhang, "Frame interpolation scheme using inertia motion prediction," Signal Process.: Image Commun., vol. 18, pp. 221-229, 2003. [9] J. Luo, I. Ahmad, Y. Liang, and Y. Sun, "Motion estimation for content adaptive video compression," in Proc. ICME, Taiwan, Jun. 2004, pp. 1427-1430. [10] X. Song, T. Chiang, and Y.-Q. Zhang, "A scalable hierarchical motion estimation algorithm for MPEG-2," in Proc. IEEE Int. Conf. Image Process. (ICIP), 1998, pp. IV126-IV129. [11] A. M. Tourapis, O. C. Au, and M. L. Liou, "Fast block-matching motion estimation using predictive motion vector field adaptive search technique (PMVFAST)," presented at the ISO/IEC JTC1/SC29/WG11 MPEG99/m5866, Noordwijkerhout, The Netherlands, Mar. 2000. [12] A. Tourapis, G. Shen, M. Liou, O. Au, and I. Ahmad, "New predictive diamond search algorithm for block-based motion estimation," in Proc. VCIP, 2000, pp. 1365-1373. [13] S. Zhu and K. K. Ma, "A new diamond search algorithm for fast block-matching motion estimation," IEEE Trans. Image Process., vol. 9, no. 2, pp. 287-290, Feb. 2000.

SPEED UP PSNR GAIN

176X144 720X480 176X144 720X480

MVFAST 1.00 1.00 0 0

FAME 1.48 1.37 -0.010 +0.070

CAST 4.27 2.23 +0.015 +0.046

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

95

Framework for Adaptive Composition and

Provisioning of Web Services M.S.Thilagavathy

#1, R.SivaRaman

*2

# 1 Department of Computer Science and Engineering, Anna University Tiruchirappalli

Tiruchirappalli ,Tamil Nadu, India, 2Lecturer, Anna University , Tiruchirappalli [email protected]

Abstract— Web service composition allows developers

to create applications on top of service oriented computing’s

native paradigm of description, discovery, and communication

capabilities. Such applications are rapidly deployable and offer

developers reuse possibilities and provide users a seamless access

to a variety of complex services. This paper describes the design

of a framework for adaptive composition of web services. In the

composition model service context and exceptions are configured

to accommodate needs of different users. Services selection is

made dynamically identifying the best set of services available at

runtime. This allows for reusability of a service in different

contexts and achieves a level of adaptiveness and

contextualization without recoding and recompiling of the

overall composed services. The execution model is adaptive in the

sense that the runtime environment is able to detect exceptions

and react to them according to a set of rules. The single

centralized repository is being substituted by dedicated

repositories that cooperate and exchange information about

stored services on demand. Aspect oriented approach is used to

provide support for service adaptation. Three core services,

coordination service, context service, and event service, are

implemented to automatically schedule and execute the

component services, and adapt to user configured exceptions and

contexts at run time. The various parts of the framework are

explained with a common case study taken from E-learning

domain.

Keywords— Web service, service composition, service-oriented

architecture, exception handling, aspect-oriented programming,

event-based service execution.

I. INTRODUCTION

In service-oriented computing (SOC), developers use

services as fundamental elements in their application

development processes. Services are platform and network

independent operations that clients or other services invoke

[1]. Web services are typical example of SOC. Definition

published by the World Wide Web consortium W3C, in the

web services Architecture document states that a web service

is a software system identified by a URL, whose public

interfaces and bindings are defined and described using XML.

Its definition can be discovered by other software systems. These systems may then interact with the web service in a

manner prescribed by its definition, using XML-based

messages given by internet protocols.

Nowadays, an increasing amount of companies and

organizations only implement their core business and

outsource other application services over Internet. Service

composition is a process of combining existing services in

order to satisfy the functionality required by the user. Service

composition accelerates rapid application development,

service reuse, and complex service consummation. Every

composition approach must guarantee connectivity, non-

functional QoS properties, such as timeliness, security, and

dependability, correctness and scalability. To make service composition dynamic selection of Web services composition

focus on context aware business processes.

A current trend is to provide adaptive service

composition and provisioning solutions that offer better

quality of composite Web services[2],[3]. The pervasiveness

of the Internet and the proliferation of interconnected

computing devices (e.g., laptops, PDAs, 3G mobile phones)

offer the technical possibilities to interact with services

anytime and anywhere. Service adaptation refers to the

problem of modifying a service so that it can correctly interact

with another service, overcoming functional and non-functional mismatches and incompatibilities. While the

service functionality remains to a large extent the same, a

service needs to adapt to existing one. To simplify adaptation,

it is important to separate the adaptation logic from the

business logic. Such separation helps to avoid the need of

developing and maintaining several versions of a service

implementation and isolates the adaptation logic in a single

place [4].

Adaptive Composition of web services with

Distributed Registries provides a system infrastructure for

distributed, adaptive, context-aware provisioning of composite

Web services. This allows service designers to focus more on specifying service composition requirements at a high level of

abstraction such as business logic of applications, generic

exception handling policies, and contextual constraints, rather

than on low-level deployment and coordination concerns [5].

The salient features of our approach are:

Adaptable service composition model decouples the

contextual and exceptions specification from the business

logic of the actual service. The concept of process schema is

used for modeling the business logic of the composite service.

End users can then customize this process schema by

assigning users’ contextual constraints to the process schema. Generic exception behaviours (e.g., service failures, network

errors) that may happen during the service provisioning are

specified as policies that can be reused across different

process schemas. By adopting AOP approach for modeling

96

exceptions, end users can dynamically add, remove, and

modify tuples of aspects on how exceptions can be handled

without changing functionalities of composite service.

DIstributed REgistry (DIRE), which is devoted to service

publication and exploits a publish/subscribe middleware to

support the dynamic federation of heterogeneous registries

and the flexible distribution of service descriptions based on

explicit subscriptions.

An event-driven service execution model provides the

execution semantics of adaptive composite services. We

propose the use of an aspect-oriented programming (AOP) approach to weave adaptation solutions into the different

composite services that need to be modified. The execution

semantics is enforced by means of three core generic services:

coordination service, context service, and event service. These

basic infrastructure form the backbone of a middleware that

provides deployment and automatic execution of the adaptive

composite service in a robust and scalable manner.

II. RELATED WORK

Recently, automatic service composition approaches

typically exploit the Semantic Web and artificial intelligence

(AI) planning techniques. By giving a set of component services and a specified requirement (e.g., user’s request), a

composite service specification can be generated

automatically [4]. However, realizing a fully automatic

service composition is still very difficult and presents several

open issues [1], [6], [8]. The basic weakness of most research

efforts proposed so far is that Web services do not share a full

understanding of their semantics, which largely affects the

automatic selection of services. The work in [11] proposes a

trade-off between planning and optimization approaches. In a

first semiautomatic logical composition step, the goal is

translated into a workflow-based specification that introduces abstract tasks. A second physical composition step maps

abstract tasks to concrete services and is supervisioned by the

composed service designer.

Composite services orchestration is a very active area of

research and development. Our work’s distributed

orchestration model has some similarities with the work

presented in [7], which proposes a decentralized orchestration

approach. The approach, however, differs from our work, in

that it is only applicable when the assignment of activities to

their executing entities is known during the deployment of the

workflow, which is a restrictive assumption In the context of

service composition where providers can leave and join a community or alter the characteristics of their offers (e.g., the

QoS or the price) after the composite service has been defined

and deployed. Several techniques have been proposed to deal

with adaptability of composite Web services. In the effort

reported in [10], a platform has been developed where BPEL

processes can be extended with policies and constraints for

runtime configuration. The other two recent efforts for

adaptive Web services composition, reported in [2], focus on

dynamic service selection for context-aware business

processes.

III. ADAPTABLE COMPOSITION MODEL

The service composition encompasses roles and

functionality for aggregating multiple services into a single

composite service. Resulting composite services can be used

as basic services in further service compositions or offered as

complete applications and solutions to service clients. The

basic weakness of most of the service composition models

proposed so far is that Web Services do not share a full

understanding of their behaviour. Though some approaches to Web Service description consider behaviour, but it is not

publicly available. Therefore, it is not exploited in

composition, discovery, etc. All traditional service

composition model depends on process modeling notations to

choreograph the component services. This mode of

composition should know the context that the service will be

executed before composing the component services. For

adaptive service composition, Web service invocation is based

on the dynamic selection of concrete services at runtime. The

user or front-end application which invokes a Web service

may specify only its abstract interface requirements and quality of service constraints. The environments that users

interact with are dynamic by nature. Therefore it is not

possible to enumerate all the possible contexts and exceptions

at service design time.

A. Process Specification

A composed service is specified at an abstract level as a

high-level business process. We assume that a composed

service is characterized by a single initial task and a single end

task. The operational language for process implementation is BPEL . A set of semantic annotations are associated to the

process specification to specify requirements by the user of

the composed service. Some annotations specified are

Global and local constraints on quality dimensions.

Web service dependency constraints.

A process schema is a reusable and extensible business

process template devised to reach a particular goal . A process

schema can be configured by assigning a number of user

contexts to the process schema’s tasks and the schema itself.

Individual users can customize process schemas to meet their

particular requirements by assigning contexts to schema as parameters. A process schema can also be configured to

handle particular type of exceptions. Exception handling

Policies are assigned to the process schema and its tasks by

the relation policy Assignment. These policies prescribe the

knowledge on the appropriate response to a particular

exception.

A statechart is made up of states and transitions. States can

be initial, end, basic, or compound. A basic state corresponds

to an invocation of a service operation, whether an atomic

service, a composite service, or a community. The concept of

Web service community [6] is proposed to handle the large number and dynamic nature of Web services in a flexible way.

97

The statechart of a simplified process schema of the E-

learning class assistant is given in section VII. E-learning

class assistant helps students to manage their class activities.

B. Configuration of composite service

Process schemas that correspond to recurrent user needs

(e.g., booking rooms, travel planning) are defined by service

designers based on common usage patterns, and are stored in

schema repositories. The configuration is done by specifying a number of user contexts and exception handling policies and

assigning them to the process. Two abstractions for modeling

user contexts are execution contexts and data supply and

delivery contexts.

1) Execution Context: An execution context specifies that

certain conditions must be met in order to perform a particular

operation. Two constraints considered in our service

composition model are temporal and spatial constraints. These

constraints specify the time and location in which the task has

to be executed. Temporal and spatial constraints can be empty,

meaning that the corresponding task can be executed anytime and anywhere.

2) Data Supply and delivery Context: The value of a task’s

input parameter may be: 1) requested from user during task

execution. 2) obtained from user profile, or 3) obtained as an

output of another task, they are expressed as queries. Similarly,

the value of a task’s output parameter may be: 1) sent to other

tasks as input parameters and/or 2) sent to an end user.

IV. SERVICE ADAPTATION

If the invoked service changes the interface or protocol,

then all the composite services invoking it will have to

undergo analogous changes to interact with the new version of

the invoked service. Taxonomy of mismatches that can occur

between two services are Signature mismatch, Parameter

constraint, ordering mismatch, Extra message, Message split.

For each mismatch a template is provided in the AOP

approach to adaptation.Template contains a set of <pointcut,

advice> pairs that define where the adaptation logic is to be

applied, and what this logic is. An example of template for the signature mismatch is presented in table 1.

V. ARCHITECTURE OF OUR APPROACH

Fig.1 provides architecture of our framework for adaptable

web service composition. We propose DIstributed REgistries

(DIRE) instead of single UDDI registry. It provides support

for the cooperation among heterogeneous registries. It exploits

a Service Publication Bus to connect a set of registries and

allow them to share their services. The service builder, the

service discovery engine, the proxy service, and the service

deployer compose the service development and deployment

environment, which provide a service composition

environment where service designers and users can compose and invoke Web services. The runtime environment consists

of a set of generic services (coordination, context, and event)

that provide mechanisms for enacting the execution of

composite Web services.

TABLE I

TEMPLATE FOR SIGNATURE MISMATCH

Signature Template

Query Generic Adaptation Advice

query (<inputType>)

executes before receive when typebp = <inputType>

Signaturepart1 (<Ti>) Receive msgOes ; Assign msgObp .inParabp . typebp

<Ti> (msgOes .inParaes .typees); Reply msgObp;

query (<outputType>)

executes before reply when typebp = <outputType>

Signaturepart2 (<To>) Receive msgObp; Assign msgOes .outParaes . typees

<To> mgObp .outParabp .typebp);

Reply msgOes;

A. Service Development/Deployment Environment

The service discovery engine facilitates the advertisement

and location of services. Service registration, discovery, and

invocation are implemented by SOAP calls. When a service

registers with a discovery engine, a SOAP request containing

the service description in WSDL is sent to the DIRE registry.

After a service is registered in the DIRE registry, service designers and end users can locate the service by sending the

SOAP request (e.g., business name, service type) to the DIRE

registry. The service builder assists service designers in the

creation and maintenance of composite services. It provides

an editor for describing the statechart diagram of a composite

service operation and for importing operations from existing

services into composite services and communities. It should

be noted that the service builder also supports the

specification of process schemas.

The service deployer is responsible for generating

Aspect templates and control tuples of every task of a composite service. Once the control tuples and templates are

generated, the service deployer assists the service designer in

the process of uploading these tuples into the tuple spaces of

the corresponding component services and the composite

service. Coordination services communicate asynchronously

through the shared spaces by writing, reading, and taking

control tuples.

DIRE (DIstributed REgistry) fosters the seamless

cooperation among heterogeneous registries since it does not

require modifications to the publication and discovery

processes adopted by the organizations behind the registries. DIRE exploits the publish/subscribe (P/S) paradigm. Special-

purpose element, called dispatcher is used for both to

subscribe and publish. DIRE exploits both content-based and

subject-based subscriptions.

B. Runtime Environment

This layer contains the three core generic service that

provides the execution semantics for the adaptive composite

services. The coordination service provides an operation

called coordinate for receiving messages, managing service

instances (i.e., creating and deleting instances), registering

98

events to the event service, triggering actions, tracing service

invocations and communicating with other coordination

services. The coordination service relies on tuple space of the

associated service to manage service activities. The context

service detects, collects, and disseminates context information

while the event service fires and distributes events. Finally,

the event service provides operations for receiving messages,

including subscribing messages from the coordination service

of a service and context information from the context service,

and notifying the fired events to the coordination services.

.

Fig.1 Frame work for adaptable web service composition.

VI. EXECUTION MODEL

Existing service provisioning systems are centralized and

service orchestration is ensured by a single process which acts

as a central scheduler. Centralized execution models suffer

from permanent connectivity, availability and scalability

problems [7]. Accordingly, to achieve adaptive and scalable

execution of composite services in dynamic environments, the

participating services should be self-managed: they should be

capable of coordinating their actions in an autonomous way.

This leads to Decentralized Orchestration of composite web services. In Decentralized Orchestration, data and control

dependences between the components are analyzed and the

code can be partitioned into smaller components that execute

at distributed locations. We refer to this mode of execution as

decentralized orchestration. There are multiple engines, each

executing a composite web service specification at distributed

locations. The engines communicate directly with each other

to transfer data and control when an asynchronous manner.

Performance benefits of Decentralized Orchestration are

There is no centralized coordinator which can be a

potential bottleneck.

Distributing the data reduces network traffic and

improves transfer time.

A. Orchestration Enabling

The model consists of three core services, namely the

coordination service, the context service, and the event service. These three elements form orchestration enabling

services. When executing a composite service, orchestration

enabling services automatically schedule and execute the

component services, and adapt to user configured exceptions

and contexts. Each participating Web service is associated

with a coordination service that monitors and controls the

service execution. The coordination service determines when

should a component service be executed, and what should be

done after the execution is completed. The knowledge needed

by a coordination service in order to answer these questions at

runtime is statically extracted from the description of the

composite service (e.g., statecharts, user contexts), and placed in the corresponding tuple space. A coordination service

enforces the control tuples with the help of an event service

and a context service. The event service is responsible for

disseminating events registered by the coordination service,

and the context service is responsible for collecting context

information from context providers.

VII. IMPLEMENTATION

In our implementation, we design an E-Learning service.

We offer some courses. User has to register for a course. Once

they have registered, their username and password will be

stored. When the user accesses the service, first authentication of the user is performed. Then he is given the information

about his subjects lecture time and place. The service that

delivers lecture notes will take input from attendance

remainder service. Lecture notes are provided to the user

only during that particular lecture time. Lecture notes will be

delivered in an adaptive way. By analyzing the time required

by the users to view the pages and understand its contents, we

find an average time. If the user views the page beyond the

critical time limit, he is provided with a page that describes its

contents in a simpler form (e.g. diagrams, flowcharts etc.). If

the student has any doubts he can either post or vote his

queries based on the questions already raised. All these implementations are made in java with Apache Tomcat6 as a

Web server where Apache Axis is deployed. Apache Axis

provides not only a server-side infrastructure for deploying

and managing services, but a client-side API for invoking

these services. Each service has a deployment descriptor that

includes the unique identifier of the Java class to be invoked,

session scope of the class, and operations in the class available

for the clients. Each service is deployed using the service

management client by providing its descriptor and the URL of

the Axis servlet rpcrouter.

Service Builder

Proxy Service Service Discovery Engine

Service

Deployer

DIRE

Registry

Registry

Service Publication Bus

Registry

Registry

Tuple Space

Coordination service

Context

Context service

Event Service

Web Services

Composite Services

Development/

Deployment

Environment

Run time

Environment

99

Fig .2 class assistant process schema

VIII. CONCLUSION AND FUTURE WORK

The paper presented the framework for the deployment of

adaptable Web service compositions. First, we introduced a

adaptable service composition model. The innovative aspect

of our model is to provide distinct abstractions for service context and exceptions, which can be embedded or plugged

into the process schemas through simple interaction with end

users. It comprises DIRE for the user-controlled replication of

services onto distributed registries. We proposed the use of

AOP for service adaptation to interface and protocol

mismatches. The notion of template also promotes reusability

of adaptation logic that occurs repetitively across different

locations in an implementation of a service. In the future, we

plan to extend the Development Environment to offer a semi-

automated identification of mismatches and a graphical

interface that allows the user to create queries over process

specifications and navigate through the results

REFERENCES

[1] N. Milanovic and M. Malek, “Current Solutions for Web Service

Composition,” IEEE Internet Computing, vol. 8, no. 6,

pp. 51- 59 , Nov./Dec. 2004 [2] D. Ardagna and B. Pernici, “Adaptive Service Composition in

Flexible Processes,” IEEE Trans. Software Eng., vol. 33, no. 6,

pp. 369-384, June 2007.

[3] L. Baresi, E. Di Nitto, C. Ghezzi, and S. Guinea, “A Framework

for the Deployment of Adaptable Web Service Compositions,”

Service Oriented Computing and Applications, vol. 1, no. 1, pp.

75-91, 2007.

[4] W. Kongdenfha, R. Saint-Paul, B. Benatallah, and F. Casati, “An

Aspect-Oriented Framework for Service Adaptation,” Proc. Fourth

Int’l Conf. Service-Oriented Computing (ICSOC ’06), Dec. 2006. [5] Q. Z. Sheng, B. Benatallah, Z. Maamar, and A. H. H.

Ngu,“Configurable composition and adaptive provisioning of web

services”, IEEE Services Computing, vol. 2, NO. 1, pp.34- 49,

Mar. 2009.

[6] D. Berardi, G.D. Giacomo, and D. Calvanese, “Automatic

Composition of Process-Based Web Services: A Challenge,” Proc.

14th Int’l World Wide Web Conf. (WWW ’05), May 2005.

[7] G.B. Chafle, S. Chandra, V. Mann, and M.G. Nanda,

“Decentralied Orchestration of Composite Web Services,” Proc.

13th Int’l World Wide Web Conf. (WWW ’04), May 2004.

[8] Q. Yu, X. Liu, A. Bouguettaya, and B. Medjahed, “Deploying and

Managing Web Services: Issues, Solutions, and Directions,” The

VLDB J., vol. 17, no. 3, pp. 537-572, 2008.

[9] M.P. Papazoglou, P. Traverso, S. Dustdar, and F. Leymann,

“Service-Oriented Computing: State of the Art and Research

Challenges,” Computer, vol. 40, no. 11, pp. 38-45, Nov. 2007. [10] B. Srivastava and J. Koehler, “Web Service Composition—

Current Solutions and Open Problems, Proc. Int’l Conf.

Automated Planning and Scheduling (ICAPS ’03), 2003. [11] V. Agarwal, K. Dasgupta, N. Karnik, A. Kumar, A. Kundu,

S. Mittal, and B. Srivastava, “A Service Creation Environment

Based on End to End Composition of Web Services,” Proc. 14th

Int’l Conf. World Wide Web (WWW ’05), May 2005.

Course registration

User login

Attendance remainder

Attendance guide

Lecture notes

Browse

question

s

Vote question

Post question

consultation

feedback

Proceedings of the Third National Conference on RTICT 2010 Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

100

AN ENHANCED APPROACH FOR CLASS-DEPENDENT FEAUTRE SELECTION

K. Shanmuga Sundaram . B.E., J. Arumugam.M.E

Lecturer of Computer Science & Engg. Lecturer of Computer Science & Engg.

Hindustan University Hindustan University

Chennai-603 103 Chennai-603 103

[email protected] [email protected]

Abstract

Today data sets that we process are

becoming increasingly larger, not only in terms

of the number of patterns (instances), but also

the dimension of features (attributes), which may

degrade the efficiency of most learning

algorithms, especially when there exist irrelevant

or redundant features. The predictive accuracy

of the learning algorithms are reduced in the

presence of irrelevant features. The distribution

of truly relevant features for the main task are

blurred by irrelevant or redundant features.

Deleting those irrelevant features cannot only

improve the classification accuracy, but also

reduce the structural complexity of the radial

basis function (RBF) neural network and

facilitate rule extraction. All these reasons urge

us to carry out data dimensionality reduction

(DDR).

1. Literature Review

There are numerous techniques for

DDR. Depending on whether the original

features are transformed to new features, one

may categorize these techniques into feature

extraction or feature selection techniques,

respectively. Depending on whether a classifier

is used to evaluate the performance of the new

feature set, these DDR techniques can be

categorized into wrapper or filter methods,

respectively. Feature extraction methods, e.g.,

principal component analysis (PCA) and linear

discriminant analysis (LDA) transform the

original set of features into a new set of features.

Because the new features are different from the

original features, it may be difficult to interpret

the meaning of the new features. Feature

selection selects a desirable subset of the

original features while maintaining or increasing

acceptable classification accuracy. Feature

selection eliminates unimportant features, e.g.,

redundant and irrelevant features, and obtains

the best subset of features that discriminates

well among classes. Thus how to decide which

features are important so as to form the

desirable feature subset is the main objective of

feature selection.

Many factors affect the success of machine

learning on a given task. The representation and

quality of the instance data is first and foremost.

If there is much irrelevant and redundant

information present or noisy and unreliable data,

then knowledge discovery during the training

101

phase is more difficult. In real-world data, the

representation of data often uses too many

features, but only a few of them may be related

to the target concept. There may be

redundancy, where certain features are

correlated so that is not necessary to include all

of them in modeling; and interdependence,

where two or more features between them

convey important information that is obscure if

any of them is included on its own.

2.Proposed System

The general wrapper approach to

selection of class-dependent features consists of

the following three steps (Fig. 3.1).

1. In the first step, a C-class

classification problem has been converted into

two-class classification problems, i.e., problem

1, problem 2,…,and problem C. The goal of

problem c is to correctly separate the original

class c from the other patterns, where

c=1,2,……..,C. For problem c , All training

patterns have been into two classes: class A is

the original class c, and class B consists of all

the other training patterns.

2. The second step is to search for

a desirable feature subset for each binary

classification problem. In this paper, a forward

search has been adapted as follows. We rank

the attribute importance using a ranking

measure, such as RELIEF weight measure,

CSM, information-theoretic measure, and the

minimal-redundancy–maximal-relevancy

(mRMR) measure based on mutual information

(MI).

In this project, RELIEF weight measure

has been adapted, CSM, and the mRMR to

evaluate the importance of the features for each

class (in each of the C sub-problems). Three

feature ranking measures are briefly reviewed in

the next section. The attribute importance

ranking has been regarded thus obtained class-

dependent attribute importance ranking. After

obtaining the attribute importance ranking of

each class, choose a desirable feature subset

for each class through a classifier, e.g., the

support vector machine (SVM), for a given

feature subset search algorithm.

For each class, forward selection search

(or bottom-up search) has been used to form

different attribute subsets, that is, it has been

started with the most important feature as the

first feature subset, and then each time add one

attribute into the previous subset in the order of

importance ranking to form a new feature

subset, until a stopping condition is satisfied,

e.g., the validation accuracy starts to decrease

or all the attributes in this class have been

added. For high-dimensional data sets, it will be

computationally expensive to incrementally form

all feature subsets and evaluate them through a

classifier. Therefore, first select certain number

of features, e.g., the top 70 from 649 ranked

features for the handwritten digits

recognition(HDR) multi feature (or HDR) data

set from the LIBSVM data, and then

incrementally form different feature subsets. In-

putting each feature subset into the classifier,

we can obtain different classification accuracy

for different feature subsets. Then choose the

attribute combination with the highest

classification accuracy or lowest error rate as

the desirable feature subset for a given feature

102

subset search algorithm.

A feature mask has been used to

represent a feature subset, i.e., a “0” or “1” in a

feature mask indicates the absence or presence

of a particular feature, respectively. For

example, if originally there are five features,

i.e., x1, x2, x3, x4, x5, and the desirable

feature subset turns out to be x1,x2,x3 with the

fourth and fifth features deleted, the feature

mask will be 1,1,1,0,0.

In the testing stage after training, the

input test pattern has been classified according

to the classifier (Fig. 3.2) with the maximum

response. The classifier in Fig. 3.2 includes

several models. When using those class-

dependent feature subsets to train and test

models, each model has an output, which

provides the probability that each sample

belongs to the current class in terms of the

current class’ feature subset. If we use SVM in

all models, we have the probability estimate, of

each cur-rent class as the output of each model.

In SVM models, each class has a probability

estimate which indicates the probability the

testing sample belongs to each class. A

common way to generate an overall output from

individual outputs of all models is to choose the

maximum individual as the overall output of the

class-dependent classifier. However, it is not fair

to directly compare outputs produced from

different feature spaces proposed the pdf

projection theorem and utilized it to project those

class-specific features pdfs back into the original

space where a fair comparison is possible. In

our paper, we propose a heuristic method, i.e., a

weight measure, to deal with the problem of

unfair comparisons on outputs from different

feature spaces.

3. Feature Ranking Measures

3.1 Relief Weight Measure

Relief is a relevance-based function

selection algorithm inspired by instance based

learning, techniques. It defines relevance in

terms of consistency. It assumes that samples

belonging to the same class should be closer, in

terms of their feature values, than samples

from different classes. As a result, if two

instances (samples) belong to the same class

and are generally similar, the similar features

for the two instances are rewarded and the

different ones are punished. When the samples

belong to different classes the opposite occurs .

One important question addressed

here is whether Relief is able to identify

relevance when the learning task is function

approximation. Medications have been

proposed to deal with function approximation.

Some modifications proposed by Kononenko

have been used that generalize the algorithm to

multiple classes and reduce its sensitive to

noise. As part of this process, different versions

of modified algorithm has to be tested and tune

its parameters to best perform in out context.

The basic Relief algorithm selects

samples randomly from the sample space. For

each sample, it finds the closest sample of the

same class (hit) and the closest sample of a

different class (miss). The difference in feature

values between the sample and the hit and

between the sample and the miss are then used

to update the weights of each feature. If the hit

103

difference is smaller than the miss difference the

weight will increase, indicating that the feature is

a good one. The relief algorithm is given below:

1. Introduce a weight vector and

initialize it to zero:

w1,…,wi,…,wD=0. D is the number of

features.

2. Randomly select an instance X from

training instances S and find its nearest hit Xh

and nearest miss Xm

3. Calculate and update the weight wi of

the ith feature xi:wi=wi+(xi-xmi)2-(xi-xhi)

2,

i=1,2,….,D. Where xmi is the ith element of Xm

and xhi is the ith element of Xh. Each feature’s

weight is updated by the differences ( Euclidean

distance) and between the samples and the

hits(hit differences). The weight is increased if

the miss differences are greater than the hit

differences, which means that the feature is

relevant.

4. Repeat steps 2 and 3 over all training

instances. In our approach to class-dependent

feature selection, multi-class problems are

already converted into multiple two class

problems, and therefore we can use the original

RELIEF algorithm to rank the features.

However, for class-independent feature

selection method, it is multi-class problem so

that we will use the extended version RELIEF to

rank the features importance. Each sample will

have one hit and more than one miss, and the

differences between the misses and the sample

are weight averaged. The weight adopted here

is the a priori probability of the miss class.

After obtaining the weight of each

feature, the importance of the features has been

ranked according to rationale: the larger the

weights, the more important the features.

3.2 Class Separability Measure

The CSM evaluates how well two classes

are separated by a feature vector. The greater

the distance is between different classes, the

easier the classification task. Therefore, the

feature subset that can maximize the distances

between different classes may also maximize

classification accuracy and is therefore

considered more important. The jth sample is

represented as Xj,tj, where Xj= xj1,xj2,….,xjDs

the input data and tj=1,2,…..,C is the class

label of Xj. Class separability consists of two

elements, i.e., the distance between patterns

with each class Sw

C

c

nc

j

/TmcXcjmcXcjPcSw1 1

21]))([(

and the distance between patterns among

different classes Sb

C

c

/TmmcmmcPcSb1

21]))([(

Here Pc is the probability of the cth class

and nc is the number of samples in the cth class.

Xcj is the jth sample in the cth class, mc is the

mean vector of the cth class, and m is the mean

vector of all samples in the data set.

104

The ratio Sw/Sb can be used to measure

the separability of the classes: the smaller the

value, i.e., the smaller the distances within each

class and the greater the distance among

different classes, the better the separability. The

importance of a feature may be evaluated by

ratio Sw/Sb calculated after the feature is

removed from the data set, i.e., The greater

S’w/Sb’ is, the more important the removed

attribute is. For example, if removing attribute 1

from the data set leads to greater S’w/Sb’,

compared with removing attribute 2, we may

consider attribute 1 more important compared to

attribute 2 for classifying the data set, and vice

versa. Hence, we may evaluate the importance

level of the attributes according to ratio Sw/Sb

with an attribute deleted each time in turn.

3.3 The Minimal-Redundancy Maximal-

Relevance Measure:

The mRMR feature selection method was

proposed by Peng et al.. It is based on the MI

theory to select features in terms of their

relevancy and redundancy. Let FD=xi |

i=1,2,…,D denote a feature set. According to

the definition of MI, the MI value of two features

xi and xj is denoted as I(xi; xj) (i, j = 1,2,…,D),

which describes statistical dependence between

the two features. In the same manner, the MI

value I(xi; xj) (c is one of C classes) is used to

denote the statistical dependency of the feature

xi to the class c. Peng et al. used the

incremental search strategy to sequentially

select features with the maximal relevancy and

minimal redundancy. The sequence in which

features are included corresponds to the ranking

order of features. Suppose Fm -1 denote the

feature subset consisting of m-1 top ranked

features, the mth ranked feature xm should be

selected according to the following:

Although RELIEF cannot detect redundancy in

features, RELIEF was often adopted due to its

efficiency in computation. The CSM is used to

detect feature’s relevancy in classification

problems by many authors, e.g., Fu et al., even

as a linear ranking measure. Compared with

these two feature importance ranking

measures, the mRMR measure may be more

effective since it is a nonlinear method and it

evaluates features in terms of both

relevancy

Fig. 3.2. Architecture of the general classifier

with class-dependent feature subsets

(after training and during testing).

105

redundancy.

Fig. 3.1. Schematic diagram for general wrapper

approach to class-dependentfeature selection.

.

Suppose the output of the cth model is

denoted as Oc (c = 1, 2, …,C), which is the

probability that one sample belongs to the cth

model to classify the sample between two

classes, i.e., class A (the cth class) and class B (

all other classes except the cth class). For C

models, we have C outputs, i.e., O1,

O2,…..,O

c,…O

c, each representing the

probability that the sample belongs to the

current class, respectively. Instead of

comparing the C outputs obtained from different

feature spaces, we assign a weight to each

output and compare the weighted outputs, e.g.,

pc.O

c in Fig 3.2. These weights are obtained

feature space. That is, we first combine all

class-dependent feature subsets together to

form one union set, Xc. Then we classify

the sample using the union set, as the feature

subset with C classes as target classes,

together with a classifier, such as an SVM. The

classifier produces outputs p1,…..,p

c,…..,p

C

each representing the probability that the

sample belong to each of C class, respectively.

We assign these probabilities obtained from the

common feature space to the corresponding

O1,…..,O

c,…..,O

C for a fair comparison. Since

class-dependent feature subsets vary with

different training samples in the training process,

those weights model in Fig 3.2. has C outputs,

i.e., probability estimates for all classes.

4.Implementation Result

We train and test the artificial data sets

, then to select feature subset as optimum from

binary classifier using class-dependent feature

selection method. It is outperform compared to

class-independent feature selection method.

The figure 5.1 shows the class

dependent feature set input to the MATLAB

program. The figure 5.2 shows the feature

subsets output obtained from the MATLAB

program using class dependent feature

selection.

106

Fig. 5.1 Feature set (Input)

Fig. 5.2 Feature subsets (Output)

5.Conclusion and Future work

In this project, a general wrapper

approach to class-dependent feature selection is

proposed. In order to demonstrate the scenario

mentioned in the Introduction, two data sets A

and B have been generated. The experimental

results for the two artificial data show that our

class-dependent feature selection method can

select those specific feature subset for each

class, i.e., class-dependent feature subsets, so

that the number of noise features or redundant

features can be greatly reduced and the

classification accuracy is substantially

increased.

Future work is to decrease the

computational cost which may be worthwhile in

certain applications where improvements of

accuracy or reduction of data dimensionality are

very important and meaningful.

6.Reference

1. Lipo Wang, Nina Zhou, and Feng Chu

“A General Wrapper Approach to Selection of

Class-Dependent Features” IEEE Transactions

on neural networks vol. 19. No.7, July 2008

2. P. M. Baggenstoss, “Class-specific

classifier: Avoiding the curse of

dimensionality,” IEEE Aerosp. Electron.

Syst. Mag., vol. 19, no. 1, pt. 2, pp. 37–

52, Jan. 2004.

3. P. M. Baggenstoss, “The PDF projection

theorem and the class-specific method,”

IEEE Trans. Signal Process., vol. 51,

no. 3, pp. 672–685, Mar. 2003.

4. P. M. Baggenstoss, “Class-specific

features in classification,” IEEE Trans.

Signal Process., vol. 50, no. 12, pp.

3428–3432, Dec. 2002

5. J. Bins and B. Draper, “Feature

selection from huge feature sets,” in

Proc. Int. Conf. Comput. Vis.,

Vancouver, BC, Canada, 2001, pp.

159–165 [Online]. Available:

107

http://citeseer.ist.psu.edu/bins01feature.

html

6. A. L. Blum and P. Langley, “Selection of

relevant features and examples in

machine learning,” Proc. Nat. Acad. Sci.

USA, vol. 95, pp. 933–942, 1998.

Proceedings of the Third National Conference on RTICT 2010 Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

108

DESIGN AND IMPLEMENTATION OF A MULTIPLIER

WITH SPURIOUS POWER SUPPRESSION TECHNIQUE

(SPST)

JEYAPRABA.A

[email protected]

ABSTRACT

This project provides the

experience of applying an advanced

version of Spurious Power

Suppression Technique (SPST) on

multipliers for high speed and low

power purposes. When a portion of

data does not affect the final

computing results, the data controlling

circuits of SPST latch this portion to

avoid useless data transition occurring

inside the arithmetic units, so that the

useless spurious signals of arithmetic

units are filter out.

Modified Booth Algorithm is

used in this project for multiplication

which reduces the number of partial

product to n/2. To filter out the

useless switching power, there are

two approaches, i.e. using registers

and using AND gates, to assert the

data signals of multipliers after data

transition.

The simulation result shows

that the SPST implementation with

AND gates owns an extremely high

flexibility on adjusting the data

asserting time which not only

facilitates the robustness of SPST but

also leads to a speed improvement

and power reduction.

MODIFIEDBOOTH

ENCODER ALGORITHM

A generator that creates a

smaller number of partial products

will allow the partial product

summation to be faster and use less

hardware. A method known as

Modified Booth’s Algorithm reduces

the number of partial products by

about a factor of two, without

requiring a pre add to produce the

partial products.

The multiplier is partitioned

into overlapping groups of 3 bits, and

each group is decoded to select a

single partial product as per the

selection table. In general the there

will be n/2 products, where n is the

operand length.

The encoding of the multiplier

Y, using the modified booth

algorithm, generates the following

109

five signed digits, -2, -1, 0, +1, +2.

Each encoded digit in the multiplier

performs a certain operation on the

multiplicand, X, as illustrated in

figure 2.

Figure 1: Multiplication using

Modified Booth encoding

Figure 2 – 16 bit Booth 2 Example

Fig. 2 shows a computing example of

Booth multiplying two numbers

“2AC9”and “006A,” where the

shadow denotes that the numbers in

this part of Booth multiplication are

all zero so that this part of the

computations can be neglected.

Saving those computations can

significantly reduce the power

consumption caused by the transient

signals.

DETECTION LOGIC

The first implementing

approach of the control signal

assertion circuits is using registers as

shown in figure 3. The registers may

latch the wrong values because of

delay problem. To solve this problem,

we adopt the other implementing

approach of the control signal

assertion circuits illustrated in the

shadow area in Fig. 4, using an AND

gate in place of the registers to control

the signal assertion. Efficient timing

control and power consumption

achieved through this implementation.

Figure 3 - Detection Logic using

Register

110

Figure 4 – Detection Logic using

AND gate

SPSTADDER

The SPST is illustrated through

a low-power adder/subtractor design

example. The adder/subtractor is

divided into two parts, i.e., the most

significant part (MSP) and the least

significant part (LSP). The MSP of

the original adder/subtractor is

modified include detection logic

circuits, data controlling circuits,

denoted as latch-A and latch-B in Fig.

5, sign extension circuits, and some

glue logics for calculating the carry in

and carry out signals.

Figure 5 – Low Power SPST

ADDER

LOW POWER MULTIPLIER

DESIGN USING SPST

APPLYING SPST ON MODIFIED

BOOTH ENCODER

We already discussed the

computing example of Booth

Multiplying of two numbers “2AC9”

and “006A” in figure 2, where the

shadow denotes that the numbers in

this part of Booth multiplication are

all zero so that this part of the

computations can be neglected. We

propose the SPST-equipped modified-

Booth encoder, which is controlled by

a detection unit. The detection unit

has one of the two operands as its

input to decide whether the Booth

encoder calculates redundant

computations. As shown in figure 6,

the latches can, respectively, freeze

the inputs of MUX-4 to MUX-7 or

only those of MUX-6 to MUX-7

when the PP4 to PP7 or the PP6 to

PP7 are zero, to reduce the transition

power dissipation. Such cases occur

frequently in e.g., FFT/IFFT,

DCT/IDCT, and Q/IQ which are

adopted in encoding or decoding

multimedia data.

111

Figure 6 – SPST equipped Modified

Booth Encoder

APPLYING THE SPST ON

COMPRESSION TREE

The proposed SPST-equipped

multiplier is illustrated in Fig. 7. The

PP generator generates the five

candidates of the partial products, i.e.,

2A, A, 0, +A, +2A, which are then

selected according to the Booth

encoding results of the operand B.

Moreover, when the operand besides

the Booth encoded one has a small

absolute value, there are opportunities

to reduce the spurious power

dissipated in the compression tree.

According to the redundancy

analysis of the additions, we replace

some of the adders in compression

tree of the multiplier with the SPST-

equipped adders, which are marked

with oblique lines in Fig. 7.

The bit-widths of the MSP and

LSP of each SPST-equipped adder are

also indicated in fraction values

nearing the corresponding adder in

Fig. 7.

Figure 7 – Proposed High

Performance and Low Power

Multiplier

SIMULATION RESULT OF

PARTIALPRODUCT

GENERATOR USING MODIFIED

BOOTH ENCODER

112

CONCLUSION

The high speed low

power multiplier adopting the new

SPST is designed. The Multiplier is

designed by equipping SPST on a

modified Booth encoder which is

controlled by a detection unit using

AND gate. The modified Booth

encoder will reduce the number of

partial products generated by a factor

two. The SPST adder will avoid the

unwanted addition and thus minimize

the switching power dissipation. The

SPST implementation with AND gate

have an extremely high flexibility on

adjusting the data asserting time. This

facilitates the robustness of SPST can

attain significant speed improvement

and power reduction when compared

with the conventional tree multipliers.

This design verified using Xilinx 9.1

using Verilog HDL coding.

REFERENCES

Kuan-Hung Chen and

Yuan-sun Chu “Low

Power Multiplier with

Spurious Power

Suppression Technique”

IEEE Trans on VLSI

Systems, VOL 15, July

2007.

O. Chen, R. Sheen, and

S. Wang, “A low power

adder operating on

effective dynamic data

ranges,” IEEE Trans.

Very Large Scale Integr.

(VLSI) Syst., vol. 10, no.

4, pp. 435–453, Aug.

2002

S. Henzler, G.

Georgakos, J. Berthold,

and D. Schmitt-

Landsiedel, “Fast power-

efficient circuit-block

switch-off scheme,”

Electron. Lett., vol. 40,

no. 2, pp. 103–104, Jan.

2004

Proceedings of the Third National Conference on RTICT 2010 Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

113

ROBOTICS MOBILE SURVEILLANCE USING ZIGBEE

H.Senthil Kumar1 H.Saravanakumar

2 V.Suresh

3 S.T.Prabhu Sowndharya

4

1Assistant Professor, Mepco Schlenk Engineering College, Sivakasi.

2,3,4Final ECE, Mepco Schlenk Engineering College, Sivakasi.

ABSTRACT

Recently, increasing research attention

has been directed toward wireless

technology,Collections of small, low power

nodes, that can intelligently deliver high level

sensing results to the user. Mobile sensor provide

an inexpensive and more convenient way to

monitor physical environments. Mobile sensor

system can enrich human life in applications

such as home security, building monitoring, and

healthcare.

Our robotics mobile surveillance system

uses numerous sensors, zigbee, line follower

robot, web cam etc. to monitor the physical

environment. Conventional approaches in the

implementation of surveillance systems

continuously capture events, which require huge

computation or manpower to analyze. Robotics

mobile surveillance system can reduce such

overhead while providing more advanced

services.

Key word: Zigbee, Line follower robot,

wireless camera, sprinklers

1.INTRODUCTION

Robotics mobile surveillance system

consists of zigbee, line follower robot, wireless

web cam, microcontroller and various sensors

which monitors human intrusion, temperature,

smoke, fire, flood, gas leakage. When any

unwanted events happening in the environment,

the corresponding sensor indicates the

microcontroller which transmits the

corresponding data to the remote place via

zigbee module. If necessary the web cam

capture that events and transmits the snapshot

wirelessly. So this eliminates the continues

monitoring of camera and greatly reduce the

memory storage.Also if fire occurs in that

environment zigbee switch on the sprinklers via

wirelessly to cut off fire.

114

BLOCK DIAGRAM

Figure 1: Block diagram of our project

2.DESCRIPTION

2.1ZIGBEE

Zigbee is a technological standard

created for control and sensor networks. It is

based on IEEE 802.15.4 standard which is

created by zigbee alliance which aims low data

rate, low power consumption, small packet

devices. Zigbee network is a self-organizing and

supports peer-to-peer and mesh networks.

Bandwidth: 20-250KB/s

Transmission Range: 1-100+ meters

System Resources: 4KB-32KB

Network size: 2^64

Battery life: 100-1000 days

Figure 2: IEEE 802.15.4

2.1.1 ZIGBEE NETWORK FORMATION

Zigbee networks are called personal area

networks (PAN). Each network contains a 16-

bit identifier called a PAN ID. ZigBee defines

three different device types – coordinator,

router, and end device.An example of such a

network is shown below.

Figure 3: Node Types / Sample of a Basic

ZigBee Network Topology

Coordinator – Responsible for selecting the

channel and PAN ID. The coordinator starts a

new PAN. Once it has started a PAN, the

coordinator can allow routers and end devices to

join the PAN. The coordinator can transmit and

receive RF data transmissions, and it can assist

in routing data through the mesh network.

Router – A router must join a ZigBee PAN

before it can operate. After joining a PAN, the

router can allow other routers and end devices to

join the PAN. The router can also transmit and

receive RF data transmissions, and it can route

data packets through the network

115

End Device – An end device must join a

ZigBee PAN, similar to a router. The end

device, however, cannot allow other devices to

join the PAN, nor can it assist in routing data

through the network. An end device can

transmit or receive RF data transmissions. Since

the end device may sleep, the router or

coordinator that allows the end device to join

must collect all data packets intended for the

end device, and buffer them until the end device

wakes and is able to receive them.

In our project zigbee is used to transmit digital

output data from microcontroller to remote

place.For example if any gas leakage occurs in

the environment it will indicate to the remote

place which is either computer or lcd

display.And if fire is detected by the

corresponding sensor, zigbee automatically

switch on the sprinklers attached in the wall.

3. LINE FOLLOWER ROBOT

Line follower robot is a machine that can follow

a path. The path can be visible like a black line

on a white surface (or vice-versa) or it can be

invisible like a magnetic field.In our project we

are using differential drive line follower.

Differential drive means the wheel gets the

power from separate source respectively. And

the sensor we are using is white line sensor.

White line sensor is the most critical part of the

robot. It is basically a light intensity sensor. It

consists of infrared LED and photo diode.

Infrared LED illuminates the surface and photo

diode senses the reflected light from the surface.

We are using Infrared LED because ir LED

emits light at 940 nanometer and photo diode

shows maximum sensitivity at 940 nanometer

wavelength. Based on the output from the white

line sensor, the power to each wheel differs.So

the speed of each wheel can vary. The power

applied to the wheel is based on principle of

pulse width modulation (PWM).The

microcontroller we are used is P89V51RD2.

Figure 4: Line Follower Block Diagram

Figure 5:Assembled Line Sensor

SENSORS

In our project the following sensors are

used:

116

PIR module

LM35(Temperature)

SY-HS-220(Humidity) MQ5(Gas Sensor)

4.RESULTS AND DISCUSSION

4.1 HUMAN DETECTION

The PIR (Passive Infra-Red) Sensor is a

pyroelectric device that detects motion by

measuring changes in the infrared levels emitted

by surrounding objects. This motion can be

detected by checking for a high signal on a

single I/O pin. Pyroelectric devices, such as the

PIR sensor, have elements made of a crystalline

material that generates an electric charge when

exposed to infrared radiation. The changes in

the amount of infrared striking the element

change the voltages generated, which are

measured by an on-board amplifier. The device

contains a special filter called a Fresnel lens,

which focuses the infrared signals onto the

element. As the ambient infrared signals change

rapidly, the on-board amplifier trips the output

to indicate motion.

4.2 TEMPERATURE DETECTION

The LM35 series are precision

integrated-circuit temperature sensors, whose

output voltage is linearly proportional to the

Celsius (Centigrade) temperature

4.3 GAS LEAKAGE SENSOR

The gas sensor we are used is MQ5.

They are used in gas leakage detecting

equipments in family and industry, are suitable

for detecting of LPG, natural gas , town gas,

avoid the noise of alcohol and cooking fumes

and cigarette smoke

5. CONCLUSION

The sensor network community has

moved out of its infancy and is now embarking

on more serious design efforts—large-scale,

long-lived systems that truly require self-

organization and adaptivity.In this project we

developed zigbee based robotics mobile

surveillance.This can greatly reduce the

manpower and enhance secutity.

6. FUTURE WORK

Our project can be extended to send the

information to users via GSM modem.So the

user can get the information from anywhere in

the world.

7. REFERENCES

[1].A zigbeebased home automation syste

m,Gill, K.; Shuang-Hua Yang; Fang Yao; Xin

Lu; Consumer Electronics, IEEE

Transactions on Volume: 55 , Issue: 2 ,2009

117

[2].Wireless Home Securityand Automation

System Utilizing ZigBee based Multi-hop

Communication,Sarijari, M.A.B.; Rashid,

R.A.; Rahim, M.R.A.; Mahalin, N.H.; Telecommunication Technologies 2008 and 2008 2nd Malaysia Conference on Photonics. NCTT-MCP 2008.

[3]. Realization of Home Remote Control

Network Based on ZigBee,ZhangShunyang

XuDu,JiangYongping,WangRiming. Electr

onic Measurement and Instruments, 2007. ICEMI'07.

[4]. D. Johnson et al., “Mobile Emulab: A

Robotic Wireless and Sensor Network Testbed,” Proc. 25th Ann. IEEE Conf. Computer Comm. (Infocom 06), IEEE Press, 2006.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

118

Abstract - Many people have irregular heartbeats from time

to time. Some heart problems occur during certain activities,

such as eating, exercise or even sleeping. Sometimes the

irregular heartbeats don’t influence life style and are usually

harmless in normal hearts. But it is also possible that these

irregular heartbeats with pre-existing illness can cause heart

attacks that lead to death. Holter monitoring system records

continuous electrocardiographic measurements of the

heart’s rhythm. Usually the recording time is around 24 to

48 hours. That means even when the heartbeat is normal, the

Holter monitor also works as well. The system discussed

here can automatically record the ECG wave when the user

is not feeling good or the heartbeats are not regular. The

recording algorithm is not continuous any more. This

increases the recording time up to 72 hour or more. The R-

Peak detection is used for ECG analysis. The Hilbert

transform of ECG signal is used for efficient R-Peak

detection and is simulated using Matlab.

Keywords - Electrocardiogram (ECG), Holter Monitor, Irregular

Heartbeats, R-Peak, Hilbert transform.

I. INTRODUCTION

The standard ECG is a representation of the

heart’s electrical activities recorded from electrodes that

are placed on different parts of patient’s body. The

electrocardiogram is composed of complexes and waves.

In normal sinus rhythm, waves and complexes are the P

wave, QRS complex, ST Segment, T wave and U wave.

Measurements are PR interval, QRS duration, QT

interval, RR interval and PP interval.

Since the development of medical science, many

instruments for improving people’s health have been developed. The electrocardiogram (ECG) monitoring

system is one of them. The most common type of ECG

monitoring is Holter monitoring. Holter monitoring is a

portable recording tool and can help doctors make a

precise diagnosis[1].

The system discussed here can automatically

record the ECG wave when the user is not feeling good or

the heartbeats are not regular. The recording algorithm is

not continuous any more. It also can record the heartbeats

manually; the wearers can record the heartbeat if wanted

when the heart rhythms are ordinary.

II. BASICS OF ELECTROCARDIOGRAM (ECG) AND

HOLTER MONITOR

A. Electro cardiogram

The standard ECG is a representation of the

heart’s electrical activities recorded from electrodes that

are placed on different parts of patient’s body. The

electrocardiogram is composed of complexes and waves.

In normal sinus rhythm, waves and complexes are the P

wave, QRS complex, ST Segment, T wave and U wave.

Measurements are PR interval, QRS duration, QT interval, RR interval and PP interval[2,3]. Figure 1

illustrates a typical waveform of normal heartbeats and

intervals as well as standard time and voltage measures.

Figure 1. ECG signal

B. Advantanges of Holter Monitoring System The Holter monitoring system records the

electrical activity of heart during usual daily activities.

The system discussed here can automatically record the

ECG wave when the user is not feeling good or the

heartbeats are not regular[4]. Comparison with New

system is shown in Table 1. TABLE 1

COMPARISON BETWEEN GENERAL HOLTER MONITOR WITH

NEW SYSTEM DEVELOPED HERE

General

Holter

Monitor

New system

Continuity Continuous Intermittent

Saving time 24-48

hours

More than 48 hours

Operation Manually

operated

Automatic/Manually

operated

Effective Electrocardiographic Measurements of The Heart’s Rhythm Using

Hilbert Transform

Manoj Kumar Verma1, P. Muthu

2

1P.G Scholar, Karunya University. Coimbatore,India

2Assistant Professor, Karunya University, Coimbatore,India

119

C.Structure of The Holter Monitoring System

The structure of Holter monitoring system is shown in

Figure 2.

Figure 2. The structure of the Holter Monitoring System

1) Differentiation: The differentiation of ECG data is to

remove the baseline drift and to minimize other

artifacts.

2) Hilbert transform: One of the properties of the

Hilbert transform is that it is an odd function. Similarly a crossing of the zero between consecutive

positive and negative inflexion points in the original

waveform will be represented as a peak in its Hilbert

transformed conjugate[5,6]. This interesting property

can be used to develop an elegant and much easier

way to find the peak of the QRS complex in the ECG

waveform corresponding to a zero crossing in its first

differential waveform d/dt (ECG).

3) R-Peak Detection: R-Peaks are detected using an

adaptive thresholding and maximum value in the

sequence. The second stage detector uses the

information provided by the first approximation. A defined width window subset (i.e., ±10 samples from

the location of the peak found in the corresponding

h(n) sequence) is selected in the original ECG

waveform to locate the real R peak[7]. Once again a

simple maximum peak locator in the values of this

subset sequence is used.

D.Holter Monitoring Review

The Holter monitor is battery-powered and can

continuously record the electrical activities of the heart over a specified period of time, normally 24 to 48 hours.

Usually the patient will undergo Holter monitoring as an

outpatient, meaning that the monitor will be placed on the

body of the patient by a technician in a cardiologist’s

office. Then the patient will go home and do normal

activities[8,9]. With the development of technology, the

Holter monitor is greatly reduced in size. It is now very

compact and combined with digital recording and used to

record ECGs.

As Figure 3 shows, the Holter monitor is a

small-size recording device. The monitor has wires called

leads. The leads attach to metal disks called electrodes, which the user wears on his chest. These electrodes are

very sensitive, and they can pick up the electrical

impulses of the heart. The impulses are recorded by the

Holter monitor record the heart’s electrical activity.

Figure 3. A man with the Holter monitor

Advanced Holter monitors have been developed

that employ digital electrocardiographic recordings,

extended memory for more than 24 hours recording, pacemaker pulse detection and analysis, software for

analysis of digital ECG recordings that are downloaded

and stored on a computer, and capability of transmission

of results over the internet.

E. Hilbert Transform Review

In 1893, the physicist Arthur E. Kennelly (1861-

1939) and the scientist Charles P. Steinmetz (1865-1923)

first used the Euler formula

e jz = cos (z) + j sin (z) (1)

which was derived by a famous Swiss

mathematician Lenonard Euler (1707-1783) to introduce

the complex notation of harmonic wave forms in

electrical engineering, that is:

e jwt = cos (ωt) + j sin (ωt) , (2)

Where j is the imaginary unit.

In the beginning of the 20th century, the German

scientist David Hilbert (1862-1943) proved that the

Hilbert transform of the function cos (ωt) is sin (ωt). This

is the one of properties of the Hilbert transform, i.e., basic ± π/2 phase-shift[10].

III. IMPLEMENTATION AND SIMULATION

RESULTS

A. R-Peak Detection Using Adaptive Thresholding

Since this algorithm for Hilbert transformation

works well with short sequences, a moving 1024 points

window is used to subdivide the input sequence y(n)

before obtaining its Hilbert transform. In this work, the

sample frequency used is 360 Hz. To optimize accuracy,

the starting point of the next window should match the last R point located in the previous ECG subset. Because

the P and T waves are minimized in relation to the relative

peak corresponding to QRS complex in the Hilbert

sequence, simple threshold detection is used to locate the

peaks in the h (n) sequence[11]. The threshold must be

adaptive in order to guarantee accurate detection of the R

peaks.

Raw ECG

Data

Differentiation

of ECG signal

Hilbert

transform of

differentiated

ECG signal

R-Peak

detection algorithm

Second stage detection

Recording in flash memory

120

The decision is made in base of the amplitude of

the peak and their position in relation to the last R peak

located using an adaptive time threshold based on the

average R–R interval length of the previous R peaks

located.

B. Second Stage Detector

The second stage detector uses the information

provided by the first approximation. A defined width

window subset (i.e., ±10 samples from the location of the

peak found in the corresponding h(n) sequence is selected

in the original ECG waveform to locate the real R peak.

C. Simulation Results

The algorithm discussed above is simulated

using Matlab and the results are shown in Figure 4.

Figure 4. Matlab simulation result waveforms

We can see in the figure that the R-peaks are clear

in the Hilbert transformed sequence and baseline drift is

removed by first differential of ECG signal. Here a

random subset window of 1024 samples is taken and R-

peak is detected using adaptive thresholding and second

stage detector.

IV. CONCLUSION

From the Matlab simulation it can be seen that

the R-peaks in the Hilbert transformed data are clearer

and the base line drift is removed by first differential of

the ECG waveform. So, the P and T waves are minimized

in relation to the relative peak corresponding to QRS

complex in the Hilbert sequence. So, simple threshold

detection is used to locate the peaks in Hilbert

transformed sequence.

REFERENCE

[1] Bolton, R.J., L.C. Westphal, On the use of the Hilbert Transform

for ECG waveform processing, in: Computers in Cardiology, IEEE

Computer Society, Silver Spring, MD, 1984, pp. 533–536.

[2] Bracewell, R.N., The Fourier Transform and its Applications,

McGraw-Hill, New York, 1978, pp. 267–274.

[3] Oppenheim, A.V., R.W. Schafer, Discrete-Time Signal Processing,

Prentice-Hall, Englewood Cli7s, NJ, 1989, p. 775.

[4] Friesen, G.M., T.C. Jannett, M.A. Jadallah, S.L. Yates, S.R. Quint,

H.T. Nagle, A comparison of the noise sensitivity of nine QRS

detection algorithms, IEEE Trans. Biomed. Eng. 37 (1990) 85–98.

[5] Hamilton, P.S., W.J. Tompkins, Quantitative investigation of QRS

detection rules using the MIT=BIH arrhythmia database, IEEE Trans.

Biomed. Eng. 33 (1986) 1157–1165.

[6] Lee, J., K. Jeong, J. Yoon, M. Lee, A simple real-time QRS detection

algorithm, Proceedings of the 18th Annual International Conference of

the IEEE Engineering in Medicine and Biology Society, Amsterdam,

1996.

[6] Pan, J., W.J. Tompkins, A real time QRS detection algorithm, IEEE

Trans. Biomed. Eng. 32 (1985) 230–236.

[7] Afonso, V.X., W.J. Tompkins, T.Q. Nguyen, S. Luo, ECG beat

detection using filter banks, IEEE Trans. Biomed.Eng. 46 (1999) 192–

201.

[8] Ruha, A., S. Sallinen, S. Nissila, A real-time microprocessor QRS

detector system with a 1ms timing accuracy for the measurement of

ambulatory HRV, IEEE Trans. Biomed. Eng. 44 (1997) 159–167.

[9] Zheng, Li.C., C. Tai, Detection of ECG characteristic points using

wavelet transforms, IEEE Trans. Biomed. Eng. 42 (1995) 21–28.

[10] Xue, Q., Y.H. Hu, W.J. Tompkins, Neural-network-based adaptive

matched 6ltering for QRS detection, IEEE Trans. Biomed. Eng. 39

(1992) 315–329.

Proceedings of the Third National Conference on RTICT 2010 Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

121

Visual Analysis for Object Tracking Using Filters

and Mean Shift Tracker B.Vijayalakshmi

#1, K.Menaka

#2,

#1 Assistant professor, KLN College of Information Technology, Madurai #2M.E Student, KLN College of Engineering(2) , Madurai.

[email protected],[email protected]

Abstract

This paper represents a technique for the

target representation and localization, the

central component in visual tracking of

objects is proposed. Colour histogram is used

as a reliable feature to model object and its

adaptation handles with elimination changes.

Tracking plays a key role in an effective

analysis of an object. Many experimental

results demonstrate the successful tracking of

targets whose visible colors change drastically

and rapidly during the sequence, where the

basic Mean Shift tracker obviously fails.

Definite measure minimization during

tracking enhances the guarantees of

convergence. . Proposed method extends the

traditional mean shift tracking, which is

performed in the image coordinates, by

including the scale and orientation as

additional dimensions and simultaneously

estimates all the unknowns in a few number of

Asymmetric mean shift iterations. We

describe only few of the potential applications:

exploitation of background information,

Kalman tracking using motion models, and

face tracking. The object tracking using

Gaussian approach is revealed by Kalman

Filter whereas the non Gaussian approach is

by means of condensation algorithm. The

retrieval of an image during its motion is

revealed by Mean Shift Tracker. The

extension of this paper can be revealed in

features of stable convergence and speed.

Keywords: Object tracking, Filters, Mean shift

tracker.

I. INTRODUCTION

Tracking is the propagation of shape and motion

estimates over time, driven by a temporal stream

of observations. The noisy observations that arise

realistic problems demand a robust approach

involving propagation of probability distribution

122

over time. Modest levels of noise may be treated

satisfactorily using Gaussian densities, and this is

achieved effectively by Kalman Filtering. More

pervasive noise distributions as commonly arise

in visual background clutter, demand a more

powerful, a non –Gaussian approach. The

condensation algorithm described, combines

random sampling with learned dynamical model

to propagate an entire probability distribution for

object position and shape over time. Mean Shift

tracker works by searching in each frame for the

location of an image region whose color

histogram is closest to the reference color

histogram of the target. The distance between

two histograms is measured using their

Bhattacharyya coefficient, and the search is

performed by seeking the target location via

Mean Shift iterations. Traditional mean shift

method requires a symmetric kernel, such as a

circle or an ellipse, and assumes constancy of the

object scale and orientation during the course of

tracking.

1.1 Object Tracking Filters

1.1.1 Kalman Filter

When the noise sequences are Gaussian and are

linear functions, the optimal solution is provided

by the Kalman filter which yields the posterior

being also Gaussian. When the functions are

nonlinear, by linearization the Extended Kalman

Filter (EKF) the posterior density being still

modeled as Gaussian. A recent alternative to the

EKF is the Unscented Kalman Filter (UKF)

which uses a set of discretely sampled points to

parameterize the mean and covariance of the

posterior density. The most general class of

filters is represented by particle filters also called

bootstrap filters. The filtering and association

techniques discussed above were applied in

computer vision for various tracking scenarios.

Boykov and Huttenlocher [1] employed the

Kalman filter to track vehicles in an adaptive

framework. Rosales and Sclaroff [2] used the

Extended Kalman Filter to estimate a 3D object

trajectory from 2D image motion. Particle

filtering was first introduced in vision as the

Condensation algorithm by Isard and Blake [3].

Probabilistic exclusion for tracking multiple

objects was discussed in [4]. Wu and Huang

developed an algorithm to integrate multiple

target clues [5]. Li and Chellappa [6] proposed

simultaneous tracking and verification based on

particle filters applied to vehicles and faces.

Applications are smoothness of the similarity

function allows application of a gradient

optimization method which yields much faster

target localization compared with the (optimized)

exhaustive search. In our case the Bhattacharyya

coefficient has the meaning of a correlation

score. The new target representation and

localization method can be integrated with

various motion filters and data association

techniques.

1.1.2 Condensation Filtering Algorithm

123

The Condensation algorithm is based on factored

sampling but extended to apply iteratively to

successive images in a sequence. Given that the

estimation process at each time-step is a self-

contained iteration of factored sampling, the

output of an iteration will be a weighted, time-

stamped sample-set, denoted St(n), , n = 1, ... ,

N with weights ∏t(n), representing

approximately the conditional state-density

p(xt/Zt) at time t, where Zt = (Z1……..Zt). How

this is sample-set obtained? Clearly the process

must begin with a prior density and the effective

prior for time-step t should be p(xt/Zt-1).

This prior is of course multi-modal in general and

no functional representation of it is available. It is

derived from the sample set representation (S t-1

(n)), ∏ (t-1) (n) ), n. = 1, ... , N of (x t-1(n) ,Z t-1n),

the output from the previous time-step, to which

prediction must then be applied. The iterative

process applied to the sample-sets is depicted in

Figure 1.

At the top of the diagram, the output from time-

step t - 1 is the weighted sample-set (S t-1 (n) ,∏

t-1 (n)), n = 1, ...N the aim is to maintain at

successive time steps, sample sets of fixed size

N, so that the algorithm can be guaranteed to run

within a given computational resource. The first

operation therefore is to sample (with

replacement) N times from the set St-1(n)

choosing a given element with probability ∏t-

1(n) .Some elements, especially those with high

weights, may be chosen several times, leading to

identical copies of elements in the new set.

Others with relatively low weights may not be

chosen at all.

Each element chosen from the new set is now

subjected to a predictive step. The dynamical

model we generally use for prediction is a linear

stochastic differential equation (s.d.e.) learned

from training sets of sample object motion (Blake

et al., 1995). The predictive step includes a

random component, so identical elements may

now split as each undergoes its own independent

random motion step. At this stage, the sample set

St(n), for the new time-step has been generated

but, as yet, without its weights; it is

approximately a fair random sample from the

effective prior density p(x t /Z t-1) for time-step t.

Finally, the observation step from factored

sampling is applied, generating weights from the

observation density p(Zt/xt) to obtain the sample-

set representation S t (n), ∏ t (n) of state-

density for time t.

Figure 1.1 One time step in the condensation algorithm. Blob centers

represent the sample values and sizes depict the sample weights

The algorithm is specified in detail in Figure 1.1.

The process for a single time-step consists of N

iterations to generate the N elements of the new

124

sample set. Each iteration has three steps,

detailed in the figure, and explained as follows.

1. Select nth new sample St(n)

to be some

St-1(j)

from the old sample set, sampled

with replacement with probability ∏t-

1(j)

This is achieved efficiently by using

cumulative weights Ct-1(j)

(constructed

in step 3).

2. Predict by sampling randomly from

the conditional density for the dy-

namical model to generate a sample

for the new sample-set.

3. Measure in order to generate weights

∏t(n)

for the new sample.

1.2 Mean Shift Tracker

Recently, Mean-Shift Tracking [2] has attracted

much attention because of its efficiency and

robustness to track non-rigid objects with partial

occlusions, significant clutter, and variations of

object scale. As pointed out by Yang and

Duraiswami [7], the computational complexity of

traditional Mean-Shift Tracking is quadratic in

the number of samples, making real-time

performance difficult. Although the complexity

can be made linear with the application of a

recently proposed fast Gauss transform [7],

tracking in real-time remains a problem when

large or multiple objects are involved. The main

focus on the advantage enhancing high quality

edge preserving filtering and image segmentation

can be obtained by means of applying the mean

shift in the combined spatial-range domain. The

method we developed is conceptually very

simple being based on the same idea of

iteratively shifting a fixed size window to the

average of the data points within. Details in the

image are preserved due to the non parametric

character of the analysis which does not assume a

priori any particular structure for the data.

1.2.1 Asymmetric kernel mean shift

The asymmetric kernel mean shift, in which the

scale and orientation of the kernel adaptively

change depending on the observations at each

iteration. Proposed method extends the

traditional mean shift tracking, which is

performed in the image coordinates, by including

the scale and orientation as additional dimensions

and simultaneously estimates all the unknowns in

a few number of mean shift iterations.

Asymmetric kernels [7] have been used in the

area of statistics for over a decade and have been

shown to improve the density estimation. The

main advantage of anisotropic symmetric kernels

is that most of the non-object region resides

outside of the kernel. These kernels, however, do

not represent the object shape and still contain

non-object regions as part of the object. An ideal

kernel has the shape of the tracked object which

may be asymmetric as shown in Figure 1.2.

Asymmetric kernels, however, may not have an

analytical form. This observation suggests the

125

use of implicit functions for defining the profile

of arbitrarily shaped kernels.

Figure 1.2 Asymmetric kernel mean shift tracker image

II EXPERIMENTAL RESULTS

The following tracking ball picture represents the

various results obtained using Kalman and

condensation filters as well as Mean Shift

Tracker.

Figure 2.1 Input Tracking Ball

Figure 2.2 Removed Background

Figure 2.3 Erosion

Figure 2.4 Detected Ball

Figure 2.5 Kalman Filtered Ball

Figure 2.6 Condensed Filtered ball

Moving objects are characterized by their color-

histograms. Therefore the key operation of the

object tracking algorithm is histogram estimation.

Mean-shift tracking algorithm is an iterative

126

scheme based on comparing the histogram of the

original object in the current image frame and

histogram of candidate regions in the next image

frame. The aim is to maximize the correlation

between two histograms.

Figure 2.8. Moving object tracking results; frame 1, 10, 20, 30, 60 and 70.

III CONCLUSION

Thus the object is tracked visually using Kalman

Filter, a parametric and a Gaussian approach. The

existing pervasive in the visual of Kalman Filter

is enhanced and overcome by non Gaussian

approach involved condensation Filter algorithm.

The retrieval of an image during its motion is

revealed by Mean Shift Tracker. The extension of

this paper can be revealed in features of stable

convergence and speed.

REFERENCES

[1]Y. Boykov and D. Huttenlocher. A new Bayesian

approach to object recognition. In Proc. of IEEE CVPR,

1999.

[2] R. Rosales, S. Sclaroff, Learning body pose using

specialized maps, in: T.G. Dietterich et al. (Eds.),

Advances in Neural Information Processing Systems 14,

MIT Press, 2002, pp. 1263-1270.

[3] Blake,A. and Israd, M. (1997). Condensation-

conditional density propagation for visual tracking.,

International Journal of Computer Vision, v.29 n.1, p.5-28,

Aug. 1998.

[4] MacCormick, J. Blake, A. A probabilistic exclusion

principle for tracking multiple objects. The Proceedings of

the Seventh IEEE International Conference on Computer

Vision 1999.

[5] Y.Wu and T. Huang, “A co-inference approach to

robust tracking,” in Proc. 8th Intl. Conf. on

ComputerVision, Vancouver, Canada, volume II, 2001, pp.

26–33.

[6] B. Li and R. Chellappa, “Simultaneous tracking and

verification via sequential posterior estimation,” in Proc.

IEEE Conf. on Computer Vision and Pattern Recognition,

Hilton Head, SC, volume II, 2000, pp. 110–117.

[7] Yilmaz, A “Object Tracking by Asymmetric Kernel

Mean Shift with Automatic Scale and Orientation

Selection” . IEEE Conference on Computer Vision and

Pattern Recognition, 2007. CVPR '07

Proceedings of the Third National Conference on RTICT 2010 Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

128

DEVELOPMENT OF DATA ACQUSITION SYSTEM USING

SERIAL COMMUNICATION

1K.Eunice, S. Smys

2

1Student of ECE Department,

2 Assistant professor, ECE Department, Karunya university, Coimbatore.

Eunicekarasala, smys @karunya.edu

Abstract: This system is used for real time capture

and display of parameters of precision optical

instrument. The processing, capture and display can

be accomplished by three blocks; they are optical

sensor package, digital processing and display. The

sensor package contains optical sensor, FPGA

(Field Programmable Gate Array) card etc. The sensor measures the angular rotation with great

accuracy and precision. All the parameters that are

required by the sensor are processed in FPGA and

DSP, later these parameters are communicated

through the serial port. The various parameters

obtained from sensor are communicated through

the RS232 serial port through dedicated FPGA.

Introduction:

The basic principle of operation is, a

simple optical sensor that measures any rotation

about its sensitive axis. This implies that the orientation in inertial space will be known at all

times. A sensor converts a physical property or

phenomenon into a corresponding measurable

electrical signal, such as voltage, current, frequency

etc. These electrical parameters has to be known to

the external world. Hence, a data acquisition

system is required. The components of data

acquisition system include appropriate sensors that

convert any measurement parameter to an electrical

signal, which is acquired by data acquisition

hardware. Here, an RS232 serial port of a PC is used to send a command and receive the data from

the sensor package. Hence a software and GUI has

to be developed using LabVIEW.The project is

undertaken basically for designing a software

system for real time capture and display of

parameters of precision optical instrument. This is

based on modelling and simulation of FPGA based

data acquisition software scheme for precision

angle measurement. The document has been

proposed to provide an over view of the data that

has been acquired through the sensor all the

parameters that are required by the sensor are processed in FPGA and DSP, later these parameters

are communicated through the serial port.

2. Laser based rotation sensor:

Fig 1 – block diagram of Laser based rotation

sensor A Sensor is a device that responds to physical

stimuli (such as heat, light, sound, pressure,

magnetism, motion) and transmits the resulting

signal or data for providing measurement,

operating a control or both. From the sensor get the

data to be measure

The sensor electronics can be classified into the

following parts depending on their function:

Now the data obtained from the sensor is connected

to the FPGA to program and control using serial

communication. Data acquisition systems, as the name implies, are products and/or processes used to

collect information to document or analyze some

phenomenon. In the simplest form, a technician

logging the temperature of an oven on a piece of

paper is performing data acquisition. As technology

has progressed, this type of process has been

simplified and made more accurate, versatile, and

reliable through electronic equipment. Equipment

ranges from simple recorders to sophisticated

computer systems. Data acquisition products serve

as a focal point in a system, tying together a wide

variety of products, such as sensors that indicate temperature, flow, level, or pressure. Some

common data acquisition terms are shown below.

Data acquisition (abbreviated DAQ) is the

process of sampling of real world physical

conditions and conversion of the resulting samples

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

129

into digital numeric values that can be manipulated

by a computer.

3.Data acquisition system:

Data acquisition and data acquisition systems

(abbreviated with the acronym DAS) typically involves the conversion of analogy waveforms into

digital values for processing. The components of

data acquisition systems include:

Sensors that convert physical parameters

to electrical signals.

Signal conditioning circuitry to convert

sensor signals into a form that can be

converted to digital values.

Analog-to-digital converters, which

convert conditioned sensor signals to

digital values. When defining a Data Acquisition System, this

can be can be accurately described by explaining

what a Data Acquisition System does – not what it

―is.‖ A Data Acquisition System catches or

captures data about an actual system and stores that

information in a format that can be easily

retrievable for purposes of engineering or scientific

review and analysis. Another requirement of a Data

Acquisition System should be that it captures

information programmatically or automatically – in

other words, without any hands-on human

intervention or guidance. Generally speaking, there are seven elements or functions in a Data

Acquisition System. The seven elements/functions

are (in no particular order) data collection,

measurement, timing and triggering, a real-time

clock, system control, data communication and data

archiving. All seven elements must be in place for a

structure to be considered a Data Acquisition

System. If only a number of these elements are part

of the system, the module could be defined as a

component of a Data Acquisition System. If a

system has all seven elements along with additional features, it is probably a larger system with a Data

Acquisition System being part of the larger

structure.The actual components or elements of a

Data Acquisition System to perform the seven

essential functions are critical to the efficiency of

the system. There must be a series of sensors as

inputs to a Data Acquisition Board; in addition,

there must be a trigger to synchronize the sensor

inputs (the data stream), as well as a control for the

Data Acquisition Board. Between the Data

Acquisition Board and the processor of the system and system clock, a data communications bus (I/O)

is also required. While the data is being stored real-

time, the analysis and review of the information is

performed after data is gathered. By definition,

information cannot be analyzed in real-time,

otherwise data events will be missed or overlooked.

The Data Acquisition System must collect, sort,

catalog and store data to be reviewed and analyzed

in a meticulous manner.

The purpose of data acquisition is to measure an

electrical or physical phenomenon such as voltage,

current, temperature, pressure, or sound. PC-based data acquisition uses a combination of modular

hardware, application software, and a computer to

take measurements. While each data acquisition

system is defined by its application requirements,

every system shares a common goal of acquiring,

analyzing, and presenting information. Data

acquisition systems incorporate signals, sensors,

actuators, signal conditioning, data acquisition

devices, and application software.

Serial Communication Data Acquisition

Systems:

Serial communication data acquisition

systems are a good choice when the

measurement needs to be made at a location

which is distant from the computer. There are

several different communication standards,

RS232 is the most common but only supports

transmission distances up to 50 feet. RS485 is superior to RS485 and supports

transmission distances to 5,000 feet.

Fig

3

Figure 2 data acquisition system

RS232

A standard for serial communications found in many data acquistion systems.

RS232 is the most common serial

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

130

communication, however, it is somewhat

limited in that it only supports

communication to one device connected to

the bus at a time and it only supports

transmission distances up to 50 feet.

RS485 A standard for serial communications

found in many data acquistion systems.

RS485 is not as popular as RS232,

however, it is more flexible in that it

supports communication to more than one

device on the bus at a time and supports

transmission distances of approximately

5,000 feet.

RS232 is a popular communications protocol for

connecting modems and data acquisition devices to computers. RS232 devices can be plugged straight

into the computer's serial port (also known as the

COM or Comms port). Examples of data

acquisition devices include GPS receivers,

electronic balances, data loggers, temperature

interfaces and other measurement instruments.

RS232 Software:

To obtain data from your RS232 instruments and

display it on your PC you need some software.

Version 4.3 of the Windmill RS232 software is

now free from their web site. They also offer free

serial trouble-shooting software. The RS232 Standard:

RS stands for recommended standard. In the 60's a

standards committee now known as the Electronic

Industries Association developed an interface to

connect computer terminals to modems. Over the

years this has been updated: the most commonly

used version of the standard is RS232C (sometimes

known as EIA232); the most recent is RS232E. The

standard defines the electrical and mechanical

characteristics of the connection - including the

function of the signals and handshake pins, the voltage levels and maximum bit.RS232 standard

was created for just one specific situation and the

difficulties come when it is used for something

else. The standard was defined to connect

computers to modems. Any other use is outside of

the standard. The authors of the standard had in

mind the situation below:

The standard defines how computers ( it calls them

Data Terminal Equipment or DTEs) connect to

modems ( it calls them Data Communication

Equipment or DCEs). The standard says that

computers should be fitted with a 25 way plug

whilst modems should have a 25 way D socket. The interconnecting lead between a computer and a

modem should be simply pin1—pin1, pin2—pin2,

etc. The main signals and their direction of flow are

described below. It is important to note that a signal

which is an output from a computer is an input to a

modem and vice versa. This means that you can

never tell from the signal name alone whether it is an input or an output from a particular piece of

equipment. Also, instead of being a DCE device, a

data acquisition device might be configured as

DTE. In this case you need an adaptor or the

RS232 cable wired differently to normal. When the

PC is connected to a DTE instrument, some of the

cable wires must cross over.

TXD Transmitted Data, Pin 2 of 25 way D

This is the serial encoded data sent from a

computer to a modem to be transmitted over the telephone line.

RXD Received Data, Pin 3 of 25 way D

This is the serial encoded data received by a

computer from a modem which has in turn received

it over the telephone line.

DSR Data Set Ready, Pin 6 of 25 way D

This should be set true by a modem whenever it is

powered on. It can be read by the computer to

determine that the modem is on line.

DTR Data Terminal Ready, Pin 20 of 25 way D

This should be set true by a computer whenever it

is powered on. It can be read by the modem to determine that the computer is on line.

RTS Request to Send, Pin 4 of 25 way D

This is set true by a computer when it wishes to

transmit data.

CTS Clear To Send, Pin 5 of 25 Way D

This is set true by a modem to allow the computer

to transmit data. The standard envisaged that when

a computer wished to transmit data it would set its

RTS. The local modem would then arbitrate with

the distant modem for use of the telephone line. If

it succeeded it would set CTS and the computer would transmit data. The distant modem would use

its CTS to prevent any transmission by the distant

computer.

DCD Data Carrier Detect, Pin 8 of 25 Way D

This is set true by a modem when it detects the data

carrier signal on the telephone line..

PC Serial Ports

A nine pin D plug has become the standard fitting

for the serial ports of PCs, although it's nothing to

do with the RS232 standard. The pin connections

used are: Pin Direction Signal

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

131

1 Input DCD Data Carrier Detect

2 Input RXD Received Data

3 Output TXD Transmitted Data

4 Output DTR Data Terminal Ready

5 Signal Ground 6 Input DSR Data Set Ready

7 Output RTS Request To Send

8 Input CTS Clear To Send

9 Input RI Ring Indicator

Conclusion:

The speed of RS232 communications is expressed

in Baud. The unit is named after Jean Maurice-

Emile Baudot (1845-1903), a French telegraph

engineer and the inventor of the first teleprompter. It was proposed at the International Telegraph

Conference of 1927. The maximum speed,

according to the standard, is 20000 Baud. The

length of the cable also plays a part in maximum

speed. The longer the cable, the greater the cable's

capacitance and the slower the speed at which you

can obtain accurate results. A large capacitance

means voltage changes on one signal wire may be

transmitted to an adjacent signal wire. Fifty feet is

commonly quoted as the maximum distance, but

this is not specified in the standard. We generally

recommend a maximum distance of 50 metres, but this depends on the type of hardware connecting

and characteristics of the cable.

References:

[1] A. Deshpande, M. Garofalakis, and R. Rastogi.

Independence is Good: Dependency-Based

Histogram Synopses for High-Dimensional Data.

In SIGMOD, May 2001.

[2] A. Desphande, C. Guestrin, W. Hong, and S.

Madden. Exploiting correlated attributes in

acquisitional query processing. Technical report, Intel-Research, Berkeley, 2004.

[3] L. Getoor, B. Taskar, and D. Koller. Selectivity

estimation using probabilistic models. In SIGMOD,

May 2001.

[4] P. B. Gibbons. Distinct sampling for highly-

accurate answers to distinct

values queries and event reports. In Proc. of VLDB,

Sept 2001.

[5] P. B. Gibbons and M. Garofalakis.

Approximate query processing: Taming the

terabytes (tutorial), September 2001.

[6] P. B. Gibbons and Y. Matias. New sampling-based summary statistics for improving

approximate query answers. In SIGMOD, 1998.

[7] J. M. Hellerstein, R. Avnur, A. Chou, C.

Hidber, C. Olston, V. Raman, T. Roth, and P. J.

Haas. Interactive data analysis with CONTROL.

IEEE Computer, 32(8), August 1999.

Proceedings of the Third National Conference on RTICT 2010 Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

131

Analysis of Health Care Data Using Different

Classifiers.

Soby Abraham,

M.Tech.–CSE IVth Sem, University School of Information Technology, GGS Indraprastha University, Delhi

[email protected]

Abstract- Data mining is an interesting field of

research whose major objective is to acquire

knowledge from large amounts of data. With

advances in heath care related research, there is a

wealth of data available. However, there is a lack of

effective analytical tools to discover hidden and

meaningful patterns and trends in data, which is

essential for any research.

Data mining is used in fields like business, science,

engineering etc. the various data mining

functionalities that can be applied in these fields

include characterization and discrimination,

frequent pattern mining, classification and

prediction.

In recent years, human immune-deficiency virus

(HIV) related illnesses have become a threat to the

modern world. Researchers all over world, including

India, are trying hard to find suitable answer to this

and this led to lots of research in the field.

Therefore, a tool which can process data in

meaningful way is the need the time.

In this study, we briefly examine the potential use

of classification based data mining techniques such

as Naive Bayesian classifier, J48 and Zeror classifier

to massive volume of health care data. Data mining

fondly called patterns analysis on large sets of data.

Keywords- decision tree, classification, Human

immunodeficiency virus, antiretroviral therapy

I. INTRODUCTION

Data mining is the analysis of (often large)

observational data sets to find unsuspected

relationships and to summarize the data in

novel ways that are both understandable and

useful

to the data owner.

Data mining, also known as knowledge

discovery in databases [1] is defined as the

extraction of implicit, previously unknown and

132

potentially useful information from data. Data

mining starts with the raw data, which usually

takes the form of simulation data, observed

signals, or images. These data are preprocessed

using various techniques such as sampling,

analysis, feature extraction, and normalization.

It is an interactive and iterative process

involving data pre-processing, search for

patterns, knowledge evaluation, and possible

refinement of the process based on input from

domain experts. Data mining tools can answer

business questions that traditionally were too

time consuming to resolve [2]. Data mining is

the next step in the process of analyzing data.

Instead of getting queries on standard or user

specified relationships, data mining goes a step

farther by finding meaningful relationships in

data.

There are several data mining techniques, such

as, Decision Trees, Artificial Neural Networks

(ANNs), Clustering, Naïve Bayes, Association

Rules, Time series etc[2] but decision tree is one

of the more powerful technique that produce

interpretable result and thus widely used in

clinical purpose.

Hence, decision tree may be considered mining

knowledge from large amounts of data since it

involves knowledge extraction, as well as

data/pattern analysis in tree diagrams [4].

Considering decision trees, classification is the

most common data mining task and it consists

of examining the features of a newly presented

object in order to assign it to one of a

predefined set of classes. Classification deals

with discrete outcomes, estimation deals with

continuously-valued outcomes.

In this study, our objective are to: (1) present an

evaluation of techniques such as Naive Bayes

classifier, Zeror and J48 classifiers which are

used to predict the occurrence of route of

transmission based on treatment history of HIV

patients. (2) demonstrate that data mining

method can yield valuable new knowledge and

pattern related to the HIV patient; (3) assesses

the utilization of healthcare resources and

demonstrate the socioeconomic, demographic

and medical histories of patient.

Organization of paper is as follows; in section II

we give a brief explanation of the data mining

and knowledge discovery. In section III we give a

brief explanation of the data mining techniques.

In section IV we analyze of HIV patients using

decision trees as well as different classifiers.

Result and conclusion are discussed in Section V

and section VI.

II. DATA MINING AND KNOWLEDGE DISCOVERY

Knowledge plays a vital role in the lives of all

human beings and important pieces of

knowledge hold the key to our future on this

planet. With out knowledge it is not feasible to

proceed further and gain success in any

discipline. In today’s world, there is a

continuous search for data from which

important pieces of knowledge can be

extracted. Such pieces of knowledge are

required by both users and non-users of data.

With the enormous amount of data stored in

files, databases, and other repositories, it is

increasingly important, if not necessary, to

develop powerful means for analysis and

perhaps interpretation of such data and for the

extraction of interesting knowledge that could

help in decision-making.

Data mining is also known as Knowledge

Discovery in Databases refers to the nontrivial

extraction of implicit, previously unknown and

potentially useful information from data in

databases. While data mining and knowledge

discovery in databases are frequently treated as

synonyms, data mining is actually part of the

knowledge discovery process [6].

133

Figure 1: Knowledge [6] Discovery Process.

The Knowledge Discovery in Databases process

comprises of a few steps leading from raw

data collections to some form of new

knowledge. The iterative process consists of the

following steps.

1) Data Cleaning. it is a phase in which noise

data and irrelevant data are removed from the

collection.

2) Data Integration. in this stage, multiple

data sources may be combined in a common

source.

3) Data selection. at this step, the data relevant

to the analysis is decided on and retrieved from

the data collection.

4) Data transformation. also known as data

consolidation, it is a phase in which the selected

data is transformed into forms appropriate for

the mining procedure.

5) Data mining. it is the crucial step in which

clever techniques are applied to extract patterns

potentially useful.

6) Pattern evaluation. in this step, strictly

interesting patterns representing knowledge are

identified based on given measures.

7) Knowledge representation. is the final phase

in which the discovered knowledge is visually

represented to the user. It helps users to

understand and interpret the data mining

results.

It is common to combine some above

mentioned steps together. For example data

cleaning and integration can be performed

together as a pre-processing phase to generate

data warehouse. Data selection and data

transformation can also combine where

consolidation of the data is the result of

selection. Knowledge discovery is an iterative

process. Once the discovered knowledge is

presented to the user, the evaluation measures

can be enhanced, the mining can further

refined, new data can be selected or further

transformed, or new data sources can be

integrated in order to get more appropriate

result [6].

III. DATA MINING TECHNIQUES

Decision Trees

Decision tree is an important technique

in machine learning and it is used extensively

in data mining. Decision trees are able to

produce human-readable descriptions of trends

in the underlying relationships of a dataset and

can be used for classification and prediction

tasks.

Flat files

Databases

Cleaning &

Integration

Evaluation &

Presentation

Knowledge

Knowledge

Discovery

Selection &

Transformation

134

This technique has been used successfully in

many different areas, such as medical diagnosis,

plant classification, and customer marketing

strategies etc.

A decision tree is a predictive machine-learning

model that decides the target value of a new

sample based on various attribute values of the

available data. The internal nodes of a decision

tree denote the different attributes, the

branches between the nodes tell us the possible

values that these attributes can have in the

observed samples, while the terminal nodes tell

us the final value (classification) of the

dependent variable.

The principle idea of a decision tree is to split

data recursively into subsets so that each subset

contains more or less homogeneous states of

target variable (predictable attribute). At each

split in the tree, all input attributes are

evaluated for their impact on the predictable

attribute. When this recursive process is

completed, a decision tree is formed.

There are few advantages of using decision

trees over using other data mining algorithms,

for example , decision trees are quick to build

and easy to interpret.

A. Naive Bayes Classifier

Naive Bayes classifier is based [7] on Bayes’

theorem and the theorem of total probability. A

classifier that suffers relatively little from high

dimensionality is the Naive Bayes classifier.

These classifiers require an optimization

procedure that is computationally unacceptable

in settings in which the training data changes

continuously. In simple terms, a Naive Bayes

classifier assumes the presence or absence of a

feature of a class is unrelated to the presence or

absence of any other feature.

The probability model for a classifier is a

conditional model is P(C/F1,…,Fn) over a

dependent class variable C with a small number

of outcomes or classes, conditional on several

feature variables F1 through Fn. Using Bayes’

theorem , we write

p(C/F1,…,Fn) =

Depending on the nature of the probability

model, Naive Bayes classifiers can be trained

very efficiently in a supervised learning setting.

Naïve bayes method uses the method of

maximum likely hood.

An advantage of the Naive Bayes classifier is

that it requires a small amount of training data

to estimate the parameters necessary for

classification. Because independent variables

are assumed, only the variances of the variables

for each class need to be determined and not

the entire covariance matrix.

B. J48 Classifier

J48 classifier is [8] a simple C4.5 decision tree used

for classification. It creates a binary tree. This is a

standard algorithm that is widely used in

machine learning. In order to classify a new item

using J48, firstly create a decision tree based on

the attribute values of the available training

data. So, whenever it encounters a set of items,

it identifies the attribute that discriminates the

various instances most clearly. These features

tell us about the data instances so that we can

p(C) p(F1,…,Fn/C)

p(F1,…,Fn)

135

classify them the best which have the highest

information gain. If there is any value for which

there is no ambiguity, that is, for which the data

instances falling within its category have the

same value for the target variable, then we

terminate that branch and assign to it the target

value that we have obtained.

For the other cases, we then look for another

attribute that gives us the highest information

gain. Hence we continue in this manner until we

either get a clear decision of what combination

of attributes gives us a particular target value, or

we run out of attributes.

C. ZeroR Classifier

ZeroR classifier [8] simply predicts the majority

class in the training data. Although it makes

little sense to use this scheme for prediction, it

can be useful for determining a baseline

performance as a benchmark for other learning

schemes. ZeroR classifier is also used to

compare the performances. ZeroR is frequently

used as a baseline for evaluating other machine

learning algorithms. ZeroR always predict the

most common classification. For example, if

most of training data got low coverage, ZeroR

will predict low coverage on all inputs. ZeroR

provides a useful indication of a predictor’s

worst performance, since a nonrandom

prediction scheme should do better.

IV. ANALYSIS

In this study, data gathered from ART Clinic, was

analyzed using different classifiers such as Naive

Bayesian classifier, J48 classifier and ZeroR

classifier. Significant [9] research efforts have

been undertaken in investigating the association

between HIV and ART. HIV database is

considered as one of the deadly disease all over

the world.

Following tests and procedures may be used:

Physical exam and history.

Complete blood count

Stages of infection.

Chest X-ray.

A. Methods of Data Collection.

We collected the data from different

ART systems.

We derived a dataset from HIV

database that include 1054 enrolled patient out

of which we have considered only 672 unique

patients because rest of patients is defaulter.

B. Data Cleaning

Real world data, like data acquired

from ART Clinic tend to incomplete, inconsistent

and noisy. Data cleaning routine attempt to fill

missing values, smooth out noise and correct

inconsistencies in the data.

C. Missing Values.

Most datasets contain missing values.

There are number of causes for missing

136

data. Many methods were applied to solve

this,

Fill the missing values manually.

Replace the missing values with

the most popular value.

Use a global consistent.

Figure 2: Missing values.

D. Noisy Data

Noise is a random error or variance in a

measured variable. Many methods for data

smoothing are also methods for data reduction

involving discretization.

E. Data Integration.

Data integration is the merging of data

from multiple data sources. The data may also

need to be transformed into forms appropriate

for mining. Careful integration of data from

multiple sources can help reduce and avoid

redundancies and inconsistencies. It helps to

improve the accuracy and speed of the mining

process.

F. Data Selection.

Attribute selection reduces the data set

size by removing irrelevant or redundant

attributes. The goal is to find a minimum set of

attributes such that the resulting probability

distribution is close to the original distribution

obtained using all the attributes. In this paper,

HIV care was the main aim, so data concerning

the diagnosis of HIV was carefully selected from

the over all data sets.

Main focus on following data.

Figure 3: Attribute selection.

V. RESULTS.

In this study, we analyze data on Route of

transmission in the HIV database and investigate

their association with patient’s history of ART.

Classifiers like Naïve Bayes, J48 and applied to

the database for identifying the patterns.

In HIV database we have used certain codes for

route of transmission and the code is given

table1.

TABLE I

137

Code for Route of Transmission

Type of transmission Value

Heterosexual 1

Mother-to-Child 2

Blood Transfusion 3

IVD 4

Surgical Instrument 5

Unsafe Injection 6

Professional needle

stick Injury

7

Professional needle

stick Injury

8

Unknown 9

A. Analysis using Naive Bayes Classifier

When we analyzed the data using Naive Bayes

classifier, output is as shown below.

Figure 4: Naive Bayes Classifier.

Receiver Operator Characteristic areas are

commonly used to present results for binary

decision problems in machine learning. It is used

to find out the accuracy of the test. ROC areas

can also be calculated from clinical prediction

rules. Using Naive Bayes ROC area = 0.991, is

high for class2 compared to other classes, which

is mother to child. Mean absolute error in this

case is 0.0491.

B. Analysis using J48 Classifier.

When we analyzed the data using J48 Classifier,

output is as shown below.

138

Figure 5: J48 Classifier.

Using J48 Classifier, ROC area = 0.5 for all

classes. Mean absolute error in this case is

0.0728.

C. Analysis using Zeror Classifier.

ZeroR is a simple classifier that always returns a

predefined answer regardless of the test

instance; it assigns a new instance to the most

frequent class of training instances. The ZeroR

model simply classifies every data item in the

same class.

When we analyzed the data using J48 Classifier,

output is as shown below.

Figure 6: ZeroR Classifier.

Using ZeroR Classifier, ROC area = 0.5 for all

classes, which is same that we have obtained

from J48 classifier, but the mean absolute error

is lower in this case. Mean absolute error in this

case is 0.0744.

VI. CONCLUSION

In this paper we have described classification

techniques for healthcare datasets. We have

used different data mining classifiers such as

Naive Bayes, J48 and ZeroR to find out ROC area

and mean absolute error. Mean absolute error

is low in the case of Naive Bayes classifier. This

paper leads the study of data mining in health

care datasets.

139

VII. ACKNOWLEDGMENTS

The author will like to thank Dr. Anjana Gosain,

University School of Information Technology,

Guru Gobind Singh Indraprastha University for

the guidance during the course of study.

VIII. REFERENCES

[1] J. Han and M. Kamber – Data Mining: Concepts

and techniques; Second Edition; Morgan

Kaufmann, 2006.

[2] I.H. Witten, and E. Frank – Data Mining: Practical

Machine Learning Tools and Techniques; Second

Edition; Morgan Kaufmann, 2005.

[3] Lindsay, Clark. "Data Mining." Online. Internet. 3

Oct. 1997.

[4] http://www.twocrows.com.

[5] Osmar R. Zaļane- Principles of Knowledge

Discovery in Databases, 1999.

[6] A. Kumar, “Applications of data mining”, In

proceedings of National Conference on

Information Technology: Present Practices and

challenges, 2007.

[7] P.Bhargavi and Dr.S.Jyothi- Applying Naive

Bayes Data Mining Technique for Classification

of Agricultural Land Soils.

[8] http://wikipedia.org/

[9] Harleen Kaur and Siri Krishan Wasan: Empirical

Study on Applications of Data Mining Techniques

in Healthcare.

[10] Mark Hall, Eibe Frank, Geoffrey Holmes, Bernhard

Pfahringer, Peter Reutemann, Ian H.

Witten: The WEKA Data Mining Software: An

Update

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

141

Merin Jojo (M.E Computer Science and Engineering) 1,

V.Venkatesakumar M.E., 2

1Student, Department

of CSE, Anna University Coimbatore.

[email protected] 2

Lecturer, Department of CSE, Anna University Coimbatore.

Abstract- In this paper, we deals with distributed data mining

in peer-to-peer networks with the help of generic local

algorithm. For large networks of computers or wireless

sensors, each of the components or peers has some data about

the overall state of the system. System’s functionalities are

based on modeling the global state. The state of the system is

constantly changing, so it is necessary to keep the models up-

to-date. The data modeling and communication cost may be

very costly due to the scale of the system. In this paper we

describe a two step approach for dealing with these costs.

First, we describe a highly efficient local algorithm which can be used to monitor a wide class of data mining models. Then,

we use this algorithm as a feedback loop for the monitoring of

complex functions.

Keywords- Distributed data mining, Local Algorithm, Peer-

to-peer, Mean monitoring.

1. INTRODUCTON

Large distributed systems or in wireless sensor networks, produce, store and process huge amounts of status

data as a main part of their daily operation and often the need

to model the data that is distributed over the entire system. But

centralizing the data is a costly approach. The internets,

intranets, local area networks, ad-hoc wireless networks and

sensor networks are some of the examples .Mining in such

environments naturally calls for proper utilization of these

distributed resources. When data and system changes are

frequent, designers should update the model frequently and

risk wasting resources on insignificant changes, causes

inaccuracy and resulting in system degradation.

At least three algorithmic approaches can be

followed in order to address this situation. The periodic

approach is to simply rebuild the model from time to time but

waste the resources. The incremental approach is to update the

model with every change of the data and it might be

inaccurate. Last, the reactive approach is to monitor the

change and rebuild the model only when it no longer suits the

data. The benefit of the incremental approach is that its

accuracy can be optimal. If monitoring is done efficiently and

accurately, then the reactive approach can be applied to many

different data mining algorithm at low costs.

A local algorithm is one in which the complexity of

computing the result does not directly depend on the number

of participants and mainly based on distributed systems.

Instead, each processor usually computes the result using

information gathered from just a few nearby neighbors. Local

algorithms are in-network algorithms in which data is

decentralized but rather computation is performed by the peers of the network. At the heart of a local algorithm there is a data

dependent criteria dictating when nodes can avoid sending

updates to their neighbors. In a local algorithm, it often

happens that the overhead is independent of the size of the

system. Because of this reason, local algorithms exhibit high

scalability. The dependence on the criteria for avoiding

sending messages also makes local algorithms inherently

incremental. Specifically, if the data changes in a way that

does not violate the criteria, then the algorithm adjusts to the

change without sending any message. These local algorithms

were developed, for a large selection of data modeling problems. Problems such as association rule mining, facility

location, outlier detection, L2 norm monitoring classification

and multivariate regression etc can be solved.

In this work first, we generalize a common theorem

underlying the local algorithms. Next, we describe a generic

algorithm, relying on the said generalized theorem, which can

be used to compute arbitrarily complex functions of the

average of the data in a distributed system. Then, we describe

a general framework for monitoring, and consequent reactive

updating of any model of horizontally distributed data.

Finally, we describe the application of this framework for the problem of providing a mean monitoring which is a good

approximation of the k-means clustering of data distributed

over a large distributed system. Some simulation results are

also added at the last section. The last section deals with

conclusion and future works.

2. NOTATIONS AND ASSUMPTIONS

Mining In Distributed System Using Generic

Local Algorithm

142

In this section we discuss the notations, assumptions

and problem definition which will be used throughout the

paper. The main idea of the algorithm is to have peers

accumulate sets of input vectors from their neighbors. We

show that under certain conditions on the accumulated vectors

a peer can stop sending vectors to its neighbors long before it collects all input vectors.

2.1 Notations

Let V = p1. . . pn be a set of peers connected to

one another via an underlying communication infrastructure.

The set of peers with which pi can directly communicate, Ni, and is known to pi. Ni contains pi and at least one more peer. pi

is given a time varying set of input vectors in IRd. Peers

communicate with one another by sending sets of input

vectors X i,j the latest set of vectors sent by peer pi to pj . We

further define four sets of vectors that are central to our

algorithm.

Definition 2.1: The knowledge of pi, is Ki =∪pj ε Ni Xi,j

Definition 2.2: The agreement of pi and any neighbor pj ∈ Ni

is A i,j = X i,j ∪ Xj,i.

Definition 2.3: The withheld knowledge of pi with respect to a

neighbor pj is the subtraction of the agreement from the

knowledge Wi,j = Ki \ A i,j .

Definition 2.4: The global input is the set of all inputs

G = ∪ pj ∈ V X i,i

We are interested in inducing functions defined on G.

Since G is not available at any peer, we derive conditions on

K, A and W which will allow us to learn the function on G. We

denote a given cover RF respective of F if all regions except

the tie region, F is constant.

2.2 Assumptions

We make the following assumptions:

Assumption 2.1.Communication is reliable.

Assumption 2.2.Communication takes place over a spanning

communication tree.

Assumption 2.3. Peers are notified on changes in their own

data xi, and in the set of their neighbors Ni.

Assumption 2.4. Input vectors are unique.

Assumption 2.5. A respective cover RF can be precomputed

for F.

Simple approaches such as piggybacking message

acknowledgement can be implemented in even the most

demanding scenarios mainly for wireless sensor networks.

3. PROBLEM DEFINITION

Problem definition: Given a function F, a spanning network

tree G (V, E) which might change with time, and a set of time

varying input vectors X i,i at every pi ε V , the problem is to

compute the value of F over the average of the input vectors

G. While the problem definition is limited to averages of data

it can be extended to weighted averages by simulation. If a

certain input vector needs to be given an integer weight ω then

ω peers can be simulated inside the peer that has that vector

and each be given that input vector. Likewise, if it is desired

that the average be taken only over those inputs which comply

with some selection criteria then each peer can apply that

criteria to X i,i apriori and then start off with the filtered data. Thus, the definition is quite conclusive.

4. MAIN THEOREM

The main theorem of this paper lay the background

for a local algorithm which guarantees eventual correctness in

the computation of a wide range of ordinal functions. The

theorem generalizes the local stopping rule described in [1] by

describing a condition which bounds the whereabouts of the

global average vector in Rd depending on the Ki, Ai,j and Wi,j

of each peer pi.

Theorem 3.1: [Main Theorem] Let G(V,E) be a spanning

tree in which V is a set of peers and let X i,i be the input of pi,

Ki be its knowledge, and A i,j and Wi,j be its agreement and

withheld knowledge with respect to a neighbor pj ∈ Ni as

defined in the previous section. Let R ⊆ IRd be any convex

region. If at a given time no messages traverse the network

and for all pi and pj ∈ Ni Ki, A i,j ∈ R and either W i,j = Ø or

W i,j ∈ R as well, then G ∈ R.

Proof: Consider a communication graph G(V,E) in which for

some convex R and every pi and pj such that pj ∈ Ni ,it holds

that Ki, Ai.j∈ R and either Wi,j = Ø or W i,j ∈ R as well.

Assume an arbitrary leaf pi is eliminated and all of the vectors

in W i,j are added to its sole neighbor pj . The new knowledge

of pj is K’j = Kj ∪ W i,j . Since by definition Kj ∩ Wi,j = Ø ,

the average vector of the new knowledge of pj , K′ j , can be

rewritten as Kj∪ Wi,j = α·Kj+(1−α)·W i,j for some α ∈ [0, 1].

Since R is convex, it follows from Kj ,W i,j ∈ R that K′ j ∈ R

too. Now, consider the change in the withheld knowledge of pj

with respect to any other neighbor pk ∈ Nj resulting from

sending such a message. The new W′ j,k = W i,j ∪ W j,k.

Again, since Wi,j ∩ W j,k = Ø and since R is convex it

follows from Wi,,j ,W j,k ∈ R that W′ j,k ∈ R as well. Finally,

notice the agreements of pj with any neighbor pk except pi do

not change as a result of such message. Hence, following

143

elimination of pi we have a communication tree with one less

peer in which the same conditions still apply to every

remaining peer and its neighbors. Proceeding with elimination

we can reach a tree with just one peer p1, still assured that K1

∈ R. Moreover, since no input vector was lost at any step of

the elimination K1 = G. Thus, we have that under the said

conditions G ∈ R.

Figure 1: The graphical allignment of three peers having

weight, agreement and withheld knowledge are gven in

figure

The above theorem is exemplified in Figure 1. Three peers are

shown, each with a drawing of its knowledge, it agreement

with its neighbor or neighbors, and the withheld knowledge.

Notice the agreement A 1, 2 drawn for p1 is identical to A2, 1 at

p2. For graphical simplicity we assume all of the vectors have

the same weight and avoid expressing it. We also depict the

withheld knowledge vectors twice once as a subtraction of the

agreement from the knowledge by using a dotted line . If the

position of the three peers’ data is considered vis-à-vis the

circular region then the conditions of Theorem 3.1 hold. Now, assume what would happen when peer p1 is eliminated. This

would mean that all of the knowledge it withholds from p2 is

added to K2 and to W2,3. Since we assumed |W1,2| = |K2| = 1 the

result is simply the averaging of the previous K2 and W1,2.

Notice both these vectors remain in the circular region. Lastly,

as p2 is eliminated as well, W2,3 which now also includes W1,2

is blended into the knowledge of p3. Thus, K3 becomes equal

to G. However, the same argument, as applied in the

elimination of p1, assures the new K3 is in the circular region

as well

To see the relation of Theorem 3.1 to the previous the

Majority-Rule algorithm [1], one can restate the majority

voting problem as deciding whether the average of zero-one

votes is in the segment [0, λ) or the segment [λ, 1]. Both

segments are convex, and the algorithm only stops if for all

peers the knowledge is further away from λ than the agreement – which is another way to say the knowledge, the

agreement, and the withheld data are all in the same convex

region. Therefore, Theorem 3.1 generalizes the basic stopping

rule of Majority-Rule to any convex region in Rd. Two more

theorems are used for the proper working explanation of the

generic algorithm.

5. GENERIC ALGORITHM AND ITS USAGE

The generic algorithm, depicted in Algorithm 1,

receives as input the function F, a respective cover RF, and a

constant, L, whose function is explained below. Each peer pi

outputs, at every given time, the value of F based on its

knowledge Ki.

Algorithm 1 Generic Local Algorithm

Input of peer pi: F, RF = R1,R2, . . . , T , L, Xi,i, and Ni

Ad hoc output of peer pi: F (Ki)

Data structure for pi : For each pj ∈ Ni X i,j , |X i,j |, Xj,i |Xi,j |,

last_ message

Initialization: last message ← −∞

On receiving a message X, |X| from pj

– X j,i ← X, |X, j i| ← | X |

On change in Xi,i, Ni, Ki or |Ki|: call OnChange()

OnChange( )

For each pj ∈ Ni:

If one of the following conditions occurs:

1. RF (Ki)= T and either Ai,j ≠ Ki or |Ai,j | ≠ |Ki|

2. |Wi,j | = 0 and Ai,j ≠ Ki

3. Ai,j ∈ RF (Ki) or Wi,j ∈ RF (Ki)

then call SendMessage(pj)

SendMessage(pj):

If time ( ) – last_ message ≥ L If RF( Ki) = T then the new Xi,j and |Xi,j | are Wi,j and |Wi,j |,

respectively

Otherwise compute new Xi,j and |Xi,j | such that Ai,j ∈ RF

(Ki) and either Wi,j ∈ RF( Ki) or |Wi,j | = 0

last message ← time ( )

Send Xi,j , |Xi,j | to pj

Else Wait L − (time ( ) − last message) time units and then

call OnChange( )

The algorithm is event driven. Events could be one of

the following: a message from a neighbor peer, a change in the

set of neighbors due to failure or recovery, a change in the

local data, or the expiry of a timer which is always set to no

more than L. On any such event pi calls the OnChange

method. When the event is a message X, |X| received from a

neighbor pj, pi would update Xi,j to X and |Xi,j | to |X | before it

calls OnChange.

144

The objective of the OnChange method is to make

certain that the conditions are maintained for the peer that runs

it. These conditions require Ki, Ai,j , and Wi,j to all be in RF

(Ki), which is not the tie region T . Of the three, Ki cannot be

manipulated by the peer. The peer thus manipulates both Ai,j , and Wi,j by sending a message to pj , and subsequently

updating Xi,j .However, this solution is one of the many

possible changes to Ai,j and Wi,j , and not necessarily the

optimal one. We leave the method of finding a value for the

next message Xi,j which should be sent by pi unspecified at

this stage, as it may depend on characteristics of the specific

RF. The other possible case is that RF (Ki) = T. Since T is

always the last region of RF, this means Ki is outside any other

region R ∈ RF. Since T is not necessarily convex, the only

option which will guarantee eventual correctness in this case

is if pi sends the entire withheld knowledge to every neighbor

it has. Lastly, we need to address the possibility that although

|Wi,j | = 0 we will have Ai,j which is different from Ki. This

can happen, e.g., when the withheld knowledge is sent in its

entirety and subsequently the local data changes.

Had we used sets of vectors, Wi,j would not have

been empty, and would fall into one of the two cases above.

As it stands, we interpret the case of non-empty Wi,j with zero

|Wi,j | as if Wi,j is in T .According to the above theorem, the

peer send withheld knowledge and local data changes. The

peer can rely on the correctness of the general results from the

previous section which assure that if F (Ki) is not the correct

answer then eventually one of its neighbors will send it new

data and change Ki. If, one the other hand, one of the

aforementioned cases do occur, then pi sends a message. This is performed by the SendMessage method. If Ki is in T then pi

simply sends all of the withheld data. Otherwise, a message is

computed which will assure Ai,j and Wi,j are in RF (Ki).

One last mechanism employed in the algorithm is a

“leaky bucket” mechanism. This mechanism makes certain no

two messages are sent in a period shorter than a constant L.

Leaky bucket is often used in asynchronous, event-based

systems to prevent event inflation. Every time a message

needs to be sent, the algorithm checks how long has it been

since the last one was sent. If that time is less than L, the

algorithm sets a timer for the reminder of the period and calls OnChange again when the timer expires. Note that this

mechanism does not enforce any kind of synchronization on

the system. It also does not affect correctness: at most it can

delay convergence because information would propagate more

slowly.

6. REACTIVE ALGORITHMS

The above section described an efficient generic local

algorithm, capable of computing any function even when the

data and system are constantly changing. Here, we leverage

this powerful tool to create a framework for producing and maintaining various data mining models. This framework is

simpler than the current methodology of inventing a specific

distributed algorithm for each problem and may be as efficient

as its counterparts. In this section we describe the mean

monitoring application. The basic idea of this framework is

used in order to monitor the quality of the model with the help

of algorithm. Here we discuss about the mean monitoring.

A. Mean Monitoring

The problem of monitoring the mean of the input

vectors has direct applications to many data analysis tasks.

The objective in this problem is to compute a vector μ which

is a good approximation for G. Formally, we require that

||G − μ ||≤ ∈ for a desired value of ∈. For any given estimate

μ, monitoring whether ||G – μ|| ≤ ∈ is possible via direct

application of the L2 thresholding algorithm .Every peer pi

subtracts μ from every input vector in Xi,i. Then, the peers

jointly execute L2 Norm Thresholding over the modified data.

If the resulting average is inside the ∈-circle then μ is a

sufficiently accurate approximation of G not otherwise.

The basic idea of the mean monitoring algorithm is to

employ a convergecast-broadcast process in which the

convergecast part computes the average of the input vectors

and the broadcast part delivers the new average to all the

peers. The trick is that, before a peer sends the data it collected

up the convergecast tree, it waits for an indication that the

current μ is not a good approximation of the current data.

Thus, when the current μ is a good approximation,

convergecast is slow and only progresses as a result of false alerts. During this time, the cost of the convergecast process is

negligible compared to that of the L2 thresholding algorithm.

When, on the other hand, the data does change, all peers alert

almost immediately. Thus, convergecast progresses very fast,

reaches the root, and initiates the broadcast phase. Hence, a

new μ is delivered to every peer, which is a more updated

estimate of G. The details of the mean monitoring algorithm

are given persist for a given period of time before the

convergecast advances. Experimental evidence suggests that

setting τ to even a fraction of the average edge delay greatly

reduces the number of convergecast without incurring a significant delay in the updating of μ. The second step is the

separation of the data used for alerting the input of the L2

thresholding algorithm from that which is used for computing

the new average. If the two are the same then the new average

may be biased. This is because an alert, and consequently

advancement in the convergecast, is bound to be more

frequent when the local data is extreme. Thus, the initial data,

and later every new data, is randomly associated with one of

two buffers: Ri, which is used by the L2 Thresholding

algorithm, and Ti, on whom the average is computed when

convergecast advances. In third step of implementation is the converge cast

process. First, every peer tracks changes in the knowledge of

the underlying L2 thresholding algorithm. When it moves

from inside the ∈-circle to outside the ∈-circle the peer takes

note of the time, and sets a timer to τ time units. When a timer

expires or when a data message is received from one of its

neighbors pi checks if currently there is an alert and if it was

145

recorded τ or more time units ago. If so, it counts the number

of its neighbors from whom it received a data message. If it

received data messages from all of its neighbors, the peer

moves to the broadcast phase, computes the average of its own

data and of the received data and sends it to itself. If it has

received data messages from all but one of the neighbors then this one neighbor becomes the peer’s parent in the

convergecast tree; the peer computes the average of its own

and its other neighbors’ data, and sends the average with its

cumulative weight to the parent. Then, it moves to the

broadcast phase. If two or more of its neighbors have not yet

sent a data messages pi keeps waiting.

Next deals with broadcast process. Every peer which

receives the new μ vector, updates its data by subtracting it

from every vector in Ri and transfers those vectors to the

underlying L2 thresholding algorithm. Then, it re-initializes

the buffers for the data messages and sends the new μ vector to its other neighbors and changes the status to convergecast.

There could be one situation in which a peer receives a new μ

vector even though it is already in the convergecast phase.

This happens when two neighbor peers concurrently become

roots of the convergecast tree (i.e., when each of them

concurrently sends the last convergecast message to the other).

To break the tie, a root peer pi which receives μ from a

neighbor pj while in the convergecast phase ignores the

message if i > j it ignores the message. Otherwise if i < j pi

treats the message just as it would in the broadcast phase.

7. SIMULATION RESULTS

The following are required simulations for analyzing

the performance of the algorithm. The dependency of cost and

quality of mean monitoring is based on many factors such as

alert mitigation period, alert threshold and length of epoch.

The below simulation gives the information about the

dependency of quality and cost on mean monitoring on the

alert mitigation period τ.

(a)Quality versus τ

(b)Quality versus τ

Some of the implementation results are given below.

(c) One Client side output for viewing events

(d) One server side output. This using incremental

algorithmic approach.

8. CONCLUSION

In this paper we present a generic algorithm which can

compute many function of the average data in large

distributed system. We present a number of interesting

applications for this generic algorithm. Besides direct

146

contributions to the calculation of L2 norm, the mean, and k-

means in peer-to-peer networks, we also suggest a new

reactive approach in which data mining models are computed

by an approximate or heuristic method and are then efficiently

judged by an efficient local algorithm. The future work is

based on the usage of local algorithm in network topology and the limitations of lcal computation.

REFERENCES

[1] R. Wolff and A. Schuster, “Association Rule Mining in Peer-to-Peer

Systems,” in Proceedings of ICDM’03, Melbourne, Florida, 2003, pp.

363–370.

[2]R. Wolff, K. Bhaduri, and H. Kargupta, “Local L2 Thresholding based

Data Mining in Peer-to-Peer Systems,” in Proceedings of SDM’06,

Bethesda, Maryland, 2006, pp. 428–439.

[3].D. Krivitski, A. Schuster, and R. Wolff, “A Local Facility Location

Algorithm for Sensor Networks,” in Proceedings of DCOSS’05, Marina

del Rey, California, 2005, pp. 368–375.

[4].Y. Birk, L. Liss, A. Schuster, and R. Wolff, “A Local Algorithm for Ad

Hoc Majority Voting Via Charge Fusion,” in Proceedings of DISC’04,

Amsterdam, Netherlands, 2004, pp. 275–289.

[5]. K. Bhaduri, “Efficient Local Algorithms for Distributed Data Mining in

Large Scale Peer to Peer Environments: A Deterministic Approach,”

Ph.D. dissertation, University of Maryland, Baltimore County, Baltimore,

Maryland, USA, May 2008.

[6].K. Das, K. Bhaduri, K. Liu, and H. Kargupta, “Distributed Identification

of Top-l Inner Product Elements and its Application in a Peer-to-Peer

Network,” IEEE Transactions on Knowledge and Data Engineering

(TKDE), vol. 20, no. 4, pp. 475–488, 2008.

[7]. S. Bandyopadhyay, C. Giannella, U. Maulik, H. Kargupta, K. Liu,

and S. Datta, “Clustering Distributed Data Streams in Peer-to-Peer

Environments,” Information Science, vol. 176, no. 14, pp. 1952–1985,

2006.

[8]. W. Kowalczyk, M. Jelasity, and A. E. Eiben, “Towards Data Mining

in Large and Fully Distributed Peer-to-Peer Overlay Networks,” in

Proceedings of BNAIC’03, Nijmegen, Netherlands, 2003, pp. 203–210.

[9].S. Datta, C. Giannella, and H. Kargupta, “K-Means Clustering over

Large, Dynamic Networks,” in Proceedings of SDM’06, Maryland, 2006,

pp. 153–164.

[10].M. Rabbat and R. Nowak, “Distributed Optimization in Sensor

Networks,”

in Proceedings of IPSN’04, California, 2004, pp. 20–27.

[11].N. Jain, D. Kit, P. Mahajan, P. Yalagandula, M. Dahlin, and Y. Zhang,

“STAR: Self-tuning aggregation for scalable monitoring,” in Proceedings

of VLDB’07, Sept. 2007, pp. 962–973.

[12]. R. van Renesse, K. P. Birman, and W. Vogels, “Astrolabe: A robust and

scalable technology for distributed system monitoring, management, and

data mining,” ACM Transactions on Computer Systems (TOCS), vol. 21,

no. 2, pp. 164–206, 2003.

[13].D. Kempe, A. Dobra, and J. Gehrke, “Computing Aggregate Information

using Gossip,” in Proceedings of FOCS’03, Cambridge, Massachusetts,

2003, pp. 482–491.

[14].Y. Afek, S. Kutten, and M. Yung, “Local Detection for Global Self

Stabilization,” Theoretical Computer Science, vol. 186, no. 1-2, pp. 199–

230, 1997.

[15].N. Linial, “Locality in Distributed Graph Algorithms,” SIAM Journal of

Computing, vol. 21, no. 1, pp. 193–2010, 1992.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

147

PROVINCE BASED WEB SEARCH

A.Suganthi1, M.Gowdhaman

2

1Lecturer,

2 Post Graduate Student

Department of Computer Science And Engineering,

PSG College of Technology,

Coimbatore.. [email protected] ,

[email protected]

Abstract

In this project, we search a given set of

keywords in categorized documents. Searching is

done after the categorization is completed and

categories of given documents are available.

Here we do two separate operations. First we

generate the categories and its related

categories. After that we give required web site

links to find categories of those links. Here each

website contents are parsed into keywords list

and using those keys the corresponding category

is determined. Now the documents and its

categories are computed to search using keys.

Second, we give keywords to search engine to

search the document and its corresponding

category. If keyword is composite of multiple

keywords then all keys are searched and its

corresponding document and its corresponding

category will be retrieved. The category contains

name, keys, and weights for corresponding keys.

Category is sorted using those weights and key

occurrences.

Keywords: Category, Search, Weights.

I. INTRODUCTION

IN the last 10 years, content-based document

management tasks have gained a prominent status in

the information system field, due to the increased

availability of documents in digital form and the ensuring need to access them in flexible ways.

Among such tasks, Text Categorization assigns

predefined categories to natural language text

according to its content. Text categorization has

attracted more and more attention from researchers

due to its wide applicability. Since this task can be

naturally modeled as a supervised learning problem.

Support Vector Machines (SVM) is the machine

learning algorithm introduced by Vapnik. SVM was

first applied for text categorization task by Jaochim. SVM is based on the Structural Risk Minimization

principle with the error-bound analysis. The method

is defined over a vector space where the problem is to

find a decision surface that “best” separates the data

points in two classes. While a wide range of classifiers have been used,

virtually all of them were based on the same text

representation, “bag of words,” where a document is represented as a set of words appearing in this

document. Values assigned to each word usually

express whether the word appears in a document or

how frequently this word appears. These values are

indeed useful for text categorization. However, are

these values enough?

Considering the following example, “Here you are”

and “You are here” are two sentences corresponding

to the same vector using the frequency-related values,

but their meanings are totally different. Although this is a somewhat extreme example, it clearly illustrates

that besides the appearance and the frequency of

appearances of a word, the distribution of a word is also

important.

Searching a given keyword set in a given document

set and categorizes the documents. If a keyword set

is given then it will determine the documents which are most relevant to that keyword set and also the

category which it belongs to that keyword set.

Most of the users are interested in the website

contents of their desired information. Also users want

the information location where that info is found. So

this project gives a solution for user that user can

148

search where a particular text paragraph is found in a

given set of websites and corresponding category.

With the enormous growth in information on the

Internet, there is a corresponding need for tools that

enable fast and efficient searching, browsing and

delivery of textual data. The concurrent execution

will greatly simplify the complexity of the search.

II. RELATED WORK

When the features for text categorization are

mentioned, the word “feature” usually have two

different but closely related meanings. One refers to

which unit is used to represent a document or to

index a document, while the other focuses on how to

assign an appropriate weight to a given feature.

Consider “bag of words” as an example. Using the former meaning, the feature is a single word, while

tf-idf weighting is the feature given the latter

meaning. This section will focus on previous

researches about the features used for text

categorization based on these two meanings. Other

topics about text categorization can be found in a

review .

Weights for keywords in different words are

calculated using modified tf.idf -method,

Weight(i,j)= tfi,j . idfi,j.

. where i is keyword, j is document and term

frequency tfi;j is ”normalized” by dividing it by the

total number of words in the corresponding

document. We take the logarithm of the document frequency to

nullify keywords that occur in all documents:

The contribution of this paper are the following:

.1.Distributional features for text categorization are

design.. Using these features can help improve the

performance, while requiring only a little additional

cost.

.2.How to use the distributional features is answered. Combine traditional term frequency with the

distributional features results in improved

performance.

.3.The factors affecting the performance of the

distributional features are discussed. The benefit of

the distributional features is closely related to the

length of documents in a corpus and the writing style

of documents.

Text categorization is a process of automatically

assigning text documents into some predefined

categories. Text categorization adopts classification

algorithms for building the models. For text domain,

features are a set of terms extracted from the document corpus. There are many previous works

which proposed and compared among different

algorithms for categorizing texts. Yang and Liu

(1999) presented a comparative study on many

different text categorization algorithms including

Support Vector Machines(SVM), k-Nearest Neighbor

(kNN), Neural Network(NNet), Linear Least-Squares

Fit (LLSF) and Naive Bayes(NB) . Based on the

thorough valuation, SVM yieldedthe best

performance, allowed by kNN, LLSF, NNet and NB,

respectively.

The main problem issue for text categorization is the

high dimensionality of feature space. The feature set

for text documents is a set of unique terms or words

which occur in all documents. In general, the number

of terms in a document corpus can be in the range of

tens or hundreds of thousands. To construct a model

for text categorization algorithms efficiently and effectively, feature selection technique is usually

applied. Yang and Pedersen (1997) presented a

comparative study on feature selection in text

categorization.

study on text categorization algorithms with various

feature Text categorization can be considered as a

classification approach in machine learning in which

features are words or terms extracted from a given

text corpus. The size of these features can be in the

range of tens or hundreds of thousands. To effectively and efficiently train the classification

model, we need to reduce the set of terms to a scale

of few thousands. Therefore, we applied feature

selection technique by comparing among several

well-known methods including document frequency

threshold (DF), information gain (IG) and χ2 (CHI).

Although, there were some previous works

performing comparative selection methods, most of

them focused on English text collection.

.we adopt three feature selection methods as follows.

• Document Frequency (DF) threshold: Document

frequency is the number of documents in which a

term occurs. DF value can be calculated for each

term from a training document corpus. The terms

whose DF values are less than some predefined

threshold are removed from the feature set.

149

• Information Gain (IG): IG is based on the concept

of information theory. It measures the amount of

information obtained for category prediction by

knowing the presence or absence of a term in a

document. We compute the IG value for each term

and move the term whose IG value is less than some predefined threshold.

• χ2 (CHI): CHI is based on the statistical theory. It

measures the lack of independence between the term

and the category. CHI is a normalized value and can

be compared across the terms in the same category.

III. EXISTING SYSTEM MODEL

In the existing system the search is based on the

category of the keyword which is given as input. The

categorization is done by some method without using

neural network. The main disadvantage is more

number of keywords are to be predefined and new

keywords are to be added frequently. So to avoid this problem proposed system uses neural network to

develop the procedure.

IV. PROPOSED SYSTEM MODEL

In our project, we search a given set

of keywords in categorized documents. Searching is done after the categorization is completed and

categories of given documents are available. Here we

do two separate operations. First we generate the

categories and its related categories. After that we

give required web site links to find categories of

those links. Here each website contents are parsed

into keywords list and using those keys the

corresponding category is determined. Now the

documents and its categories are computed to search

using keys. Second, we give keywords to search

engine to search the document and its corresponding

category. If keyword is composite of multiple keywords then all keys are searched and its

corresponding document and its corresponding

category will be retrieved. The category contains

name, keys, and weights for corresponding keys.

Category is sorted using those weights and key

occurrences.

V. PROPOSED SYSTEM DESIGN

The proposed system consist two kinds of

process:

1. Categorizer. 2. Searching.

Categorizer

Figure.1 Categorization Process.

First store all the keywords and its

corresponding weights in one file. We feed collection

of websites and keywords file to categorizer. It process the request and store the websites into

corresponding category.

Searching

The searching process accept the query and the

selected category from user, search the given query in

the selected category documents. Finally display the

URL of the documents that are mostly corresponding

to user needs.

Keywords Websites

Categorizer

Documents +

Categories

150

Figure.2 Searching Process

V. CONCLUSION Thus the analysis, design and

implementation of text categorization and searching

are done successfully. So that the user can able to do searching of a set of keywords in a list of websites

and the user can able to view the each keyword count

for a particular website. This searching is very useful

for crawl the websites with particular perspective

view of specific content. Also the search is running

concurrently, so we can get higher performance.

.

REFERENCES

1. T. Joachims. Text categorization with

support vector machines: Learning with

many relevant features. Proc. of the10th

European Conf. on Machine Learning,

pages 137–142,1998.

2. D.Lewis ,Y.Yang and F.Li.Revi: A new

benchmark collection for text categorization research.

J.of Machine Learning Research, 5:361-397,2004.

3. J.R.Quinlan. Induction of decision trees. Machine

Learning, 1(1):81-106, 1986.

4. T.Joachims, Text categorization with support

vector machines: Learning with many relevant

features,” proc. 10th European conf. Machine

Learning,1998.

5.D.Lewis,reuters-21578 Text Categorization test

collection, Dist. 1.0,1999.

6.M.Sauban and B.Pfahringer,”Text Categorization

using documents profiling,” Proc. Seventh European

conf. Priniciples and practice of knowledge discovery

in databases(PKDD’03),pp, 616-623,2003.

User query and

category

Search

Documents

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

151

Divergence Control of Data in Replicated

Database System P.K.Karunakaran, R.Sivaraman

Department of Computer Science and Engineering,

Anna University Tiruchirappalli, Tiruchirappalli,

Tamilnadu, India.

[email protected]

[email protected]

Abstract— To provide high availability for services such as mail

or bulletin boards, data must be replicated. One way to

guarantee consistency of replicated data is to force service

operations to occur in the same order at all sites, but this

approach is expanse. For some applications a weaker causal

operation order can preserve consistency while providing better

performance. This paper describes a distributed method to

control the view divergence of data freshness for clients in

replicated database systems whose facilitating or administrative

roles are equal. Our method provides data with statistically

defined freshness to clients when updates are initially accepted

by any of the replicas, and then, asynchronously propagated

among the replicas that are connected in a tree structure a new

way of implementing causal operations. Our technique also

supports two other kinds of operations: operations that are

totally ordered with respect to one another and operations that

are totally ordered with respect to all other operations. To

provide data with freshness specified by clients, our method

selects multiple replicas using a distributed algorithm so that

they statistically receive all updates issued up to a specified time

before the present time. We evaluated by simulation the

distributed algorithm to select replicas for the view divergence

control in terms of controlled data freshness, time, message, and

computation complexity. The method performs well in terms of

response time, operation-processing capacity, amount of stored

state, and number and size of messages: it does better than

replication methods based on reliable multicast techniques.

Keywords— Data replication, weak consistency, freshness, delay,

asynchronous update.

1. INTRODUCTION

Many computer based services must be highly available: they

should be accessible with high probability despite site crashes

and network failures. To achieve availability, the server’s state must be replicated. Consistency of replicated state can be

guaranteed by forcing service operations to occur in the same

order at all sites. However, some applications can preserve

consistency with a weaker, causal ordering, leading to better

performance. This paper describes a new technique that

supports causal order. An operation call is executed at just one

replica; updating of other replicas happens by lazy exchange

of ―gossip‖ messages—hence the name ―lazy replication‖.

The replicated service continues to provide service in spite of

node failures and network partitions. We have applied the

method to a number of applications, including distributed

garbage collection, deadlock detection [6], orphan detection [12], locating movable objects in a distributed system [9], and

deletion of unused versions in a hybrid concurrency control

scheme. Another system that can benefit from causally

ordered operations is the familiar electronic mail system.

Normally, the delivery order of mail messages sent by

different clients to different recipients, or even to the same

recipient, is unimportant, as is the delivery order of messages

sent by a single client to different recipients. However

suppose client c1 sends a message to client c2 and then a

later message to c3 that refers to information in the earlier

message.

If, as a result of reading class message, c3 sends an

inquiry message to c2, c3 would expect its message to be

delivered to c2 after c1’s message. Therefore, read and send

mail operations need to be causally ordered. Applications that use causally ordered operations may occasionally require a

stronger ordering. Our method allows this. Each operation has

an ordering type; in addition to the causal operations, there are

forced and immediate operations. Forced operations are

performed in the same order (relative to one another) at all

replicas. Their ordering relative to causal operations may

differ at different replicas but is consistent with the causal

order at all replicas. They would be useful in the mail system

to guarantee that if two clients are attempting to add the same

user-name on behalf of two different users simultaneously,

only one would succeed. Immediate operations are performed at all replicas in the same order relative to all other operations.

They have the effect of being performed immediately when

the operation returns, and are ordered consistently with

external events [7].

152

Fig.1. Need for data freshness in data warehousing.

They would be useful to remove an individual from a

classified mailing-list ―at once,‖ so that no messages

addressed to that list would be delivered to that user after the

remove operation returns. It is easy to construct and use

highly-available applications with our method. The user of a

replicated application just invokes operations and can ignore

replication and distribution, and also the ordering types of the

operations. The application programmer supplies a non

replicated implementation, ignoring complications due to

distribution and replication, and defines the category of each operation (e.g., for a mail service the designer would indicate

that send-mail and read-mail are causal, add–user and delete–

user are forced, and delete at once is immediate). To

determine the operation categories, the programmer can use

techniques developed for determining permissible

concurrency [8].

Our method does not delay update operations (such as

send-mail), and typically provides the response to a query

(such as read_ mail) in one message round trip. It will perform

well in terms of response time, amount of stored state, number

of messages, and availability in the presence of node and communication failures provided most update operations are

causal. Forced operations require more messages than causal

operations; immediate operations require even more messages

and can temporarily slow responses to queries. However,

these operations will have little impact on overall system

performance provided they are used infrequently. This is the

case in the mail system where sending and reading mail is

much more frequent than adding or removing a user. Our

system can be instantiated with only forced operations. In this

case it provides the same order for all updates at all replicas,

and will perform similarly to other replication techniques that

guarantee this property (e.g., voting or primary copy ). Our method generalizes these other techniques because it allows

queries to be performed on stale data while ensuring that the

information observed respects causality. The idea of allowing

an application to make use of operations with differing

ordering requirements appears in the work on ISIS [2] and

Psync , and in fact we support the same three orders as ISIS.

These systems provide a reliable multicast mechanism that

allows processes within a process group consisting of both

clients and servers to communicate; a possible application of

the mechanism is a replicated service. Our technique is a

replication method. It is less general than a reliable multicast mechanism, but is better suited to providing replicated

services that are distinct from clients and as a result provides

better performance: it requires many fewer messages than the

process-group approach; the messages are smaller, and it can

tolerate network partitions.

Our technique is based on the gossip approach first

introduced by Fischer and Michael , and later enhanced by

Wuu and Bernstein and our own earlier work [10]. We have

extended the earlier work in two important ways: by

supporting causal ordering for updates as well as queries and

by incorporating forced and immediate operations to make a

more generally applicable method. The implementation of

causal operations is novel and so is the method for combining

the three types of operations. The rest of the paper is

organized as follows. Section 2 describes the implementation

technique. The following sections discuss system

performance, scalability, and related work. We conclude with

a discussion of what we have accomplished.

2. MOTIVATION

Lazy-group replication is most suitable for large-scale data

processing systems, as described in Section 1. In lazy-group

replication, data freshness depends on replicas.The goal of an application will often affect the type of clustering algorithms

being used. For example, when trying to discover some good

locations for setting up the stores, a supermarket chain may

like to cluster their customers such that the sum of the

distance to the cluster center is minimized. For such

applications where the distance to the cluster centre is desired

to be short, partitioning algorithms like k-means and k-

mediods are often used. On the other hand, in applications

like finger print and facial recognition .In such case, clusters

discovered should have certain uniformity in density, colors,

etc. and can be of arbitrary shape and size. Since algorithms like k-means and k-mediods tend to discover clusters with

spherical shape and similar size, density-based algorithms will

be more suitable for these applications.

Fig. 2. System architecture used in divergence control for lazy-group replication.

3. CENTRALIZED VIEW DIVERGENCE CONTROL

3.1 System Architecture

The system architecture used in our method is shown in Fig.2.

This system processes two types of transactions: refer and

update. To process them, there are three types of nodes: client,

front-end, and replica nodes. Each replica node is composed of

more than one database server for fault tolerance, as

described in [8]. Replicas are connected by logical links.

As described in Section 1, our system is designed for

enterprise and cross-enterprise systems. A system architecture

for enterprise use must be simple for easy administration

and operation. For example, when update propagation

troubles occur in complicated systems, tracing where and how they occurred is difficult. Cross- enterprise systems in

particular require simple interfaces between enterprises for

troubleshooting. In addition, simple system architecture

leads to simple estimation of update delays for controlling

153

data freshness. From the viewpoint of fault tolerance, the

system architecture should have redundancy of processing

nodes and update propagation paths. To solve the

disadvantage of a tree-topology network, where there is only

one path between replicas, every replica is composed of

multiple database servers for node failures in our system.

An update to change data in replicated database systems

is propagated through the links of a tree as a refresh

transaction to which some updates are aggregated. How to

aggregate updates to one refresh transaction depends on

the application. For example, in mobile host tracking, old data are overwritten by the newest data. As a result, old

updates are omitted. For decision - support systems to

handle the sales of a product in an area, a refresh

transaction includes the update of totaled sales included

in all updates.

In our system, a replica may join and leave a tree-topology

network. When a replica joins a tree-topology network, the

replica’s administrator determines how it should be connected

based on the distances between replicas.

A client retrieves and changes the values of data objects

stored in replicated database systems through a front-end node. The functions of a client are to send refer and update

requests to a front-end node to retrieve and change data,

respectively, and to receive in reply processing results from

a front-end node. A refer request includes the degree of

freshness that is required by a client. The degree of

freshness, called read data freshness (RDF), is statistically

defined in this system. The RDF is formally described in

Section 3.3. When clients request the same RDF, our method

restricts the differences in RDF accordingly. We call this

difference in RDF among clients view divergence.

A front-end node is a proxy for clients to refer and update data in replicas. When a front-end node receives an

update from a client, it simply transmits the update to a

replica. In Fig. 2, when an update from client c1 is received

by front-end f1, it is forwarded to replica r3. When a

frontend node receives a refer request from a client, it

calculates data values satisfying the degree of freshness

required by the client after it retrieved data values and transaction processing logs of one or more replicas. We call

this set of replicas accessed by a front-end node a read

replica set. In Fig. 2, when front-end f3 receives a refer

request from client c3, it retrieves data values and

transaction processing logs from replicas r3; r9, and r12.

To process refer and update transactions, a replica has

five functions:

1. processing refer and update requests to its local

database;

2. processing asynchronously propagated updates in refresh transactions;

3. propagating update and refresh transactions to other

replicas;

4. calculating read replica sets for multiple degrees of

freshness by cooperating with other replicas; and

5. informing all front-end nodes about whether or not

it is a read replica for a degree of freshness.

3.2 Available Transactions

In lazy-group replication, any node can initially process

an update. Updates are then asynchronously propagated

among replicas. Because the orders of transactions processed

by replicas can vary, updates may conflict with each

other. Conflicts between updates cause inconsistent states

of data objects. To eliminate inconsistency, various ways are

used to process updates and record deletions in lazy-group

replication. For update processing, attributes associated

with update requests and data objects such as time stamps

and version numbers are used [4]. For eliminating inconsistent

states caused by the difference in the order of updates and record deletions among replicas, a method called

death certificate or tombstone is used.

Timestamped replace a value replaces a value with a newer

value. If the current value of the object already has a

timestamp greater than this update’s timestamp, the incoming

update is discarded.

3.3 Read Data Freshness As a measure of data freshness, there are two options:

the probabilistic degree and the worst boundary of data

freshness. When we use the probabilistic degree of freshness,

updates issued before a particular time are reflected in obtained data with some probability. On the other hand, data

reflect all updates issued before a particular time when data

freshness is defined by its worst boundary. There are a

number of applications whose users are tolerant to some

degree of errors inherent in the data. Therefore, we use the

probabilistic degree of data freshness as its measure, referring

to this measure as RDF. The degree of RDF represents a

client’s requirements in terms of the view divergence. The

formal definition of RDF, or Tp, is Tp= tc - te, where time

te is such that all updates issued before te in the network are

reflected in the data acquired by a client with probability p,

and tc is the present time. If the last update reflected in the

acquired data was at tl (<te), there is no update request

between tl and te with probability p.

3.4 Centralized Algorithm

3.4.1 Assumptions To determine a read replica set so that the client’s required

RDF is satisfied, we use the delay in update propagation

between replicas and assume three conditions.

1. The node clocks are synchronized.

2. The delay in update propagation can be statistically estimated.

3. The time it takes a client to obtain data, Tr, is less

than the degree of RDF required by the client, Tp.

Condition 1 means that the clocks on a client, front-end

node, and replica nodes are synchronized with each other.

Condition 2 means that we can estimate the upper confidence limit of the delay between replicas with some

probability by using samples of the measured delay time.

For obtaining data, we need the time for communication

between a client and a front-end node and for communication

between the front-end node and replica nodes. Therefore,

when a client requires data with a degree of RDF, Tp, it

can request only data that satisfy condition 3.

3.4.2 Terminology

154

We define four terms for explaining our algorithm: a range

originator, a read replica (set), a classified replica, and a mandatory replica. Table 1 describes their formal definitions.

A range originator for replica r is a replica from which a

refresh transaction can reach replica r within time Tp with

probability p, where Tp is the degree of RDF required by a

client.

TABLE 1

Formal Definition of Terminology

3.4.3 Calculating a Minimum Read Replica Set In our method, we have to choose a read replica set such that

For any replica r, at least one element of the set can receive

updates from replica r within Tp with probability p. This

means that the sum of the range originator sets of all read

replicas is the set of all replicas. The flow of the centralized

algorithm to calculate a minimum read replica set is shown in

Fig. 3a. The main part of the algorithm is composed of three

processes: classified-replica, mandatory-replica, and minimum-

subtree determination. The three processes are iterated until one

replica covers all replicas in a tree.

4 DISTRIBUTED VIEW DIVERGENCE CONTROL

4.1 Assumptions for Distributed View Divergence Control In addition to the assumptions for the centralized view

divergence control described in Section 3.4.1, we assume the

following conditions: 1. A replica can directly communicate with only its

adjacent replicas.

2. A replica knows the adjacent replica to which it

should send a message destined for a particular

replica. This can be achieved by various routing

protocols .

3. A front-end node knows the set of all replicas. This

can be accomplished using diffusing computation

and the termination detection for it .

4. Every replica knows the set of all front-end nodes.

On termination of the distributed algorithm, all replicas inform all front-end nodes about whether or not they are

read replicas. A front-end node can determine whether it

learns the set of all read replicas because it knows the set of

all replicas and every replica informs front-end nodes

whether or not it is a read replica.

Fig. 3. Flow of centralized and distributed algorithms to calculate a minimum read replica set.

4.2 Overview A distributed view divergence control method needs a

distributed algorithm to calculate a minimum read replica

set, which includes classified-replica, mandatory-replica, and

minimum-subtree determination. For efficiency, distributed view divergence control achieves low time and computation

complexity of the algorithm to calculate a minimum read

replica set. When the centralized algorithm to calculate a

minimum read replica set is executed in a distributed manner,

classified-replica, mandatory-replica, and minimum-subtree

determination are iterated a number of times. Therefore, each

replica needs to detect the termination of a current process to

determine when it should start the next process.

4.3 Classified-Replica Determination In the classified-replica determination of the centralized

algorithm, the range originator set of every replica is

compared with those of all the other replicas. If a replica

interacts with all the other replicas in distributed classified-

replica determination, it takes time D(bp + bc) + bp in

the worst case that a replica receives information from the

furthest replica, where D is the diameter of a tree for update

propagation. However, when new data are required, the range

originator set of a replica tends to include only replicas close

to it. Therefore, the time complexity of classified-replica

determination in the distributed algorithm can be improved

by eliminating comparisons between the range originator sets of replicas far from each other.

4.4 Mandatory-Replica Determination As is described in the centralized algorithm to calculate a

minimum read replica set, a classified replica is selected as

a mandatory replica when its range originator set has one or

more elements that are not included in the range originator

set of any other classified replica. In mandatory-replica

determination, we can also decrease its time and computation

complexity for the same reason as in classified-replica

155

determination because the range originator sets of replicas

far from each other tend to have no intersection.

4.5 Minimum-Subtree Determination Minimum - subtree determination in the distributed algorithm

consists of three functions: 1) calculating the minimum subtree for the next iteration as the centralized algorithm does;

2) removing replicas covered by mandatory replicas from the

range originator sets of replicas in the current subtree; and

3) detecting the termination of the distributed algorithm that

calculates a minimum read replica set. A replica in the

minimum subtree for the next iteration satisfies at least one of

the following two conditions: 1) it is not included in the range

originator set of any mandatory replica or 2) it is along the

path between replicas satisfying condition 1).

4.6 Time, Message, and Computation Complexity In each iteration, the time and message complexity of our

distributed algorithm to calculate a minimum read replica set

are the sums of those of the above processes. The computation

complexity of our distributed algorithm to calculate a

minimum read replica set is decreased from the centralized

algorithm by using the information of topology that connects

all replicas and the properties of probabilistic delay.

4.7 Dynamic Addition and Deletion of Replicas Dynamic addition and deletion of replicas cause the change

in paths along which refresh transactions are propagated

among replicas, which leads to the recalculation of range

originator sets in replicas. We divide replicas deleted from

replicated database systems into two types: leaf and nonleaf

replicas. When a leaf replica leaves replicated database

systems, our distributed algorithm works by deleting it from

the range originator sets of all replicas. When a nonleaf

replica leaves replicated database systems, a tree-topology

network for update propagation in our system is divided into two or more subtrees.

5 EVALUATION

5.1 Controlled Data Freshness Data freshness controlled by our algorithm depends on the

delay of update propagation and topology that connects

replicas. When our method is practically used, topology that

connects replicas should have a minimum diameter in order

to improve data freshness because the number of hop counts

for message propagation has the minimum value in a tree

with a minimum diameter.

5.2 Efficiency of Distributed Algorithm We evaluated the efficiency of our distributed algorithm in

Terms of time, message, and computation complexity. For

evaluation, we used 50 randomly generated tree-topology networks. They satisfied only the condition that the degrees

of nodes are in the range from 1 to 5. The numbers of

replicas in the evaluation were 100, 500, and 1,000.

6 RELATED WORK

Data freshness is one of the most important attributes of

data quality in a number of applications, To guarantee and

improve data freshness, various methods have been proposed.

When using lazy-group replication, a replica with the most

up-to-date data does not always exist as a source database.

Therefore, the aforementioned methods are not available in

lazy-group replication.

7 CONCLUSION

We have proposed a distributed method to control the view

divergence of data freshness for clients in replicated

database systems. We evaluated by simulation the

distributed algorithm to calculate a minimum read replica set

in terms of controlled data freshness time, message, and

computation complexity.

REFERENCES

[1] P.A. BERNSTEIN, V. HADZILACOS, AND N. GOODMAN, CONCURRENCY

Control and Recovery in Database Systems. Addison-Wesley, 1987.

[2] R . Ladin, B. Liskov, and S. Ghemawat, ―Providing High

Availability Using Lazy Replication, ‖ ACM Trans. Computer

Systems, vol. 10, no. 4, pp. 360-391, 1992.

[3] C. Pu and A. Leff, ―Replica Control in Distributed Systems: An

Asynchronous Approach,‖ Proc. ACM SIGMOD ’91, pp. 377-386,

May 1991.

[4] J. Gray, P. Helland, P. O’Neil, and D. Shasha, ―The Dangers of

Replication and a Solution,‖ Proc. ACM SIGMOD ’96, pp. 173-182,

June 1996.

[5] J.J. Fischer and A. Michael, ―Sacrificing Serializability to Attain

High Availability of Data in an Unreliable Network,‖ Proc. First

ACM Symp. Principles of Database Systems, pp. 70-75, May

1982.

[6] P. Cox and B.D. Noble, ―Fast Reconciliations in Fluid Replication,‖

Proc. Int’l Conf. Distributed Computing Systems, pp. 449-458, 2001

[7] M. Bouzeghoub and V. Peralta, ―A Framework for Analysis of

Data Freshness, ‖ Proc. Int’l Workshop Information Quality in

Information Systems, pp. 59-67, 2004.

[8] T. Yamashita and S. Ono, ―Controlling View Divergence of Data

Freshness in a Replicated Database System Using Statistical

Update Delay Estimation, ‖ IEICE Trans. Information and Systems,

vol. E88-D, no. 4, pp. 739-749, 2005.

[9] J. Han and M. Kamber, Data Mining, second ed. Morgan

Kaufmann, 2006.

[10] L.P. English, Improving Data Warehouse and Business Information

Quality. John Wiley & Sons, 1999.

[11] Distributed Systems , S. Mullender, ed. ACM Press, 1989.

[12] A. Demers, D. Greene, C. Hauser, W. Irish, J. Larson, S. Shenker,

H. Sturgis, D. Swinehart, and D. Terry, ―Epidemic Algorithm for

Replicated Database Maintenance, ‖ Proc. Sixth Ann. ACM Symp.

Principles of Distributed Computing, pp. 1-12, 1987.

[13] J. Gray and A. Reuter, Transaction Processing: Concepts and

Techniques. Morgan Kaufmann Publishers, 1993.

[14] E. Pacitti, E. Simon, and R. Melo, ―Improving Data Freshness in

Lazy Master Schemes, ‖ Proc. 18th IEEE Int’l Conf. Distributed

Computing Systems, pp. 164 - 171, May 1998.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

156

INTRODUCTION

Overview:

Similarity profiled temporal association

mining can reveal interesting relationships

of data items that co-occur with a particular

event over time. For example, the

fluctuation of consumer retail sales is

closely tied to changes in weather and

climate. While bottled water and generator

sales might not show any correlation on

normal days,a sales association between the

two items may develop with the increasing

strength of a hurricane in a particular region.

Retail decision makers may be interested in

such an association pattern to improve their

decisions regarding how changes in weather

affect consumer needs.Recent advances in

data collection and storage technology have

made it possible to collect vast amounts of

data every day in many areas of business

and science. Common examples are

recordings of sales of products, stock

exchanges, web logs, climate measures, and

Similarity-Profiled SpatioTemporal Association Mining

Esther Priscilla.E under the guidance of Ms.J.Preethi H.O.D(i/c) Dept of CSE

Anna University Coimbatore

Abstract—Given a time stamped transaction database and a user-defined reference sequence of interest over time,

similarity-profiled temporal association mining discovers all associated item sets whose prevalence variations over

time are similar to the reference sequence. The similar temporal association patterns can reveal interesting

relationships of data items which co-occur with a particular event over time.Most works in temporal association

mining have focused on capturing special temporal regulation patterns such as cyclic patternsand calendar scheme-

based patterns. The dissimilarity degree of the sequence of support values of an item set to the reference sequence is

used to capture how well its temporal prevalence variation matches the reference pattern. By exploiting interesting

properties such as an envelope of support time sequence and a lower bounding distance for early pruning candidate

item sets,an algorithm for effectively mining similarity-profiled temporal association patterns is developed..Mining

based on space is also included in our work referring spatio temporal mining

157

so on. One major area of data mining from

these data is association pattern analysis.

Association rules discover interrelationships

among various data items in transactional

data. However, most works in temporal

association mining have focused on special

temporal regulation patterns of associated

item sets such as cyclic patterns and

calendar-based patterns. For example, it may

be found that beer and chips are sold

together primarily in evening time on week

days. The temporal regulation patterns can

be explained with binary sequences where

the 1’s correspond to the time units in which

the pattern is prevalent (i.e., its support

value is greater than a minimum prevalence

threshold), and the 0’s correspond to the

time units in which the pattern is not

prevalent. For instance, when the unit of

time is day, a repetitive temporal pattern

on Monday is defined as 10000001000000. .

Fig. 1a illustrates an example of temporal

regulation patterns. It shows the prevalence

values of three item sets I1, I2, and I3 over

time, and the binary temporal sequences of

I1 and I2 under a fixed prevalence threshold

(e.g., support threshold ). It can be noticed

that item sets I1 and I2 show the same .Fig.

1b illustrates an example of temporal

similarity patterns. The prevalence values of

item sets I2 and I3 show a very similar

variation with a user guided reference

sequence R. The reference sequence can

represent the change of prevalence of an

item of interest (e.g., a product sale in

market basket data, a stock exchange in the

stock market, a climate measure such as

temperature or precipitation, and a scientific

phenomenon), or a user guided prevalence

sequence pattern showing special shapes

(e.g., a seasonal, emerging, or diminishing

pattern over time). Current methods for

temporal regulation pattern mining cannot

reveal these types of temporal patterns that

are based on actual prevalence similarity.

Application examples. similarity-profiled

temporal association patterns can give

158

important insight into many application

domains such as business, agriculture, earth

science, ecology, and biology. They can also

be used as filtered information for further

analysis of market trends, prediction, and

factors showing strong connections with a

certain scientific event. For example, the

weather-to-sales relationship is a very

interesting problem in retail analysis.

Managers in the merchandise and food sales

division of Walt Disney World are hoping to

find correlations between daily temperatures

and sales. Wal-Mart discovered a surprising

customer buying pattern during hurricane

season in one of its sales regions. Not only

did survival kits (e.g., flashlights,

generators, tarps) show similar selling

patterns with bottled water, but also did the

sales of Strawberry Pop-Tarts (a snack item)

Our method can help in finding such item

sets whose sales similarly change to that of a

specific item (event) for a period of time.

The mining results can improve supply

chain planning and retail decision making

for maximizing the visibility of items most

likely to be in high demand during special

time periods. Another example in the

business domain from the online website

Weather.com, which offers weather-related

lifestyle information including travel,

driving, home and garden, and sporting

events, as well as weather information. The

website reports that almost 40 percent of

weather.com visitors shop home

improvement products when temperatures

rise . The website may attract more

advertisers if it can analyze the relationships

of visited websites through weather.com

with changes of weather. Our temporal

patterns may be used for finding such

weather-related sponsor sites.

OBJECTIVE:

In this project, similarity-profiled temporal

association pattern is formalized. The

problem of mining the temporal pattern is

formulated with a similarity-profiled subset

specification, which consists of a reference

time sequence, a similarity function, and a

dissimilarity threshold. The subset

specification is used to define a user interest

temporal pattern and guide the degree of

approximate matching of prevalence values

of associated item sets for it. Similarity-

profiled temporal association mining

presents challenges during computation. The

straight-forward approach is to divide the

mining process into two separate phrases.

The first phrase computes the support values

of all possible item sets at each time point

and generates their support sequences. The

159

second phrase compares the generated

support time sequences with a given

reference sequence and finds similar item

sets. In this step, a multidimensional access

method such as an R-tree family can be used

for a fast sequence search. However, the

computational costs of first generating the

support time sequences of all combinatorial

candidate item sets and then doing the

similarity search become prohibitively

expensive with increase of items. Thus, it is

crucial to devise schemes to reduce the item

set search space effectively for efficient

computation. It is explored that interesting

properties for similarity-profiled association

mining. First, to estimate support sequences

without examining an input data set, we

define tight upper and lower bounds of true

support sequences. For early pruning of

candidate item sets, the concept of a lower

bounding distance, which is often used for

indexing schemes in the time series

literature and define the lower bounding

distance with the upper and lower bounds of

support sequences is utilized. upper lower

bounding distance is especially noteworthy

because it is monotonically nondecreasing

with the size of the item set. This property is

used for further reducing the item set search

space. A Similarity-Profiled temporal

Association Mining method (SPAMINE)

algorithm is developed on our algorithmic

design concept. For comparison, an

alternative algorithm, a sequential method

using a traditional support-pruning scheme,

is also presented. It is analytically shown

that the SPAMINE algorithm is complete

and correct, i.e., there is no false dropping or

false admission for similarity-profiled

associated item sets.

PROBLEM DEFINITION:

Problem Statement Given:

1. a finite set of items I;

2. an interest time period T t1 tn, where

ti is a time slot by a time granularity, ti tj

,i j

3. a time-stamped transaction database

DD1 Dn, Di Dj ,i j

, wherein each transaction d D is a tuple

<timestamp; itemset>, where timestamp is a

time T that the transaction is executed,

and itemset I. Di is a set of transaction

included in time slot ti; and

4. a subset specification:

4a. a reference sequence R <r1; . . . rn>

over time slots t1; . . . ; tn,

4b. a similarity function

similarityf ( ,x y

) I RN where x

and y

are

numeric sequences, and

160

4c. a dissimilarity threshold

Find. A set of item sets I I that satisfy the

given subset specification, i.e.,

similarityf (

,IS

R

)≤ where

IS

= <s1; . . . ; sn> is the sequence of

support values of an item set I over time

slots t1. . . tn.

Objective. Find the complete and correct

result set while reducing the computational

cost.

SYSTEM ANALYSIS

EXISTING SYSTEM:

The system used for mining just takes the

transaction no time.

Comparison of two standard can only be

done individually

Either manual or system input is been given.

PROPOSED SYSTEM:

For the first time both transaction and time

are taken under consideration

More than two standard can be compared at

the same transaction

Both manual and system input is been given

Systemflow diagram

yes

yes

Username

&

Password

Authenticatio

n

Load the

data

Data Input

path

Threshold

value

Apply

association

Final

Result

Exit

no

161

CONCLUSION AND FUTURE WORK

It is formulated that the problem of mining

similarity-profiled temporal association

patterns and proposed a novel algorithm to

discover them. The proposed SPAMINE

algorithm substantially reduced the search

space by pruning candidate item sets using

the lower bounding distance of the bounds

of support sequences, and the monotonicity

property of the upper lower bounding

dist

anc

e

wit

hout compromising the

correctness and completeness of the mining

results. Experimental results on synthetic

and real data sets showed that the

SPAMINE algorithm is computationally

An example of similarity-profiled temporal association mining: (a) input data,

(b) generated prevalence (support) time sequences and sequence search, and (c)

output item sets.

162

efficient and can produce meaningful results

from real data. However, our pruning

scheme effect depends on data distribution,

dissimilarity threshold, and type of reference

sequence. In the future, It is explored that

different similarity models for our temporal

patterns. The current similarity model using

a Lp norm-based similarity function is a

little rigid in finding similar temporal

patterns. It may be interesting to consider

not only a relaxed similarity model to catch

temporal patterns that show similar trends

but also phase shifts in time. For example,

the sale of items for cleanup such as chain

saws and mops would increase after a storm

rather than during the storm

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

Effective Integration any Inter Attribute Dependency Graphs of Schema

Matching M. Ramkumar, M.E (Computer science and Engineering)

Email:[email protected] Mrs. V.S. Akshaya, Senior Lecturer

Department of Computer Science and Engineering Kumaraguru College of Technology, Coimbatore

163

Abstract—Schema matching is one of the key challenges in information integration. It is a

labor-intensive and time-consuming process. To

alleviate the problem, many automated solutions

have been proposed. Most of the existing solutions mainly rely upon textual similarity of

the data to be matched. However, there exist

instances of the schema-matching problem for which they do not even apply. Such problem

instances typically arise when the column names

in the schemas and the data in the columns are opaque or very difficult to interpret. In our

previous work, we proposed a two-step

technique to address this problem. In the first

step, we measure the dependencies between attributes within tables using an information-

theoretic measure and construct a dependency

graph for each table capturing the dependencies among attributes. In the second step, we find

matching node pairs across the dependency

graphs by running a graph-matching algorithm.

Index Terms—Schema matching, attribute

dependency, graph matching.

1 INTRODUCTION

The schema-matching problem at the most basic

level refers to the problem of mapping schema elements (for example, columns in a relational

database schema) in one information repository

to corresponding elements in a second

repository. While schema matching has always been a problematic and interesting aspect of

information integration, the problem is

exacerbated as the number of information sources to be integrated, and hence, the number

of integration problems that must be solved,

grows. Such schema-matching problems arise both in “classical” scenarios such as company

mergers and in “new” scenarios such as the

integration of diverse sets of queriable

information sources over the web.

To clarify our aims and provide some context, consider a classical schema mapping

problem where two employee tables are

integrated. How should we determine which

attributes in one table should be mapped to which attributes in the other table? First, one

logical approach is to compare attribute names

across the tables. Some of the attribute names will be clear candidates for matching, due to

common names or common parts of names.

Using the classification given in such an approach is an example of schema-based

matching. However, for many columns, schema-

based matching will not be effective because

different institutions may use different names for semantically identical attributes, or use similar

names for unrelated attributes.

To gain insight into our approach,

consider the example tables in Table 1. Suppose

these tables are from two automobile plants in

different companies. Imagine that the column

names of the second table and data instances in

columns B and C are some incomprehensible

values to the schema-matching tools.

Conventional instance-based matchers may find

correspondence between the columns Model and

A due to their syntactic similarity. However, no

further matches are likely to be found because

the two columns B and C cannot be interpreted,

and they share exactly same statistical

characteristics; that is, they have the same

number of unique values, similar distributions,

and so forth

When schema-based matching fails, the

next logical approach is to look at the data

values stored in the schemas. Again referring to

the classification from this approach is called

instance-based matching [18], [32], [39].

[Type text]

164

Instancebased matching also will work in many

cases. For example, if we are deciding whether

to match Dept in one schema to either

DeptName or DeptID in the other, by looking at

the column instances, one may easily find the

mapping because DeptName and DeptID are

likely to be drawn from different domains.

Unfortunately, however, instance-based

matching is also not always successful.

When instance-based mapping fails, it is

often because of its inability to distinguish

different columns over the same data domain and, similarly, its inability to find matching

columns using values drawn from different

domains. For example, EmployeeID and CustomerID columns in a tableare unlikely to be

distinguished if both the columns use similar

IDs. By the same token, if the two tables use different types of IDs for the same column, the

traditional

TABLE 1 Two Tables from Car Part Databases

To make progress in such a difficult

situation, our technique exploits dependency

relationships between the attributes in each table. For instance, in the first table in Table 1,

there will exist some degree of dependency

between Model and Tire if model partially determines the kinds of tires a car can use. On

the other hand, perhaps Model and Color are

likely to have very little interdependency. If we can measure the dependency between columns

A and B and columns A and C, and compare

them with the dependency measured from the

first table, it may be possible to find the remaining correspondences.

2 UNINTERPRETED MATCHING

In this section, we describe in detail our uninterpreted structure-matching technique. The

algorithm takes two table instances as input and

produces a set of matching node pairs. Our

approach works in two main steps as shown below:

1. G1 <- Table2DepGraph(S1);

G2 <- Table2DepGraph(S2) and

2. ( G1(a) , G2(b ) ) GraphMatch(G1,G2),

where Si = input table, Gi = dependency graph, ( G1(a), G2(b) ) = matching node pair.

The function Table2DepGraph( ) in the first step

transforms an input table like the one shown in Fig. 1a into a dependency graph shown in Fig.

1c. The function GraphMatch( ) in the second

step takes as input the two dependency graphs

generated in the first step and produces a mapping between corresponding nodes in the

two graphs.

The two steps are described in detail later in this section

Fig 1 Two input table examples and their dependency graphs. A weight on an edge represents mutual information between the two adjacent attributes and a weight on a node

represents entropy of the attribute (or equivalently, self-information MI(A;A)).

3 WEIGHTEDGRAPH-MATCHING

ALGORITHMS FOR SCHEMA

MATCHING

In this section, we investigate efficient

approximation algorithms for the graph-matching problem in the second step of our

approach. We focus our discussion particularly

on the bijective mapping problem for two

reasons. 1) The solution to this problem can be used as an integral part of the general solutions

for the other two problems because the other

[Type text]

165

problems can be formulated with multiple

bijective mappings. For example, an injective mapping between graphs S (m nodes) and T (n

nodes, where m > n) can be solved by finding an

n node subgraph of S that minimizes the

bijective mapping distance to T. The partial mapping problems can be formulated similarly.

Of course, this may not be an ideal solution for

them. Evaluating this approach versus approaches specifically tailored to the injective

and partial mapping problems is an interesting

area for future work. 2) The problem can be formulated in a clean mathematical optimization

framework and because of that a large number

of approximation algorithms have been

developed. We will investigate a spectrum of the solutions covering a wide range of optimization

techniques that can work for our problem

context.

The bijective mapping problem between

two dependency graphs S and T is essentially the problem of finding a permutation matrix P

that minimizes the euclidean distance between

the two dependency graphs’ adjacency matrices

AS and AT , respectively, and can be formulated as

MinPЄ || AS - PTATP||

2 F

where minx f(x) is a function that finds

an x that minimizes the f(x),|| . ||2 F is a square of

a euclidean norm (or Frobenius norm, ||A||2 F

= a2

i j

and is the set of all permutation matrices.

The above problem is known as a

weighted graphmatching problem (WGMP). WGMP is a purely combinatorial problem, and

it is generally very difficult to find an exact 0-1

integral solution. The complexity of the problem is largely dependent on the choice of metric

being optimized. However, we can show that, at

the least, the general WGMP includes the graph

isomorphism problem, for which no polynomial time algorithm has yet been found . This leads

us to focus on finding an efficient approximate

algorithm that can produce a “nearly optimum” solution, rather than finding an exact search

algorithm.

In what follows, we introduce five weighted graphmatching algorithms:

1. Umeyama’s eigen-decomposition (ED)

approach, 2. linear programming (LP),

3. convex quadratic programming (QP),

4. hill climbing (HC), and, finally, 5. branch and bound.

4 Hill-Climbing Approach So far, we have considered three deterministic

approximation algorithms for WGMP. All of

them are based on the relaxation of the original

problem to an algebraic optimization framework. We now introduce a simple

nondeterministic, iterative improvement

algorithm. The HC algorithm is simply a loop that moves, in each state transition, to a state

where the most improvement can be achieved. A

state represents a permutation that corresponds to a mapping between the two graphs. We limit

the set of all states reachable from one state in a

state transition, to a set of all permutations

obtained by one swapping of any two nodes in the permutation corresponding to the current

state. The algorithm stops when there is no next

state available that is better than the current state. As we can see, it is nondeterministic;

depending on where it starts, even for the same

problem, the final states may differ. To avoid

being stuck in a local minimum after an unfortunate run, the usual practice is to perform

some number of random restarts.

Hill Climbing Algorithm

currentNode = startNode;

loop do L = NEIGHBORS(currentNode);

nextEval = -INF;

nextNode = NULL;

for all x in L if (EVAL(x) > nextEval)

nextNode = x;

nextEval = EVAL(x); if nextEval <= EVAL(currentNode)

//Return current node since no better

neighbors exist return currentNode;

currentNode = nextNode;

[Type text]

166

Modeling new Dependency Relation The dependency graph function produces such

dependency graphs by calculating the pair wise

mutual information over all pairs of attributes in

a table and structuring them in an undirected labeled graph. For instance, each edge in the

dependency graph G1 has a label indicating

mutual information between the two adjacent nodes; for example, the mutual information

between nodes A and B is 1.5, and so on. The

label on a node represents the entropy of the attribute, which is equivalent to its mutual

information with itself or self-information.

Hence, we can model our dependency graph in a

simple symmetric square matrix of mutual information, which is defined as follows:

Let S be a schema instance with n attributes and

(1 )ia i n be its ith attribute. We define

dependency graph of schema S using square

matrix M by

( ),ijM m where

( ; ),1 ,ij i jm MI a a i j n

The intuition behind using mutual information as a dependency measure is twofold: It is value

independent; hence, it can be used in

uninterpreted matching. It captures complex correlations between two probability

distributions in single number, which simplifies

the matching task in the second stage of our

algorithm.

Enhanced Weighted Graph-Matching

Algorithms For Schema Matching. Hill climbing is a simple nondeterministic,

iterative improvement algorithm. The HC

algorithm is simply a loop that moves, in each state transition, to a state where the most

improvement can be achieved. A state represents

a permutation that corresponds to a mapping

between the two graphs. We limit the set of all states reachable from one state in a state

transition, to a set of all permutations obtained

by one swapping of any two nodes in the permutation corresponding to the current state.

The algorithm stops when there is no next state

available that is better than the current state. As we can see, it is nondeterministic; depending on

where it starts, even for the same problem, the

final states may differ. To avoid being stuck in a

local minimum after an unfortunate run, the usual practice is to perform some number of

random restarts.

proposed new Graph-Matching Algorithms

For Schema Matching

The given database of known objects and a query, the task is to retrieve one or several objects from the database that are similar to query. If graphs are use for object representation this problem turns into determining the similarity of graphs, which is generally referred to as graph matching. Standard concepts in graph matching include graph isomorphism, subgraph isomorphism, and maximum common subgraph. The graph matching algorithms

I) Direct Tree Search

Searches all subgraph isomorphs between

two graph GA and GB

1. pA and qA are the number of nodes and

edges of GA

and pB and qB are likewise for GB

2. mij=1 if the degree of GB ’s j th node is

higher or equal to degree of GA ’s i th node, 0 otherwise

3. M=[mij] 4. M’= pA x pB whose elements are either

0 or 1

5. C=[cij]=M’(M’B)T

6. If (i j) (aij=1) (cij=1) 1 i pA

1 j pA

then matrix M’ results subgraph

isomorph between GA and GB

[Type text]

167

Fig 2. Entrophy each item.

Fig 3. Inter attribute mutual information.

Fig 4. Conditional entrophy.

In this project, we have proposed a two-step schema-matching technique that works even in the presence of opaque column names and data values. In the first step, we measure the pairwise attribute correlations in the tables to be matched and construct a dependency graph using mutual information as a measure of the dependency between attributes. In the second stage, we find matching node pairs across the dependency graphs by running a graph-matching algorithm. Our work is the first to introduce an uninterpreted matching technique utilizing interattribute dependency relations. We have shown that while a single column uninterpreted matching such as entropyonly matching can be somewhat effective alone, further improvement was possible by exploiting interattribute correlations.

5 CONCLUSION

We have proposed a two-step schema-matching techniquethat works even in the presence of

opaque column namesand data values. In the

first step, we measure the pairwise

[Type text]

168

attribute correlations in the tables to be matched

and construct a dependency graph using mutual information as a measure of the dependency

between attributes. In the second stage, we find

matching node pairs across the dependency

graphs by running a graph-matching algorithm. To our knowledge, our work is the first to

introduce an uninterpreted matching technique

utilizing interattribute dependency relations. We have shown that while a single column

uninterpreted matching such as entropyonly

matching can be somewhat effective alone, further improvement was possible by exploiting

interattribute correlations. In this work, we also

investigated approximation algorithms for the

matching problem and showed that an efficient implementation can be possible for our

approach. Among the algorithms we evaluated,

the HC approach showed the most promising results. It found close to optimal solutions very

quickly, suggesting that the graphmatching

problems arising in our schema-matching domain are amenable to HC. A good deal of

room for future work exists. In our work, we

have only tested two simple distance metrics—

euclidean and normal. It is possible that more sophisticated distance metrics could produce

better results. It would also be interesting to

evaluate other dependency models using different uninterpreted methods.

6 REFERENCES

1. E.D. Andersen and K.D. Andersen, “The

Mosek Interior Point Optimizer for Linear Programming: An Implementation of the

Homogeneous Algorithm,” Proc. High

Performance Optimization Techniques (HPOPT), 1997.

2. K.M. Anstreicher and N.W. Brixius, “A

New Bound for the Quadratic Assignment

Problem Based on Convex Quadratic Programming,” Math. Programming, vol.

89, pp. 341-357, 2001.

3. P. Atzeni, G. Ausiello, C. Batini, and M. Moscarini, “Inclusion and Equivalence

between Relational Database Schemata,”

Theoretical Computer Science, vol. 19, pp. 267-285, 1982.

4. P. Atzeni, P. Cappellari, and P.A. Bernstein,

“Modelgen: Model Independent Schema

Translation,” Proc. 21st Int’l Conf. Data

Eng. (ICDE ’05), pp. 1111-1112, 2005. 5. C. Beeri, A.O. Mendelzon, Y. Sagiv, and

J.D. Ullman, “Equivalence of Relational

Database Schemes,” SIAM J. Computing,

vol. 10, no. 2, pp. 352-370, 1981. 6. J. Albert, Y.E. Ioannidis, and R.

Ramakrishnan, “Conjunctive Query

Equivalence of Keyed Relational Schemas,” Proc. 16th ACM SIGACT-SIGMOD-

SIGART Symp. Principles of Database

Systems (PODS ’97), pp. 44-50, 1997 7. H.A. Almohamad and S.O. Duffuaa, “A

Linear Programming Approach for the

Weighted Graph Matching Problem,” IEEE

Trans. Pattern Analysis and Machine Intelligence, vol. 15, no. 5, pp. 522-525,

May 1993.

M. Ramkumar received B.E degrees in Computer Science

and Engineering from the Anna

University in 2008 and pursuing ME degree in Computer Science

and Engineering from the Anna

University Coimbatore in 2010. His research

interests include Data mining.

Designed a focused crawler using multi-agents system with fault tolerance involving learning component

M.Balaji Prasath, Lecturer/IT

[email protected]

Sona College of Technology Department of Information

Technology, Salem.

ArulPrakash,Student/IT

[email protected]

Sona College of Technology Department of Information

Technology, Salem

DeenaDayalan,Student/IT [email protected]

Sona College of Technology Department of Information

Technology, Salem

ABSTRACT:-

Internet has become an important source for collecting useful information. There are various search engines that indexes only the static collections of web and lack in filtering the duplicate web results and also, web search engines typically provide search results without considering user interests or context. This scenario leads to a development of more sophisticated search tool with built-in intelligence using user personalized interest. The proposed system describes the design of such an intelligent web crawling system with multiple agents mining the web online at query time. The intelligence is incorporated into agents using neural networks with reinforcement learning. To provide search results correspond to user relevance, user interest are mapped with ODP taxonomy.

Keyword: Focused Crawler, Personalized Crawler, Multi –agent Systems

Proceedings of the Third National Conference on RTICT 2010 Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

169

Introduction:

The web creates new challenges for information retrieval. The amount of information on the web is growing rapidly, as well as the number of new users inexperienced in the art of web research. People are likely to surf the web using its link graph, often starting with high quality human maintained indices such as Yahoo!

Human maintained lists cover popular topics effectively but are subjective, expensive to build and maintain, slow to improve, and can’t cover all esoteric topics. Automated search engines that rely on keyword matching usually return too many low quality matches .To make matters worse; some advertisers attempt to gain people’s attention by taking measures meant to mislead automated search engines.

The term “search engines” is often used generically to describe both crawler-based search engines and human-powered directories. These two types of search engines gather their listings in radically different ways.

Search engines use indexing algorithms to index web pages using a program called web crawler or robot that visit web pages regularly and index them. They can be distinguished on two aspects: crawling and ranking strategy. Crawling strategy decides which page to download next and the ranking strategy contributes to the retrieval of search results based on rank. The above two strategies decides the parameters like recall (fraction of relevant pages that are retrieved), recency (fraction of retrieved pages that are current) of a search engine and precision (fraction of retrieved pages that are relevant) of a search engine.

Information Retrieval algorithms support thecomputerized search of large document collections (millions of documents) to retrieval small subsets of documents relevant to a user's information need. Such algorithms are the basis for Internet search engines and digital library catalogues. This paper attempts to find a solution for scalability problem by incorporating

agent based systems for mining the hyperlinks on a web page to find more quality web pages. The solution is designed in such a way that it narrows down the gap between the current web content and the one found at the previous crawls. A significant characteristic of a web page is “quality”, which is the relevance of web pagesthat is to be retrieved as per the user’s request. In all other agent based information retrieval systems, the agent is made to crawl according to specified predefined information. The prototype system called LACrawler mines the hyperlink information and learns itself from the previous crawls and enhances the quality of web pagesduring retrieval.

This paper presents an architectural framework for locating relevant Web documents for a specific user or group of users. The personalization is incorporated in the input and crawling modules. The input module consists of a topic suggestion component that extracts search query terms from different sources. The crawle module is realized with two types of agents: retrieval agents and coordinator agent. The coordinator agent is built with multi-level frontier queue to achieve tunneling and the URLs stored there are disseminated to retrieval agents. The retrieval agents download and classify Web pages as relevant or irrelevant using personalized relevance score.

2. Research Issues

Every search engine has its own crawling and ranking algorithms that influence the parameterslike recall, recency, and precision. It is to be noted that the success of Google is attributed to its ranking algorithms.

The following are the issues which influence thedesign of intelligent agent based crawlers:

I. Incorporating a technique to calculate the relevance of a page.

II. Automatic extraction of information from unfamiliar web sites.

III. Incorporating learning ability.

170

IV. To make agents share their experience.V. To make agents fault tolerant.

When employing multiple agents, care should betaken to make the overall system manageable,flexible and robust. The main question is how toincorporate scalability and in what terms. In thisarticle, scalability is defined as the number ofrelevant pages downloaded per second. Anotherdesign issue is employing the learning capability in the system. It is known that feed-forward neural network is best suited for pattern identification problem.

As per the definition, the autonomous agents act continuously until it is eaten or otherwise dies automatically.To incorporate graceful degradation, the workload of the dead agent should be transferred to newly spanned agent. To identify the agent that is destroyed due to system crash or network interruptions, either timeout or message passing scheme or a combination of both may be employed.

3. Architecture:

The inputs to the system are: seed URLs, query terms or keyword list, and maximum number of pages to crawl. The seed URLs serve as the starting point of search. They can be obtained either from a domain expert or from a trusted search engine like Google. As the complete

information regarding relevant pages is not known apriori, getting input from domain expert is a tedious task, if not impossible. Due to this difficulty,the input is obtained from the Google. The starting documents are pre-fetched and each agent is positioned at one of these documents and given an initial amount of energy to crawl.

initial query term

topic taxonomy

User profile

Suggested list Of query terms

The main functionality of this module is to drive the crawler with appropriate query terms and an example set of URLs. This feature makes the crawler to decide whether a Web page is relevant or not apart from exploiting information available in link structures. This step is considered crucial since the user may not exactly know the keyword describing the topic of search. If these terms are extracted from other sources and suggested, the user may very well be able to guide the crawler to the correct direction. The user has to specify the initial search terms which are stored in the user profile, to query term evaluator. This component integrates topic taxonomy like Yahoo! and user profile, to extract keywords from example set of documents available in the taxonomy. The query term evaluator suggests a list of keywords and URLs of example Web documents to the user. He/she has to select one or more of them which is given as input to the crawler module. While picking the search terms, the user has to provide a weight measure which is the indicator of how important is the presence of that term in the Web document.

4. Implementation details

The web crawler design exploits computational parallelism to achieve efficiency. Hence,

Query term evaluator

User interface

171

multithreading is expected to provide betterutilization of resources. The distributed agent isimplemented using java as it has built-in support for threads. To keep the interface as simple as possible, the users are kept transparent about the parameters used. The query terms and maximum number of pages to be crawled are given as input. Once the search is started, the system obtains the seed URLs from a trusted search engine like Google and initiates crawlers. A crawled page is displayed to the user if and only if its probability of estimated relevance is above a given threshold.

LACrawler Algorithm:

User_input(query_terms, MAX); Seed_URLs = search-engine(query_terms);Prefetch_document(seed_URLsInitialize_dagent ( query_terms, MAX); for each dagent position (dagent, document); dagent_energy=initial value; for each dagent while visited < MAX get_the_hyperlink( );for each hyperlink dagent_doc := fetch_new_document( );estimate_link_relevance(query_terms, dagent_doc);estimate_prob(link_relevance); learning (dagent, link_relevance);if estimate_prob>thresholddisplay();store();dagent_energy+= link_relevance – latency( ); reproduce( );elseSend (“killed”);Kill;

5 Conclusions:

Several issues of designing a personalized crawler are discussed in this paper. A new

architecture is proposed to locate relevant information with respect to particular user. A novel strategy is proposed to compute personalized relevance score. The input module integrated in the system improves the effectiveness of the retrieval task. Experimental results suggest that the system proposedperforms well in terms of harvest rate, and coverage. This work can be extended to apply dynamic data fusion algorithm in the coordinator module allowing the retrieval agents to download duplicate pages in order to reduce the time taken on coordinating them by a centralagent.

References

[1] Aktas,M. S., Nacar,M. A., andMenzcer, F. Using hyperlinks features to personalizeweb search. Advances in Web Mining and Web Usage Analysis, LNCS, pages 104–115, 2006.[2] Altingövde, I. S. and Özgür Ulusoy. Exploiting inter-class rules for focussed crawling. IEEE Intelligent Systems, 19:66–73, 2004.[3] Bar-Yossef, Z. and Gurevich, M. Random sampling from a search engine’s index. In Proceedings of 15th International conference on WWW, pages 367–376, 2006.[4] Barfouroushi, A., Anderson, M., Nezhad, H. M., and Perlis, D. Information retrieval on the world wide web and active logic: A survey and problem definition. Technical Report, pages 1–45, 2002.[5] Bra, P. D., Houben, G.-J., Kornatzky, Y., and Post, R. Information retrieval in distributed hypertexts. In Proceedings of 4th International Conference on Intelligent Multimedia Information Retrieval Systems and Management (RIAO 94), Center of High International Studies of Documentary Information Retrieval (CID), pages 481–491, 1994.[6] Brin, S. and Page, L. The anatomy of large scale hypertextual web search engine. In Proceedings of 7thWorldWideWeb Conference, Elsevier Science, pages 107–117, 1998.

172

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

173

SECURE BANKING USING BLUETOOTH

B.Karthika D. Sharmila R.Neelaveni

II Year M.E.,(S.E) Professor Asst Professor

Bannari Aman Institute Bannari Amman Institute PSG college of

of Technology of Technology Technology

Sathyamnaglam. Sathyamangalam. Coimbatore.

[email protected] [email protected] [email protected]

ABSTRACT

Now a days mobile device are the

most popular payment mechanism for mobile payment system. Mobile devices are a pocket

sized computing device typically having a

display screen with touch input. Even though

mobile devices offer many advantages, it has been subjected to many security threats.

Bluetooth is an open wireless protocol for

exchanging data over short distances between the mobile devices. In the proposed m-

payment mechanism, Bluetooth is used as a

tool for transactions. The existing system provides a framework for m-transaction. In the

existing system, the transactions are done for

payments through using elliptic curve

cryptography algorithm. Instead of using the existing algorithm, in the proposed system

modified SAFER plus algorithm is used. It

provides very high speed, less encryption time and high data throughput. In the proposed

system, a secure m-cheque payment system

for macro payments is to be developed. In the

proposed system, the replacement of the existing elliptic curve cryptography (ECC)

algorithm with the modified SAFER plus

algorithm are to be done. Comparison of both existing algorithm and proposed algorithm is

to be done.

Key words: PDA, ECC, Modified SAFER+,

Bluetooth

1. INTRODUCTION

Bluetooth is an open wireless protocol for

exchanging data over short distances from

fixed and mobile devices. This allows users to

form ad hoc networks between a variety of devices to transfer voice and data. When two

or more Bluetooth devices, sharing the same

profile(s), come in range of one another, they establish a connection automatically. Once

Bluetooth devices are all connected, a network

is created. Bluetooth devices create a personal

area network (PAN), or commonly called a piconet. Bluetooth piconets are designed to

link up to eight different devices. A piconet

can be as small as a two foot connection between a keyboard and computer, or it can

encompass several devices over an entire

room. In order to regulate communications one of the participating devices is assigned the

role of master of the piconet, while all other

units become slaves. Masters have the duty of

directing and controlling communications, even between two slave devices. Under the

current Bluetooth specification, up to seven

slaves and one master can actively communicate. Furthermore, in order to extend

these networks, several piconets can be joined

together in what is known as a scatternet.

Today, all communication

technologies are facing the issue of privacy

and identity theft. Bluetooth technology is no exception. The information and data share

through these communication technologies

should be private and secure. Everyone knows that email services, company networks, and

home networks all require security measures.

Likewise Bluetooth communication too needs

some security mechanism where security is a major concern for example military, banking

and so on. Mobile payment is a new and

rapidly adopting alternative payment method. Instead of paying with cash, cheque, credit

174

cards and so on, a consumer can use a mobile

phone to pay for wide range of services. In the proposed system, we are going to combine

mobile payments through Bluetooth

technology. Here we are going to concentrate

in providing security when performing large amount of transactions

2. RELATED WORK

The study was made to examine

various factors for the acceptance of mobile technology model. This also includes

exploratory research on external factors

convenience, security, new technology that

affects mobile payment acceptance and use [2]. The payroll cards can serve as an

introductory financial product for consumers

who do not want to manage a checking account, but want the combined benefits of

direct deposit and a nationally branded debit

card.[4]. The document is all about a state of the art review of mobile payment technologies

[5]. It covers all of the technologies involved

in a mobile payment solution, including

mobile networks, mobile services, mobile platforms, mobile commerce and different

mobile payment solutions.

An electronic payment provides a new

payment technique which is represented and

transferable in electronic form. Book money

provided by the central bank and credit institutions has been represented digitally and

transferable through electronic networks for a

long time already[3].Don Johnson et al proposed the Elliptic Curve Digital Signature

Algorithm (ECDSA). The paper deals with

asymmetric digital signatures schemes where Asymmetric means that each entity selects a

key pair consisting of a private key and a

related public key This is all about digital

signatures integrated with elliptic curve cryptography. The author proposed a

framework for discussion of electronic

payment schemes. A comprehensive index of such schemes and a brief overview of their

relationship to the framework is provided. The

framework consists of two axes, the levels of abstraction at which the protocol is analyzed

and the payment model considered.

3. PROPOSED SYSTEM

In the proposed system, categorization

of an electronic payment according to the

amount of money conveyed from the payer (customer) to the payee (merchant) [1] is done

as

Pico payments for transaction volumes of less than Rs 1000

Micro payments for amounts between

Rs 1000 and Rs 100000 Macro payments for amounts

exceeding Rs 100000

Figure 1 depicts the block diagram of

the proposed system. All steps performed

in the proposed system are clearly

explained. In the proposed system, modified SAFER+ algorithm is used for

encryption and decryption of cheque. The

system provides high data throughput. Before the user initiates transaction, it

encrypts the bank to PDA and merchant to

PDA by invoking the authentication and

key exchange mechanisms. A principal must possess a valid digital certificate

signed by a trusted Certificate Authority

(CA).

The principals exchange certificates to

authenticate the customer identity to the merchant and bank.After the PDA

receives an invoice from the merchant, it

initiates a conversation with the bank. The

bank issues the cheque, signs it with its digital signature key, and sends it to the

PDA. Next, the PDA transfers the cheque

to the merchant, who verifies the bank’s signature and other information. Finally,

after the process’s successful completion,

the user receives a confirmation, and the merchant can subsequently clear the

cheque[1].In this system, all actors use a

Bluetooth enabled mobile device for

transaction. The laboratory setup has been made for all the actors which create

personal area networks.

175

Figure 1. Block diagram of the process

3.1 Steps in Proposed System:

The proposed system comprises of three communication sessions

PDA to merchant(via Bluetooth)

PDA to bank (via GPRS) Merchant to bank (via the Internet)

In the existing system, the elliptic

curve cryptography algorithm is used for generating the key. The key is used for the

authentication purposes. In the proposed

system, encryption and decryption are done using the modified SAFER plus algorithm.

Instead of using the existing algorithm, the

modified SAFER+ algorithm is used. It is efficient against some Bluetooth attacks

namely Blue jacking, Blue snarfing, man in

the middle attack and eavesdropping.

3.2 System Specification

JAVA is a programming language originally developed by James Gosling at Sun

Microsystems and released in 1995. The

language derives much of its syntax from C and C++ but has a simpler object model and

fewer low level facilities. My SQL is used as

a back end to store the data

3.3 Work Done

Figure 2 Invoice

Invoice from payee to payer is

depicted in Figure 2. The cheque validation

time and encryption algorithm is mentioned.

After the completion

of purchase, Invoice is

being issued to user

Invoice is being sent to

the bank

The cheque is verified

at the merchant side

using the same

The cheque is

encrypted using modified SAFER plus

algorithm

Once the bank verifies

the data, issues cheque

to the user

Confirmation is being

sent to the user

176

The amount requested as invoice to the payer

is also displayed. The payee is then asked for the amount of the purchased products by the

payer. It is then sent to the payer. In this the

process is started from the payee. Based on the

invoice only the cheque is requested. It is then encrypted at the bank and sent to the payer.

Then payer forwards cheque to the payee. At

the payee side, decryption of the cheque is done.

Figure 3. Cheque Request

Based on the invoice amount, the

cheque request is made by the payer to FSP

and is shown in Figure 3.The account number

and the amount is mentioned.

3.4 Conclusion and Future Work

In this system, the process of

implementing the Modified SAFER+

algorithm in Bluetooth networks is implemented. Yet to compare the performance

of ECC and modified SAFER plus algorithm,

REFERENCES

[1] Gianluigi Me, Maurizio Adriano

Strangio and Alexander Schuster (2006) “Mobile Local Macro

payments Security and Prototyping”,

Proceedings of IEEE Computer society.

[2] Dennis Viehland and Roslyn Siu

Yoong Leong (2007) “Acceptance and Use of Mobile Payments”,

www.acis2007.usq.edu.au/assets/pape

rs/108.pdf.

[3] K. Soramaki and B. Hanssens (2003) “E-Payments: What Are They and

What Makes Them Different?”,

www.e-pso.info/epso/papers/ePSO-DS-no1.pdf .

[4] S. Frumkin, W. Reeves, and B. Wides

(2005) “Payroll Cards: An Innovative Product for Reaching the Unbanked

and Underbanked”,

www.occ.treas.gov/cdd/payrollcards.p

df [5] David McKitterick and Jim Dowling

(2004) “State of the Art Review of

Mobile Payment Technology.” https://www.cs.tcd.ie/publications/tec

h...03/TCD-CS-2003-24.pdf

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

177

Issues in implementing E-learning in Secondary

Schools in Western Maharashtra

#1Hanmant N. Renushe,

#2Abhijit S. Desai,

#3 Prasanna R. Rasal

#1, 2,3 Bharati Vidyapeeth University, Y.M. Institute of Management Karad

#[email protected], #[email protected], #[email protected]

Abstract— this research paper highlights issues in implementing

E-learning in secondary schools in western Maharashtra. Today

enterprise networks are distributed to different geographical

locations and applications are more centrally located. This

enhancement offers new flexible opportunities and subsequently

increases efficiency of organisation’s performance. But

educational sector, especially primary and secondary level

schools are lagging behind in implementation of Information

technology in teaching- learning process. The Government of

India has formulated mission on education through ICT

[information and communication technology]. Still the focus of

this mission is to develop infrastructure and creation of the e-

content. Therefore researcher like to focuses on various issues in

implementation of E-learning methodology in secondary schools.

Keywords— E-learning, Asynchronous E-learning, Synchronous

E-learning, Internet, Radio, Cell phone, ICT

I. INTRODUCTION

In order to measure the performance of individual,

organization, state or country the key parameters are to be

considered such as Knowledge and adoption of new changing

scenario, and today world demands for teaching and learning

process as E-learning. Many organizations have been adopted

E-learning but Educational sector is still in budding stage. Govt. of India has formulated mission on education through

ICT, this program is designed to develop basic skills for

computer operations and focused on developing the

infrastructure. E-learning can be defined as learning by

making use of information technology, Communication

technology and network technology, the network

communication devices such as Radio, Television , Internet

tools , Mobile Phones can be used to in teaching learning

process. As on today in western Maharashtra every secondary

school is well equipped with computer peripherals and use of

these computer peripherals are limited to basic operations. In

every secondary school teacher uses traditional teaching methodologies [Blackboard-Chock]. E-Learning is a

combination of learning services and technology to provide

high value integrated learning; anytime, anyplace. It is being

resented in the marketplace as the next evolution of the

training and the education industry and the next phase in the

digital revolution.

II. CURRENT SCENARIO

It is noticed that most of the teachers are conscious about

the quality of their teaching. To enhance the quality, some

teachers use teaching aids, like, charts, models – static &

working, specimen, projection slides, etc. because teachers are

given training both in preparation and use of Audio-visual

Aids, but most of them do not have appropriate teaching aids.

The information technology is influencing in major area of human life and educational sector is not exception of

information technology. Therefore the Govt. of Maharastra

has taken a step forward by implementing ICT enabled

education. The focus of the ICT enabled Mission is to develop

infrastructure and creation e-content. Instead of all such

efforts, still the methodology remains as the traditional one.

Presently in various schools of Western Maharashtra, some

measures are being taken for creating awareness of

information technology amongst the students by starting

Computer education from secondary level in the curriculum.

Despite of these efforts initiated by Government of

Maharashtra, there is immense need of implementing an E-learning Framework in Secondary School Education System.

For which some following issues are need to be resolved, so

that an effective E-learning framework can be implemented

for the utmost benefit of the students.

A. Issues in implementing E-learning

1) Infrastructure Unavailability

Presently most of the schools in urban area are having

computer laboratories with sufficient computers along with printers, but in rural area most of the schools are not having

well equipped computer laboratories. In addition, these

laboratories are not maintained properly on regular basis.

One more aspect is unavailability of internet connectivity in

schools, creates gap between a Information Warehouse and

students of schools. [1]

2) Teachers Participation

There is an ongoing debate as to whether teachers are

becoming redundant as a consequence of the use of ICT in

education or whether an e-classroom is just a myth. The role

of teachers has changed and continues to change from being

178

an instructor to becoming a constructor, facilitator, coach, and

creator of learning environments. Most of teachers goes with

conventional methods for teaching, resulting an obstacle in e-

learning implementation.[2] This unchanging attitude of

teachers is affecting in adoption of e-learning process.[3]

3) Students Attitude

There have been studies undertaken to understand the

factors of the students in acceptance of e-learning, which

emphasizes the importance of learning style of students in

teaching learning process. Lack of interest to use

technological advancements in learning process by students is

the key issue in e-learning implementation

4) Absence of a unique ICT portal for specific area

Presently, there is absence of a unique ICT portal, through

which a common platform of education would have been

provided to all students in the languages widely spoken.

5) Policy Implementation

Government of Maharashtra has taken adequate steps to

frame the educational policies for Secondary education, but

the necessary actions are being taken for effective

implantation of the same.

There are other several issues needed to be addressed, such

as lack of involvement of parents,

III. NEW EDGE METHODOLGY

Information and Communication Technology [ICT] uses

the hardware and software for efficient management of

information, i.e. storage, retrieval, processing, communication,

and sharing of information. The ICT leads to development of Websites of Government, Industries, Educational sector, etc.

The information in audio, video or textual format can be

transmitted to the users and it created opportunities like,

Online learning, e-learning, online movie. The 3G Mobiles are

also part of ICT. 3G Mobile phone provides e-mail facility,

Compact Application, and compact web browsers. Therefore

these 3G mobiles can be accessed anywhere. The ICT brings

more rich material in the classrooms and libraries for the

teachers and students. It has provided opportunity for the

learner to use maximum senses to get the information. Here

we are specifying some digital equipment for teaching and learning programs

A. Television

Now a days the children’s are spending most of the time in

watching television programs, broadcasting the educational

material on specific channel at prime time can beneficial to

the students. Some channels such as National Geographic,

Discovery, Animal Planet and TATA Sky’s Learner Kit

broadcasting educational material. IGNOU, New Delhi

conducts online classes for their various classes at GYANDARSHAN channel. On this channel, various lectures

are conducted online and students can acquire knowledge

irrespective of location.

B. 3G Mobile

The mobile phones are becoming an important commodity

nowadays. It can be taken away to anywhere and it leads to

opportunity to make use of mobiles teaching and learning

process. The 3G mobile phones can process audio and video

data at very high speed and high quality. They are having

multiple applications such as word, excel PowerPoint, and

slide share and adobe reader so the user can listen, watch or

read the e-content related to education circulated by the school

online.

C. Internet2

it is well known that one can get maximum knowledge

using internet, therefore it is better to introduce to all students.

This can be done by introducing IT related tools in school

education. The Web Based Learning has the potential to meet

the perceived need for flexible pace, place. The web allows

education to go to the learner rather than the learner to their

education.

IV. FRAMEWORK OF E-LEARNING IN SCHOOL

Every industry in India trying to use the digital revolution, and as the part most of the organization adopted the new edge

scenario i.e. information and communication technology

[ICT]. To use the ICT in Education especially in secondary

school we proposed the given model.

Fig1. E-learning Framework for Secondary School

This framework is formed using database server, web

server and the students as E-learner user.

A. Database server.

The server component provides a function or service to one

or many clients, and A Database server is a server that

provides database services to other computer programs. There

are several kinds of database server available such as Flat File

Database Server, Relational Database Server, Object Database

Server, and Object Relational Database Server, Spatial

database Server, Temporal Database Server etc. The Learning

Content can be stored on database server viz Photograph,

Animated lessons, Audio speeches, Video lessons, Question

179

banks, frequently Asked Questions Laboratory Manuals, and

e-books etc.

B. Web Server

Web server is the computer on the internet which serves

web pages on the request of internet users. Every web server

has the unique IP address and other devices such as Computer,

3G Mobiles, IP Television, and PDA can communicate with

web server using IP addresses. The E-learning web portal resides on this web server and provides access to various

schools who likes to get benefitted by this e-learning project.

C. School MIS

An 'MIS' is a planned system of the collecting, processing,

storing and disseminating data in the form of information

needed to carry out the functions of management. In a way it

is a documented report of the activities those were planned

and executed.[5] In schools, a uniform Management Information systems[ MIS ] can be implemented in each

school, which will provide access to the necessary multimedia

information from the central database server through web

server. This MIS also will play role in establishing

communication between students and e-learning system.

D. 3G Mobile

The services offered in 3G mobile has ability to make voice

call, download information, and email processing and IM

(Instant Messaging) anytime and anywhere. These mobiles are supporting for IP calling and IP Connectivity with the

computer. E-content can be taken from web server by using

school MIS.

E. Web Sites

The E-content can be accessed by the student in the form

of web pages. The well formed web pages such as content of

textbook in the form animation, Video or text format class

wise can be accessed from database server using web server.

F. IP Television

Now as days most of the television set supports to IP

address. Reliance Communication has launched IPTV in

Mumbai, which includes services such as Internet, telephone

and TV Channels. The well prepared E-content which has

been stored with database server can be broadcasted to

television. The learning ability increases with audio-visual

aids. So this can be achieved using IP TV.

G. IPOD

One can use IPOD as a drive to store and transfer data files.

As it is handheld device stored data can read the content in

textual format, hear the audio or watch the content in the form

of Video or animated film.

H. Web Radio

E-radio can be used to update students about school activities and latest happening in education. In this context

Ramnarian Ruia College in central Mumbai has launched

campus e-radio station to keep update of their students

regarding college update and latest happing in the city [4]

V. CONCLUSIONS

In the nutshell, it is observed that there are various issues in

implementation of E-learning in secondary schools in western

Maharashtra. These issues can be addressed by using various

new edge methodologies in practice. [6]The Government of Maharashtra has already taken the

initiative, but the school managements can be motivated by

allocating special grants, financial aid, to build the necessary

infrastructure such as well equipped computer laboratory,

videoconferencing, broadband internet connectivity, net

meeting setup, setting up web radios and IP Television

channels etc. In addition to this, The Government can motivate to NGO’s [Non Government Organisation],

Charitable Trusts, And Voluntary bodies in building necessary

state-of-art infrastructure. Again one more way is to build

infrastructure is to introduce Build-Operate-Transfer [BOT]

schemes in this project. So that private firms can come into

play and they will bring the necessary speed in development

of infrastructure with world class quality.

Change is the universal law of nature. Every object used to

change over the time. And resistance to change is the

phenomenon which always goes along with this change. Same

thing happens with the teachers of secondary schools. Most of them are reluctant to change or enhance their teaching

methodologies. An integrated programme can be organized

by Education Department of Government of Maharashtra to

increase the level of interest of teachers in teaching learning

process selected for E-learning project. In such programme,

teacher’s participation about new edge methodologies /

technologies can be increased by reducing their resistance to

changes in teaching methods of E-learning Framework. It can

be achieved by increasing teacher’s awareness about new edge

technologies, ease of usage, and increase in effectiveness of

teaching methodologies.

The most important object, Student for whom everybody is concerned, is required to be motivated to actively participate

in effective implementation of E-learning Framework.

Various activities can be undertaken to increase the interest

amongst the student, so that they can be really benefitted by

the implementation of E-learning framework.

There must be uniqueness in E-learning portal should be

developed and updated on regular basis. For this perpetual

activity, there is necessity of an integrated information Cell,

which can lead to regular revision and updating the E-content

such as study material in the form of multimedia, tutorials,

quiz’s, online tests, etc. Last but not the least, a firm policy should be framed and

strictly implemented with the goals and objectives of the

E-learning Project. Again Progress of the E-learning Project

should be continuously tracked and necessary changes can be

made.

180

ACKNOWLEDGMENT

The researchers are grateful to the authors, writers

and editors of the book and articles, which have been referred for preparing the presented research paper.

It is the duty of the researchers to remember their

parents whose blessings are always with them.

REFERENCES

[1] http://home.nic.in/nicportal/projects.html

[2] http://learningonlineinfo.org/2006/07/02/teachers-role-and-ict-in-

education/

[3] Farida Umrani-Khan, shridhar Ayyar, ‘ELAM: A Model of Acceptance

and Use of E-learning by Teachers and Students.

[4] http://www.ruiacollegeradio.com/

[5] http://en.wikipedia.org/wiki/Management_information_system

[6] E-World-Book by Arpita Gopal, E-Learning

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

181

AN ARTIFICIAL DEVICE OF NEUROSCIENCE

ANANDRAJ.R

SAKTHI MARIAMMAN ENGINEEERING COLLEGE DEPARTMENT OF COMPUTER SCIENCE ENGINEERING

E-MAIL ID: [email protected] MOBILE NO: 9884107283

ABSTRACT:

It is a ―device which implements the brain

through the system‖. It is designed to help those

who have lost the control of their limbs, or other

bodily functions. ―The goal of the device is to

program to develop a fast, reliable and unobtrusive

connection between the brain of a severely disabled

person and a personal computer and with bionic

hands‖.

A ‗brain-computer‘ interface that uses

internal neural signal sensor and external processors

to convert neural signals into an output signal under

the users own control. It is direct communication

between the human brain and an external device (say

computer). This system is based on Cyber kinetics'

platform technology to sense, transmit, analyze and

apply the language of neurons. The System consists

of a sensor that is implanted on the motor cortex of

the human brain and a device that analyzes brain

signals. The principle of operation behind this is that

with intact brain function, brain signals are generated

even though they are not sent to the arms, hands and

legs. This device can provide paralyzed or motor-

impaired patients a mode of communication through

the translation of thought into direct computer

control to the bionic hands.

At last we conclude that the invention of this

system hopped ―to provide paralyzed individuals

with a gateway through which they can access the

broad capabilities of computers, control devices in

the surrounding environment, and even move their

own limbs‖.

INTRODUCTION:

• The device is a brain implant system.

• Designed to help those who have lost

control of their limbs, or other bodily

functions.

It consists of a surgically implanted sensor that

records the activity of dozens of brain cells

simultaneously. The system also decodes these

signals in real time to control a computer or other

external devices.

CHIP SENSOR PROCESS:

The sensor consists of a tiny chip (smaller

than a baby aspirin) with one hundred electrode

sensors each thinner than a hair that detect brain

cell electrical activity. Sensors, or electrodes, on

the silicone chip detect signals from surrounding

neurons in the brain‘s motor cortex. This area is

highly saturated with neurons, but each sensor only

needs to detect signals from 10 to 50 neurons to

trigger the system to move the computer cursor.

The sensors act as facilitators for the message,

which is carried out by the computer.

182

The electrode translates brain activity and

sends the information to the computer in a format it

can understand. The System is based on Cyber

kinetics platform technology to sense, transmit,

analyze and apply the language of neurons. The

System consists of a sensor that is implanted on the

motor cortex of the brain and a device that

analyzes brain signals. The principle of operation

behind the System is that with intact brain

function, brain signals are generated even though

they are not sent to the arms, hands and legs.

The signals are interpreted and translated into

cursor movements, offering the user an alternate to

control a computer with thought, just as individuals

who have the ability to move their hands use a

mouse.

THE DESIGN OF A DEVICE:

It focusing on motor neuroprosthetics

aim to either restore movement in paralysed

individuals or provide devices to assist them, such

as interfaces with computers or robot arms.

Researchers at Emory University in Atlanta

led by Philip Kennedy and Roy Bakay were first to

install a brain implant in a human that produced

signals of high enough quality to simulate

movement. Their patient, Johnny Ray, suffered

from ‗locked-in syndrome‘ after suffering a brain-

stem stroke.

MOTOR CORTEX:

The motor cortex appears to be par

excellence a synthetic organ for motor acts the

motor cortex seems to possess, or to be in touch

with, the small localized movements as separable

units, and to supply great numbers of connecting

processes between these, so as to associate them

together in extremely varied combinations. The

acquirement of skilled movements, though

certainly a process involving far wider areas of the

cortex than the excitable zone itself, may be

presumed to find in the motor cortex an organ

whose synthetic properties are part of the

physiological basis that renders that acquirement

possible‖.

THE INTERFACE:

It sometimes called a direct neural interface

or a brain-machine interface is a direct

communication pathway between a human or

animal brain (brain cell culture) and an external

device. In one-way, computers either accept

commands from the brain or send signals to it (for

example, to restore vision) but not both. Two-way,

would allow brains and external devices to

exchange information in both directions but have

yet to be successfully implanted in animals or

humans. In this definition, the word brain means

the brain or nervous system of an organic life form

rather than the mind. Computer means any

processing or computational device, from simple

circuits to silicon chips (including hypothetical

future technologies such as quantum computing).

EXISTING VERSUS PROPOSED:

Neuroprosthetics is an area of neuroscience

concerned with neural prostheses—using artificial

devices to replace the function of impaired nervous

systems or sensory organs. The most widely used

neuroprosthetic device is the cochlear implant,

which was implanted in approximately 100,000

people worldwide as of 2006. There are also

several neuroprosthetic devices that aim to restore

183

vision, including retinal implants, although this

article only discusses implants directly into the

brain. The differences between Neuro and

neuroprosthetics are mostly in the ways the terms

are used: neuroprosthetics typically connect the

nervous system, to a device, whereas ―My system‖

usually connects the brain (or nervous system) with

a computer system. Practical neuroprosthetics can

be linked to any part of the nervous system, for

example peripheral nerves, while the term "Neuro"

usually designates a narrower class of systems

which interface with the central nervous system.

The terms are sometimes used interchangeably and

for good reason. Neuroprosthetics and Neuros seek

to achieve the same aims, such as restoring sight,

hearing, movement, ability to communicate, and

even cognitive function. Both use similar

experimental methods and surgical techniques.

NON-INVASIVE NEURO’S:

As well as invasive experiments, there have

also been experiments in humans using non-

invasive neuro imaging technologies as interfaces.

Signals recorded in this way have been used to

power muscle implants and restore partial

movement in an experimental volunteer. Although

they are easy to wear, non-invasive implants

produce poor signal resolution because the skull

dampens signals, dispersing and blurring the

electromagnetic waves created by the neurons.

Although the waves can still be detected it is more

difficult to determine the area of the brain that

created them or the actions of individual neurons.

Recordings of brainwaves produced by an

electroencephalogram

PARTIALLY-INVASIVE NEUROS:

Partially invasive Neuro devices are

implanted inside the skull but rest outside the brain

rather than amidst the grey matter. They produce

better resolution signals than non-invasive Neuros

where the bone tissue of the cranium deflects and

deforms signals and have a lower risk of forming

scar-tissue in the brain than fully-invasive Neuros.

Electrocorticography (ECoG) measures the

electrical activity of the brain taken from beneath

the skull in a similar way to non-invasive

electroencephalography, but the electrodes are

embedded in a thin plastic pad that is placed above

the cortex, beneath the dura mater. ECoG

technologies were first trialed in humans in 2004.

ECoG is a very promising intermediate Neuro

modality because it has higher spatial resolution,

better signal-to-noise ratio, wider frequency range,

and lesser training requirements than scalp-

recorded EEG, and at the same time has lower

technical difficulty, lower clinical risk, and

probably superior long-term stability than

intracortical single-neuron recording. This feature

profile and recent evidence of the high level of

control with minimal training requirements shows

potential for real world application for people with

motor disabilities.

Electroencephalography (EEG) is the most

studied potential non-invasive interface, mainly

due to its fine temporal resolution, ease of use,

portability and low set-up cost. But as well as the

technology's susceptibility to noise, another

substantial barrier to using EEG as a brain-

computer interface is the extensive training

required before users can work the technology. For

example, in experiments beginning in the mid-

1990s, Niels Birbaumer of the University of

Tübingen in Germany used EEG recordings of

184

slow cortical potential to give paralysed patients

limited control over a computer cursor.

Magnetoencephalography (MEG) and functional

magnetic resonance imaging (fMRI) have both

been used successfully as non-invasive BCIs. In a

widely reported experiment, fMRI allowed two

users being scanned to play Pong in real-time by

altering their haemodynamic response or brain

blood flow through biofeedback techniques. fMRI

measurements of haemodynamic responses in real

time have also been used to control robot arms

with a seven second delay between thought and

movement.

ABOUT NEURO DEVICE:

The Neural Interface Device is a

proprietary brain-computer interface that consists

of an internal neural signal sensor and external

processors that convert neural signals into an

output signal under the users own control. The

sensor consists of a tiny chip smaller than a baby

aspirin, with one hundred electrode sensors each

thinner than a hair that detect brain cell electrical

activity. Neuro consists of a surgically implanted

sensor that records the activity of dozens of brain

cells simultaneously. The system also decodes

these signals in real time to control a computer or

other external devices. In the future, Neuro could

control wheelchairs or prosthetic limbs. The long-

term goal: Pairing Neuro with a muscle stimulator

system – which would allow people with paralysis

to move their limbs again.

The Neuro technology platform was

designed to take advantage of the fact that many

patients with motor impairment have an intact

brain that can produce movement commands. This

may allow the Neuro system to create an output

signal directly from the brain, bypassing the route

through the nerves to the muscles that cannot be

used in paralyzed people.

NEURONEXUS’ CORE ADVANTAGES: NeuroNexus Technologies offers a variety

of probe designs. The extensive design freedom

offered by the micromachining technology has

resulted in a variety of different probes which

satisfy the needs of most investigators.

NeuroNexus's probes have the following

advantages over conventional microelectrodes:

One shank, with multiple sites, that is about

the size of a conventional wire electrode

Batch fabricated

High reproducibility of geometrical shape,

electrical properties and mechanical

properties

Easy customization of site placement and

substrate shape

Small size, resulting in minimal

displacement of the neural tissue

High spatial resolution at various depths up

to 1cm

Multiple parallel shanks, providing

horizontal spatial sampling

Independent recording/stimulation among

sites

Higher data output with fewer animals

TRACT OF IMPLEMENTATION:

By overcoming the cons of robotic arms

used in robotic technology and to sophistication

use of paralyzed people in individual.

Movement signals persist in the primary

motor cortex, the area of the brain

responsible for movement, long after a

spinal cord injury.

spiking from many neurons – the language

of the brain – can be recorded and routed

outside the human brain and decoded into

command signals;

185

Paralyzed humans can directly and

successfully control external devices, such

as a computer cursor and robotic limb,

using these neural command signals.

NEURON MOVEMENTS:

EARLY WORK:

ROBOTIC TECHNOLOGY:

DRAWBACKS FOUNDED:

Additional device.

Movement inaccuracy.

Environmental uncertainty.

Sensor inaccuracy.

Monkeys have navigated computer cursors

on screen and commanded robotic arms to perform

simple tasks simply by thinking about the task and

without any motor output. Other research on cats

has decoded visual signals.

NEURO NEURAL INTERFACE SYSTEM:

Neuro Neural Interface Device- a ‗BRAIN-

MACHINE‘ interface that uses internal neural

signal sensor and external processors to convert

neural signals into an output signal under the users

own control.

HOW IT WORKS:

The chip is implanted on the surface of the

brain in the motor cortex area that controls

movement. In the pilot version of the

device, a cable connects the sensor to an

external signal processor in a cart that

contains computers. The computers

translate brain activity and create the

communication output using custom

decoding software. Importantly, the entire

Neuro system was specifically designed for

clinical use in humans and thus, its

manufacture; assembly and testing are

intended to meet human safety

requirements.

• Operation behind the Neuro System is that

with intact brain function, brain signals are

generated even though they are not sent to

the arms, hands and legs.

• The signals are interpreted and translated

into cursor movements, offering the user an

alternate "Neuro pathway" to control a

computer with thought, just as individuals

who have the ability to move their hands

use a mouse.

Neuron 1 Neuron 3 Neuron 6 Neuron 8 Neuron 9 Neuron 15 Neuron 19 Neuron 20 Neuron 27

Neuron 29 Neuron 30 Neuron 31 Neuron 32 Neuron 36 Neuron 37 Neuron 38 Neuron 39 Neuron 40

Neuron 41 Neuron 45 Neuron 47 Neuron 48 Neuron 49 Neuron 50 Neuron 51 Neuron 52 Neuron 53

186

THE BRAINS BEHIND NEURO DEVICE:

1. The person thinks about moving the

computer cursor. Electrodes on a silicone chip

implanted into the person‘s brain detect neural.

Activity from an array of neural impulses in the

brain‘s motor cortex.

2 .The impulses transfer from the chip to a

pedestal protruding from the scalp through

connection wires.

3. The pedestal filters out unwanted signals or

noise, and then transfers the signal to an amplifier.

4. The signal is captured by an acquisition

system and is sent through a fiber-optic cable to a

computer. The computer then translates the signal

into action, causing the cursor to move.

BENEFITS:

Can provide paralyzed or motor-impaired

patients a mode of

Communication through the translation of thought

into direct computer control. ―The goal of the

Neuro program is to develop a fast, reliable and

unobtrusive connection between the brain of a

severely disabled person and a personal computer‖.

APPLICATIONS:

Navigate Internet.

Play Computer Games.

Turn Lights On and Off.

Control Television.

FUTURE APPLICATIONS:

Future implications are mind blowing.

We can have a robot completely

controlled by the thoughts which can do

any work and obviously very much

effectively.

There won’t be any masterminds like

astrophysics “stephen hawkings”

paralysed.

It can be a boon not only to the paralysed

persons but also to the entire humanity.

CONCLUSION:

Depending on patient‘s condition and

disability, a remedy is available and by the years

end it will be commercialized. On the other side, it

might be interesting to contact company and offer

assistance with development and testing as they are

actually working in the dark, without proper user

feedback they are limited. At last we conclude that

the invention of brain gate is been hopped ―to

provide paralyzed individuals with a gateway

through which they can access the broad

capabilities of computers, control devices in the

surrounding environment, and even move their

own limbs".

REFERENCES:

187

1. Monkey Moves Robot Using Mind Control,

Sky News, July 13, 2009

2. [http:/www.braingate2.org]

3. J. Vidal, "Real-Time Detection of Brain Events

in EEG", in IEEE Proceedings, May 1977, 65-

5:633-641.

4. S. P. Levine, J. E. Huggins, S. L. BeMent, R.

K. Kushwaha, L. A. Schuh, M. M. Rohde, E.

A. Passaro, D. A. Ross, K. V. Elisevich, and B.

J. Smith, "A direct brain interface based on

event-related potentials," IEEE Trans Rehabil

Eng, vol. 8, pp. 180-5, 2000.

5. Weymouth man, a volunteer in experiments,

dies at 27 July 25,

2007.[http://www.boston.com/news/globe/city_

region/breaking_news/2007/07/weymouth_ma

n_a.html]

6. Mind Control March 2005.

[http://www.wired.com/wired/archive/13.03/br

ain.html]

7. "Brain chip reads man's thoughts". BBC News

[http://news.bbc.co.uk/2/hi/health/4396387.stm

].

8. News about matt nagle.

[http://en.wikipedia.org/wiki/Matt_Nagle#cite_

ref-0].

9. News about Jens Naumann.

[http://www.seeingwithsound.com/etumble.htm

].

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

188

REMOTE PATIENT MONITORING

E.Arun1 ,V.Marimunthu

2 ,D.Sharmila

3 and R.Neelaveni

4

IV Year, B.Tech, Information Technology,

Bannari Amman Institute of Technology, Sathyamangalam.

Email: [email protected] 1 and [email protected]

2

Contact no: +91 99768 356911

ABSTRACT

The problem found in most hospitals

is that the physician has to frequently visit

the patient and asses his/her condition by

measuring the parameters such as

temperature, blood pressure, drip level etc.

In case of emergencies, the nurse intimates

the doctor through some means of

communication like mobile phone. A

growing selection of innovative electronic

monitoring devices is available, but

meaningful communication and decision

supports are also needed for both patients

and clinicians.

The aim is to develop a reliable,

efficient and easily deployable remote

patient monitoring system that can play a

vital role in providing basic health services

to the patients. This system enables expert

doctors to monitor patients in remote areas

of hospital. Mobile phones or personal

digital assistants (PDAs) with wireless

networking capabilities may serve as

gateways that process, store, and transfer

measured parameters to clinicians for further

analysis or diagnosis. The timely manner of

conveying the real time monitored

parameter to the doctor is given high priority

which is very much needed.

1 INTRODUCTION

Telecommunications has the

potential to provide a solution to medical

services to improve quality and access to

health care regardless of geography. The

advances in information and communication

technologies enable technically, the

continuous monitoring of health related

parameters with wireless sensors, wherever

the user happens to be. They provide

valuable real time information enabling the

physicians to monitor and analyze a

patient‟s current and previous state of

health. To remove or minimize cables, a

robust wireless link with low power

capabilities is required. Although many

wireless standards can be used, there are

important considerations such as range,

189

throughput, security, ease of implementation

and cost. Bluetooth is quickly becoming the

preferred technology for wireless patient

monitoring (Noel Baisa, 2005). A major

motivation for reducing the number of cable

or for doing away with the cables

completely is to eliminate the potentially

harmful currents to the patient. The patient

monitoring involves handling of sensitive

data. These data should be transmitted

securely without any intrusion. Even a

Bluetooth enabled gadget can try to

participate in the network which introduces

risk.

3.1 Objective

The objective of this work is to

create an automated patient monitoring

system for the patient in hospital. Presently,

the care is being provided by nurse, who

performs all the steps of patient care

manually. They take readings of patient‟s

physiological data using instruments which

are difficult to handle and require manual

tuning etc. Then, they record this data into

printed forms manually. Finally, the

collected forms are sent to a doctor who

goes through them looking for any symptom

of abnormality. The doctor then takes

decision regarding the patient‟s treatment.

The automated system that replaces all this

hectic activity, should be able to gather the

physiological data, transmit it, record it, find

any abnormality and then assist the doctor in

the decision making process.

3.2 Mode of monitoring

The mode of monitoring the patient

is determined by the nature and seriousness

of disease. A patient with mild problems

needs to be monitored only at periodic

intervals, whereas a critical patient must be

under constant monitoring. Remote

monitoring systems in general, are divided

into two basic modes of operations namely

periodic checkup mode and continuous

monitoring mode.

3.2.1 Periodic checkup mode

A typical patient under normal

circumstances can be monitored at periodic

intervals. The periodic checkups will be

performed by the health worker. In this

mode, it is rather easy to establish a reliable

and error free communication channel that

can preserve all relevant characteristics of

the transferred medical signals, regardless of

the communication service.

3.2.2 Continuous monitoring mode

Continuous monitoring mode is

required for critical patients. In this mode,

190

the patient‟s physiological data is under

constant surveillance. However, the doctors

suggest that the patients in critical condition

are admitted to the hospital instead of being

provided with remote monitoring. This

mode is particularly useful in emergency

scenarios.

3.3 Infrastructure Requirement

A rapidly increasing number of

health care professionals now believe that

wireless technology will provide accurate

data with improved patient care. A wireless

telecare system is needed along the patient

bedside to provide good health care. The

system must be interactive. The physician is

needed to send the suggestion back about

the patient to the nurse so that nurse can take

immediate action. The proposed system

consists of three main blocks. They are

mentioned in the block diagram shown in

Figure 1

The system is constructed such that it

measures the biomedical parameters at the

patient end without any intervention and

then transmits the data acquired to the

remote station. Sensors play an important

role in monitoring the parameters. The

precision at which the handheld device is to

be operated is decided by the sensor that is

used. Vital parameters such as temperature,

glucose level, ECG, drip level, pulse rate

can also be measured. The measured value is

then transmitted to the base station via

Bluetooth. The Temperature sensor is used

for sensing the temperature. This system is

constructed with low power consumption so

that it would not cause much hindrance to

the patient. The device is constructed such

that it transmits the vital information

periodically say 2 minute.

3.4 Working Model

Each patient is connected with a

temperature sensor. Measured parameters of

patients are interfaced with the system at the

patient end. The patient end system is

connected with server and doctor mobile via

Bluetooth. The server stores the central

database of all the patients. The status of the

parameter is decided in the patient end

system. If the status is normal, the parameter

is transmitted to the server and entered in the

database. If the status is abnormal, then the

parameter is immediately intimated to the

doctor end and also the data is stored in the

database of the server and is depicted in

Figure 1

.

191

Figure 1 Working Model

3.4.1 Temperature Sensor

The LM35 series are a precision

integrated circuit temperature sensor, whose

output voltage is linearly proportional to the

Celsius (Centigrade) temperature. The

LM35 has an advantage over linear

temperature sensors calibrated in ° Kelvin,

as the user is not required to subtract a large

constant voltage from its output to obtain

convenient centigrade scaling.

Figure 2 shows the block diagram of

temperature sensor module. The measured

temperature is in analog. The analog value is

converted into digital value for further

processing using 8051 microcontroller. The

digital data is transmitted to the PC through

MAX 232 driver and RS 232 serial

interface. The recorded temperature is saved

in a database.

Figure 2 Block Diagram of Temperature

Sensor Module

Temperature

Sensor

8051

Microcontroller

MAX 232

Driver

RS 232 Interface

of Computer

Temperature

Sensor

Patient end

System

Central Server

Doctor’s Mobile

192

Figure 3 Recording Setup

Figure 3 shows the recording setup.

The LM35 sensor is attached to the patient.

The data is received from the patient end in

ASCII form. The „Bluetooth Hospital

Client‟ does the conversion of ASCII values

into decimal values. The application reads

from the serial communication (COM) port.

3.4.2. Patient End System

Each patient at the Intensive Care

Unit (ICU) is provided with a computer

system, that has ‟Bluetooth Hospital Client‟

application. This application reads the

temperature sensor through the interface.

The patient parameters like patient

identification number (PID), temperature,

heart beat, sugar, inhale, exhale and blood

pressure can be transmitted to the server

from the patient end system. The user

interface of the patient end system is

inferred in Figure 4

Figure 4 Patient End System

The Bluetooth Hospital Client

application has two menus: File and Client.

The File Menu has a sub menu „Exit‟. Exit

submenu is used to close the application.

Client has two submenus: „ServerIP‟ and

„Start‟. „ServerIP‟ submenu is used to

connect with the central server and „Start‟

submenu initializes the connection.

Authentication is done after clicking the

start submenu and entering a valid pass key.

This pass key is a secret shared between

client and the server. This process is needed

to pair the client with the server. The various

patient parameters are given in the user

193

interface. The status of the patient is

displayed in the user interface.

Figure 5 Status Window

Figure 5 shows the various

parameters of the patients with the status

displayed as normal or abnormal. The

unique ID of the patient is displayed in PID

field. The temperature of the patient is

displayed in Fahrenheit. If the temperature

of the patient is less than or equal to 99 o

F,

then the status of the patient is normal. If the

temperature of the patient is greater than 99

o F, then the status of the patient is abnormal.

The readings recorded at 2 minute intervals

is displayed in Figure 6

Figure 6 Database in Patient System

This work can be extended to

measure other parameters by interfacing the

respective sensor .The work in this thesis is

done only for temperature so other fields are

displayed as NA. The data collected for a

period is stored in client database. The

database consists of patient ID, patient

name, time, temperature and the status

3.4.3 Server

Bluetooth enabled server is the

centralized system. An application

„Bluetooth Hospital Server‟ is present in the

server. This application is used to store the

various details of the patients. The server is

always active. Server updates the patient

details from each patient end system and

194

provides authentication for each patient end

system.

The server application has two

menus namely „File‟ and „Bluetooth‟. „File‟

has a submenu „Exit‟. Exit submenu is used

to close the application. „Bluetooth‟ menu

has a submenu „start‟. This start menu is

used to begin the application. A passkey is

provided to disable the unauthorized access.

The same pass key is used in the patient end

system in order to participate in the network.

After authentication, the server displays the

details of any received patient data and the

attack list.

The server displays as „Bluetooth

Started‟ when server become active. The

data is encrypted in the patient end system

(client system). The encrypted information

is transmitted to the server via Bluetooth.

Server decrypts the data and the decrypted

information is stored in the server. The

message is displayed in the server window

as shown in Figure 7.

Figure 7 Server Interface

The display consists of a block that

displays the received patient information.

IP24342 is the Identification number of the

patient. 100 o F

is the measured temperature

of the patient. NAs are other patient

parameters which are not applicable in this

work.

The server database stores the

parameters of all patients in ICU as shown

in Figure 8. It displays the recent updates of

each patient

195

Figure 8 Server Database

3.4.4 Mobile of Doctor

An application „Bluetooth Hospital

Mobile‟ is present in the Bluetooth enabled

mobile of the doctor. This application

presents the critical information of the

patient as shown in Figure 9. This mobile

application has two menus, namely „File‟

and „Client‟. „File‟ is used to close the

application. „Client‟ has two sub menus:

Server IP and Start. The server IP submenu

is used to connect with the server. Start

submenu is enabled after providing server IP

address, and now the mobile is connected

with the server. Start menu starts the

application and the details of the patient

with unique ID are displayed. Search option

is also provided in this application. Doctor

can use the search option to know the details

of the patients.

Figure 9 Mobile Interface

The mobile displays the PID and the

monitored parameters. Temp is the

196

temperature of the patient measured in

Fahrenheit. The analysis is made at the

patient end system. If abnormal, the doctor

receives the patient information in his

mobile.

4 CONCLUSION

The field of telemedicine has seen a

tremendous technical development in the

developed countries. Sustained efforts are

also underway in the developing countries

like China, Egypt and India in this field. The

timely manner of conveying the real time

monitored parameter to the doctor and

Providing 24 hours per day supervision

increases the costs of treatment. A real time

patient monitoring system is implemented

for transmission of patient data to doctor‟s

mobile

REFERENCES

V.Thulasi Bai et al, “Design and

implementation of mobile telecardiac

system” in the Journal of Scientific &

Industrial Research Vol. 67, December

2008, pp. 1059-1063 , 2008.

V.Thulasi Bai et al, “Wireless Tele Care

System for Intensive Care Unit of Hospitals

Using Bluetooth and Embedded

Technology” in Information technology

Journal at Asian Network for Scientific

Information, 2006.

K. E. B. Wallin et al, from the Department

of Anesthesiology and Intensive Care,

Karolinska Hospital, Sweden, “Evaluation

of Bluetooth as a Replacement for Cables in

Intensive Care and Surgery”, 2004.

M. Schwaibold, “A Wireless, Bluetooth-

Based Medical Communication Platform”,

2003.

Michaël Setton et al, "“Bluetooth sensors for

wireless home and hospital healthcare

monitoring paper”, 2006.

Emil Jovanov et al, Electrical and Computer

Engineering Department, University of

Alabama, “Patient Monitoring Using

Personal Area Networks of Wireless

Intelligent Sensors” 2004.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

197

An Efficient Search in

Unstructured Peer - to - Peer Networks

1A.TAMILMARAN,

PG Scholar,

Department of CSE,

Kumaraguru College of Technology,

Coimbatore – 641 006

Email: [email protected]

2Mr.V.VIGILESH,

Senior Lecturer,

Department of IT,

Kumaraguru College of Technology,

Coimbatore – 641 006

Email: [email protected]

Abstract

Peer-to-peer (P2P) networks

establish loosely coupled application-level

overlays on top of the Internet to facilitate

efficient sharing of resources. They can be

roughly classified as either structured or

unstructured networks. Without stringent

constraints over the network topology,

unstructured P2P networks can be

constructed very efficiently and are therefore

considered suitable to the Internet environment. However, the random search

strategies adopted by these networks usually

perform poorly with a large network size. In

this paper, we seek to enhance the search

performance in unstructured P2P networks

through exploiting users’ common interest

patterns captured within a probability-

theoretic framework termed the user interest

model (UIM). A search protocol and a routing

table updating protocol are further proposed

in order to expedite the search process

through self organizing the P2P network into a small world.

Index Terms - Unstructured peer - to - peer

network, self organization, user interest model

1 INTRODUCTION

(P2P) networks establish loosely coupled

application-level overlays on top of the Internet

to facilitate efficient sharing of vast amount of

resources. One fundamental challenge of P2P networks is to achieve efficient resources

discovery. In the literature, many P2P networks

have been proposed in an attempt to overcome

this challenge. Those networks can be largely

classified into two categories, namely, structured

P2P networks based on a distributed hash table

(DHT) and unstructured P2P networks based on

diverse random search strategies (e.g., flooding) .

Without imposing any stringent constraints over

the network topology, unstructured P2P

networks can be constructed very efficiently and

have therefore attracted far more practical use in

the Internet than the structured networks. Peers

in unstructured networks are often termed blind, since they are usually incapable of determining

the possibility that their neighbor peers can

satisfy any resource queries. An undesirable

consequence of this is that the efficiency of

distributed resource discovery techniques will

have to be compromised.

In practice, resources shared by a peer often

exhibit some statistical patterns. The

fundamental idea of this paper is that the

statistical patterns over locally shared resources

of a peer can be explored to guide the distributed resource discovery process and therefore

enhance the overall resource discovery

performance in unstructured P2P networks.

Three essential research issues have been

identified and studied in this paper in order to

save peers from their blindness [1], [2]. For ease

of discussion, only one important type of

resources, namely, data files will be considered

in this paper. The first research issue questions

the practicality of modeling users’ diverse

interests. To solve this problem, we have

introduced the user interest model (UIM) based on a general

198

probabilistic modeling tool termed Condition

Random Fields (CRFs) . With UIM, we are able

to estimate the probability of any peer sharing a certain resource (file) upon given the fact that it

shares another resource (file) fi. This estimation

further gives rise to an interest distance between

any two peers [3].

The second research issue considers the actual

exploration of users’ interests as embodied by

UIM. To address this concern, a greedy file

search protocol is presented in this paper for fast

resource discovery. Whenever a peer receives a

query for a certain file that is not available

locally, it will forward the query to one of its

neighbors that have the highest probability of actually sharing that file.

The third research issue has been highlighted

with the insight that the search protocol alone is

not sufficient to achieve high resource discovery

performance [5], [6]. This paper proposes a

routing table updating protocol to support our

search protocol through self organizing the

whole P2P network into a small world. Different

from closely related research works that are also

inspired by the small-world model , in order to

reduce the overall communication and processing cost, in this paper, the updating of

routing tables are driven by the queries received

by each peer.

In a P2P network, queries handled by a peer

may be satisfied by any peer in the network with

uneven probability.

This uneven distribution has a significant

impact on our routing table updating protocol

and is demonstrated in this paper. To ensure the

effectiveness of our protocol, a filtering

mechanism is further introduced to mitigate the

impact of uneven updating. To support our discussion in this paper,

theoretical analysis has been performed with two

major outcomes:

1) Using deterministic limiting models (DLMs),

we analytically compared four routing table

updating strategies, with the strong indication

that our protocol is most suitable for managing

neighbor peers.

2) It is shown in this paper that the expected

number of search for any query, provided that

each peer can maintain at least log n routing entries. n stands for the total number of files

shared by the P2P network [7]. Although several

assumptions have been utilized to make our

analysis applicable, these two results clearly

justify the effectiveness of our protocols in

unstructured P2P networks.

Finally, to evaluate all the proposed protocols

and algorithms. We have specifically compared

four protocols for updating routing tables via

experiments. The experiment results suggest that our protocol outperforms the rest three protocols,

including least recently used (LRU) and

enhanced clustering cache replacement (ECCR).

Experiments under varied network conditions are

reported, including different network sizes,

different routing table sizes, and inaccurate

UIMs[4],[9]. The robustness of our protocols is

further evaluated in a dynamic environment,

where peers continuously join and leave the

network. Our experiment results indicate that the

protocols presented in this paper are effective.

2 RELATED WORK

A variety of techniques for locating resources

(files) in P2P networks have been devised over

the last few years. Initial approaches such as

Napster [13], emphasize on a centralized

architecture with designated indexing peers. Due

to the single point of failure problem and the

lack of scalability, recent research in P2P

networks focuses more on distributed search

technologies. Two basic categories of P2P

networks have been proposed to support

distributed search, namely, structured P2P

networks and unstructured P2P networks.

Peers in unstructured P2P networks enjoy more

freedom to choose their neighbors and locally

shared files. In purely unstructured P2P networks

such as Gnutella [12], blind search through

flooding mechanisms is usually explored for

resource discovery. To find a file, a peer sends

out a query to its neighbors on the overlay. These

neighbors, in turn, will forward the query to all

of their neighbors until the query has traveled a certain radius. Despite its simplicity and

robustness, flooding techniques, in general, do

not scale. In large networks, the probability of a

successful search may decrease dramatically

without significantly enlarging the flooding

radius.

Different from semantic-based measurement, this

paper introduces a probability-theoretic

framework for capturing Users’ common

interests. Our UIM does not rely on any

predetermined semantic structure and is flexible to change in response to new evidence regarding

199

the files shared by peers [10]. Moreover, UIM

does not restrict the type of features employed

for describing a file, and it works well with

features that assume symbolic values. Another

prominent strength of UIM is due to its

encapsulation of users’ common interests statistically based on real file sharing patterns.

During the process of searching for a certain file,

all files discovered along the search path can be

evaluated using UIM to identify their degree of

matching the user’s interests.

Those files with a matching degree (or

probability) above a certain level can be returned

to the user as a useful supplement to the file

being searched for Inspired by the seminal

work on small-world networks, methods have

been proposed recently to improve the overall

resource discovery performance by forming a loosely structured P2P network that exhibits the

small-world property (i.e., peers are only a few

steps away from each other). For example, in,

Zhang et al. introduced an ECCR scheme to

obtain small-world networks through self

organization [11]. ECCR requires every peer in

the network to store some files that are not to the

interest of its users. As a result, the locality of

files is essentially destroyed, resulting in the

same problem as in structured P2P networks.

In comparison, we are able to achieve similar search performance, without restricting peers to

store files that they do not want. Both theoretical

and experimental analyses will be performed in

this paper to further compare ECCR and our

protocols.

3 GUIDED SEARCH BASED ON USERS

COMMON INTEREST

3.1 P2P Network Architecture

This paper considers a loosely connected P2P network. We use p to denote a single peer in

the network. P is further utilized to denote the set

of all peers in the network. The main type of

resource, namely, a data file, is represented by f.

For every peer p, Fp is used to represent the

group of files shared by p. In order to conduct

distributed search over the P2P network, every

peer p maintains locally a list of neighbor peers.

3.2 USER INTEREST MODEL

For many Internet applications, users’ diverse

interests often exhibit important statistical

patterns, which can be effectively exploited to

improve the quality of service of these

applications. This section deals with one

essential problem as to how user’s interests can

be modeled properly.

3.3 Learning User Interest Model

To actually apply our model in practice

domain-specific UIMs have to be established and

maintained based on observed file sharing

patterns across the whole P2P network [8]. The

primary concern of this paper is to manage

network topology and to enhance resource

discovery performance with the help of UIM.

However, to make our discussion complete, this

section will briefly introduce the process through

which UIM can be learned and updated

3.4 The Search Protocol

A file search protocol is presented to regulate

the activities of every peer p in a P2P network

upon receiving a query. The local decision

involved in the search protocol demands three

main types of information:

1) The search history hq stored in the query q,

2) The routing table Rp of the peer p that handles

the query q, and

3) The UIM, which are readily available in many

P2P networks.

Fig. 1. The guided search protocol.

4 PROTOCOL FOR UPDATING ROUTING

TABLES

4.1 Routing Table Updating Protocol

The details of our protocol for up dating.

routing table are described in Fig.2

200

Fig. 2. The protocol for updating routing tables.

Whenever the search process driven by any

query q is completed successfully, a new routing

entry, indicating that peer pi shares the queried

file fq, will be temporarily added into the routing

table Rp of every peer p recorded in the search

history hq. If Rp is not full, no entries of Rp will

be removed. Otherwise, the size of Rp will be

reduced to below Br by deleting one or more

selected entries.

5 DESIGN METHODOLOGY

5.1 Network construction

Construct the dynamic network topology. We

can add any number of Peers dynamically and

construct the connections dynamically. Peer can

easily leave from the network, and share the

resources dynamically. Uploading process can

do the dynamic resource sharing.

5.2 Resource Search

Search the need resource by using the resource

name. Peer search the resource using guided

search method. It check the routing table if any

address is available it directly go to that peer else

it search throughout the network. It takes the

resource from the available peer at the same time

it takes address of that particular Peer and other

sharing resource names.

5.3 Routing Table Updating

Routing table contains the information about the

past successful search results. It contains name

of the peer, address, resource name, and count of

the search. Routing table will update every

possible search result. It also has the other

sharing resource names. It could help for future

reference. It leads the guided search.

5.4 UIM

The search based on the user’s common interest.

It will estimate the user’s interest between two

peers. It gives the priority for the user’s interest

resources. It maintains the details about the particular resource in the memory in certain

time. Memory is cleared at the certain time

intervals.

6. CONCLUSION

In this paper, we have shown that the search

performance in unstructured P2P networks can

be effectively improved through exploiting the

statistical patterns over users’ common interests.

Our solution toward enhancing search

performance was presented in three steps:

1. A UIM has been introduced in order to capture

users’ diverse interests within a probability-

theoretic framework. It leads us to further

introduce a concept of interest distance between

any two peers.

2. Guided by UIM, a greedy protocol has been

proposed to drive the distributed search of

queried files through peers’ local interactions.

3. Finally, a routing table updating protocol has

been proposed to manage peers’ neighbor lists.

With the help of a newly introduced filtering

mechanism, the whole P2P network will

gradually self organize into a small world.

Theoretical analysis has also been performed

to facilitate our discussion in this paper.

Specifically, the search protocol was shown to be

quite efficient in small-world networks.

Succeeding analysis further justifies that by

using our routing table updating protocol, the P2P network will self organize into a small

world that guarantees search efficiency.

The search performance in unstructured P2P

networks can be effectively improved through

exploiting the statistical patterns over users’

common interests.

201

To conclude, it is noticed that our solution

presented can be extended from several aspects.

For example, UIM can include extra attributes

that characterize a network user (e.g., gender,

age, and occupation). Meanwhile, live data from

real P2P networks are expected to be collected and applied to further test the effectiveness of

our protocols.

7. REFERENCES

[1] V. Cholvi, P.A. Felber, and E.W. Bier

sack, “Efficient Search in Unstructured Peer-to-

Peer Networks”, European Trans. Telecomm.,

vol. 15, no. 6, 2004.

[2] S. Tewari and L. Kleinrock, “Optimal Search Performance in Unstructured Peer-to-Peer

Networks with Clustered Demands”, IEEE J.

Selected Areas in Comm., vol. 25, no. 1, 2007.

[3] E. Cohen, A. Fiat, and H. Kaplan,

“Associative Search in Peer-to-Peer Networks:

Harnessing Latent semantics”, Proc.IEEE

INFOCOM, 2003.

[4] H. Jin, X. Ning, and H. Chen, “Efficient

Search for Peer-to-Peer Information Retrieval Using Semantic Small World”, Proc. Int’l Conf.

World Wide Web (WWW ’06), pp. 1003-1004,

2006.

[5] T.M. Adami, E. Best, and J.J. Zhu, “Stability

Assessment Using Lyapunov’s First Method”,

Proc. 34th Southeastern Symp. System Theory

(SSST ’02), pp. 297-301, 2002.

[6] S. Androutsellis-Theotokis and D. Spinellis,

“A Survey of Peer-to-Peer Content Distribution

Technologies”, ACM Computing Surveys, vol. 36, no. 4, pp. 335-371, 2004

[7] Cohen, E. and Shenker, S., "Replication

Strategies in Unstructured Peerto- Peer

Networks", in Proc. of ACM SIGCOMM,

August 2002.

[8] Tewari, S., Kleinrock, L. "On Fairness,

Optimal Download Performance and

Proportional Replication in Peer-to-Peer

Networks", in Proc. of IFIP Networking, May 2005.

[9] Sarshar, N., Oscar Boykin., Roy

chowdhury, V. P., "Percolation Search in Power

Law Networks: Making Unstructured Peer-To-

Peer Networks Scalable", in Proc. of IEEE Peer-

to-Peer Computing, September 2003.

[10] B. Y. Zhao, L. Huang, J. Stribling, A. D.

Joseph, and J. D.Kubiatowicz, “Exploiting routing redundancy via structured peer-to-peer

overlays”, in Proceedings of ICNP, Atlanta, GA,

Nov 2003, pp. 246–257.

[11] A. Rowstron and P. Druschel, “Storage management and caching in PAST, a large-scale,

persistent peer-to-peer storage utility”, in

Proceedings of SOSP, Banff, Canada, Oct 2001,

pp. 188–201.

[12]TheGnutellaWebsite,

http://gnutella.wego.com, 2003.

[13]TheNapsterWebsite,

http://www.napster.com/, 2007.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

202

AN ETHICAL HACKING TECHNIQUE FOR PREVENTING

MALICIOUS IMPOSTER EMAIL

R.Rajesh Perumal

Department of Information Technology

Ponjesly College of Engineering

Nagercoil.

Abstract - Active hacking methods are easily

identifiable with the analysis of firewall and

intrusion detection/prevention device (IDS or

IPS) log files. However too Many

organizations & individuals fail to identify the

potential threats from the Malicious Impostor

Emails (the next generation attack), and not

normally identifiable from software or human

beings. In this paper, we see about a next

generation attack, much more powerful than

previous ones, called malicious impostor

emails. A malicious impostor email looks

perfectly legitimate in every way (e.g., it can

pass any statistical filter and human

inspection).There are 3 major defenses, such

as Prevention, Containment and Detection.

We are focusing on prevention. Prevention

can be obtained by Encrypted Address Book

(EAB). EAB can significantly slow down

malicious email spreading. Encrypting those

email addresses may theoretically defeat their

self-spreading.

I.INTRODUCTION

Ethical hacking is essentially the act of

unearthing vulnerabilities in a web based application before going live so that they can be

fixed before being accessed by anyone. People

who do it are IT professionals, not by hackers

with darker intentions. Many companies use

different third party providers for ethical hacking

services. For example, one large bank or large

internet vendor might utilize outside professional

services yearly to test their major applications

yearly, using a different firm each time. The idea

is to get a different perspective, because

methodologies differ from firm to firm, not to mention the different habits of the people

performing the test. R.Rajesh Perumal, Department of

Information Technology, Ponjesly College of

Engineering, Nagercoil.

Email: [email protected]

Active hacking methods are easily identifiable with the analysis of firewall and

intrusion detection/prevention device (IDS or

IPS) log files. However too Many organizations

& individuals fail to identify the potential threats

from the Malicious Impostor Email Problem (the

next generation attack), and not normally

identifiable from software or human beings. In

this paper, we see about a next generation attack,

much more powerful than previous ones, called

malicious impostor emails.

II.SCOPE OF THE PAPER Email has become an indispensable part

of most people’s daily routines, but abuses such

as spam, worms, and viruses have undermined

its utility. In this paper, we focused on a next

generation attack, much more powerful than

previous ones, called malicious impostor emails.

A malicious impostor email looks perfectly

legitimate in every way (e.g., it can pass any

statistical filter and human inspection). It has a

harmful executable as an attachment, which is

dangerous because it appears so authentic that the recipient would have no qualms with opening

its attachment.

There are 3 major defenses, such as

Prevention, Containment and Detection. Always

prevention is better than cure, so in our proposed

approach, we are focusing on prevention. This

prevention can be obtained by Encrypted

Address Book (EAB). EAB can significantly

slow down malicious email spreading. Many

email viruses and malicious impostor emails

proliferate by exploiting the email address books on the infected hosts. Thus, encrypting those

email addresses may theoretically defeat their

self-spreading.

III. PROBLEM DESCRIPTION

SMTP is a “push” protocol, allowing a

client to send an email to its server, which in turn

203

sends it to the recipient’s server. There are some

commands in SMTP to process it. They are

given below

Each received email contains a header

with the email ID, sender, recipient,

time, subject, and path it took from

server to server.

They are easy to create and falsify by

the sender or a relay.

Due to a lack of security safeguards,

SMTP headers are a security

vulnerability that are simple for

adversaries to exploit.

A. Existing Threats to Email: Less Harmful

1) Spam

o unsolicited advertising emails

o commercially-motivated o mass-produced and distributed

2) Phishing

o social engineering attack by

impersonating an authority

o Gleans passwords, account numbers,

cryptographic keys.

B. Existing Threats to Email: More Harmful

1) Email Viruses

o malevolent program sent by email as

an attachment

o inserts itself into other executables and corrupts files

2) Email Worms

o self-replicating, self-propagating

program

o Harms the network and consumes

bandwidth.

Malicious Impostor Email

Problem will do both the less harmful as well as

more harmful threats. Now we can see the

solution for this problem.

C. Malicious Impostor Email Definition-A malicious impostor email is an

email sent to a recipient U with mechanisms

White List U, Filter U, and Scanner U such that

Pr [sender (email) Є White List U] = 1, meaning

that the email possesses a sender address that is

on U’s white list.

Pr [Filter U (email) outputs “suspicious”] = 0,

meaning that the nonattachment content cannot

be detected as malicious, even by a human being.

Pr [Scanner U (email) outputs “suspicious”] =

θ for some 0 ≤ θ < 1.

Consequences of these attacks are severe. They are as follows

The attachment payload could exploit

any system vulnerability.

A PKI could be rendered ineffective.

They may go unnoticed for long periods

of time.

IV.COMMON ATTACK PHASES

Figure 1 The basic phases used to attack are shown in the

figure shown above.

Penetrate perimeter: bypassing the

firewalls and scanning the target

Acquire target : It is referred to the set

of activities undertaken where the tester

subjects the suspect machine to more intrusive challenges such as

vulnerability scans and security

assessment

204

Escalate privileges: once the target has

been acquired, the tester attempts to

exploit the system and gain greater

access to protected resources.

V. OUR PROPOSED APPROACH

We have used four building

blocks to achieve this encrypted address book.

They are as follows

Building-block I: Embedding

passphrases onto pictures

Building-block II: Encryption scheme

and its security requirement

Building-block III: How to encrypt so as to avoid offline dictionary attack

(e.g., @ etc.) Our mechanism for

encrypting email addresses in address

books and folders is called EAB.

• Each address book entry is encrypted with a

unique key.

• The users are relieved of memorizing any

passwords.

• There is no need for any special purpose

hardware.

EAB takes advantage of current hard AI

problems such as and image recognition puzzles.

A. EAB under the Hood

An address book consists of records A

= (A0, A1, A2), where A0 is the

address, A1 is the username, and A2 are

other attributes.

EAB substitutes A0 with several new

attributes, but keeps A1 and A2 intact.

EAB uses symmetric key encryption to

encrypt each A0 (RC4 stream cipher in

practice) with a unique passphrase as

the key.

B.. Securely Encrypting an Email Address

Email addresses are well-formed and

follow a specified format.

EAB encrypts each address, although

top-level domains, the first character,

and “@” and “.” characters are not

encrypted.

EAB maps valid email characters (A-Z,

a-z, 0-9, “-” and “_”) to the first six bits

per character, with the remaining set to

“0”.

C. An Entry in an Address Book

An icon is associated with a contact,

with the passphrase (RC4’s key)

embedded in the icon.

Correctly typing the passphrase

decrypts the email address.

To avoid typos, a second encrypted image may be used, with the passphrase

as its key. If the image is consistent

with the icon the passphrase was

correctly typed

D. Functionalities of the Prototype System

eab.setup(ces.ab)

eabØ

for each(a0,a1,a2) Є ces.ab

Select a random string r0

user picks a passphrase pa

user picks a picture image

icon embed(pa,image)

c0 Ench (r0 , pa) ( a0 )

eab eab U (icon, r0 ,c0 , a1,a2)

eab.modifyaddr(eab)

User clicks an icon Є eab

User enters passphrase pa from icon

User enters new address newaddr

c0 Ench (r0 , pa) ( newaddr )

update(icon, r0 ,c0 , a1,a2 ) in eab

eab.deleteaddr(eab,icon)

If(icon, r0 ,c0 , a1,a2 ) Є eab

eab eab (icon, r0 ,c0 , a1,a2 )

Initialize(ces.ab,ces.folders)

eab.setup(ces.ab)

For each folder Є ces.folders

For each email Є folders

For each u Є email

u.addr eab.geticon(eab,u.name)

205

erase ces.ab

receive(eab)

EMAILS ces.receive() Table projection (eab,icon,c1)

For each em Є EMAILS

For each u Є EMAILS

If (icon,u,name) Є Table

u.addr icon

Else

User picks an icon for u.name

eab.addaddr(eab,u.addr,u.name)

u.addr eab.geticon(eab,u.name)

eab.addaddr(eab,addr,user)

Select a random string r0

user picks a passphrase pa

user picks a picture image

icon embed(pa,image)

c0 Ench (r0 , pa) ( addr )

c1 user c2 info

eab eab U (icon, r0 ,c0 , c1,c2 )

eab.getaddr(eab,icon)

If(icon, r0 ,c0 , c1,c2 ) Є eab

User enters pa from icon

return Dech (r0 , pa) (c0)

Else return NULL

eab.geticon(eab,user)

If(icon, r0 ,c0 , user,c2 ) Є eab

return icon

Else return NULL

deleteaddr(eab,new.folder, icon)

For each folder Є new.folders

For each email Є folder

If icon = email.addr

eab.deleteaddr(eab,icon)

return

insertaddr(eab,email)

If user clicks “insert address” button

User clicks on an icon Є eab

addr eab.getaddr(eab,icon)

ces.insertaddr(addr)

If user types in an email address addr User enters a username user

If eab.geticon(eab,user) = NULL

eab.addaddr(eab,addr,user)

E. Security Assumptions

Image recognition is a difficult

problem. Specifically, given two images, do they correspond to the same

person or not?

CAPTCHAs are hard for programs to

decipher; conversely, CAPTCHAs are

easy for a human to decipher.

Users will type low-entropy

passphrases.

VI. OUR EXPERIMENTS

Example Email Address Book Entry is shown in

the diagram below

Example Email in the Inbox shown in the

206

diagram below

A. Verifying the CAPTCHA

These are the captcha verifying screenshots.

B. Related Prior Works

Actually some commercial software,

which however “encrypts all addresses

with a single password”

But can be integrated with our so that we get “two layers of encryption”

(EAB for the email addresses you care

most)

VII. Conclusions

We systematically explored the feasibility of

encrypting email address books so that malicious

impostor emails and email viruses cannot

automatically spread themselves within a short

period of time.

Beyond: Email folders are also protected

VIII. References

1) How to Secure Your Email Address Books -

Erhan J Kartaltepe, Paul Parker, and Shouhuai

Xu Department of Computer Science

University of Texas at San Antonio

2) SMTP Information gathering –Liuis

Mora,Neutralbit

3) Web Application Hacking -Matthew Fisher,

SPI Dynamics

4) Gray Hat Hacking The Ethical Hacker’s Handbook -Shon Harris, Allen Harper, Chris

Eagle, Jonathan Ness, and Michael Lester

5) Passive Information Gathering The Analysis of Leaked Network Security Information -Gunter

Ollmann, Professional Services Director

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

207

STERLING NETWORK SECURITY

– (New dimension in computer Security)

K.YAMINI (3rd

year )

Department of computer science and

engineering

Anand institute of higher technology

[email protected]

ABSTRACT:

Network security is a complicated

subject, historically only tackled by well-

trained and experienced experts.

However, as more and more people

become ``wired'', the number of security

threats have increased exponentially in

recent times. In spite of so many security

services provided, still the hackers find it

possible to break the code.

The security function of the

system is divided into two equal partitions

and rests equally on both firewalls and

VPN. Virtual Private Network is a

method of providing secure data and

voice communication across a public

network, usually the Internet.

Virtual Private Network should basically

possess

Encryption

Authentication

Access control

M.RANJITH (3rd

year )

Department of computer science and

engineering

Anand institute of higher technology

[email protected]

Virtual Private Network security is

established by tunneling.

Tunneling would establish a set of

protocols legs which complaint Internet

components could create their own

channel inside the Internet. This channel

would be protected by authentication and

encryption counter-measures

Here is our effort to provide a stern

resistance to the hackers by combining

already existing features like firewall and

VPN. The two security measures though

proved already, do contain certain flaws

of their own. Here we have tried to

combine the merits of the two systems

and exclude the demerits.

1. INTRODUCTION:

People choose a variety of security

models, the commonly referred model is

208

“Security Through Obscurity” this

model, a system is presumed to be

secure simply because nobody knows

about it-its existence, measure or

anything else.

Probably the most for computer security

is “Host Security”

A firewall is a system or a group of

systems that enforces an access control

policy between two networks. It is

basically a protective device. Firewall is

a set of hardware components –a router,

a host computer or some combination of

routers, computers and networks with

appropriate software. Packet filtering

proves to provide insufficient security.

And also the advanced types of firewalls

have some security holes for hackers.

Therefore using encryption,

authentication and stegnography along

with Firewalls can strengthen security.

1.1 EFFECTIVE BORDER

SECURITY:

To maintain the absolute minimum level

of effective Internet security, we must

control our border security using

firewalls that perform all the three of the

basic firewall functions.

Packet filtering

Network address translation

High level service proxy

1.2 COMMENTS:

Many services contain logon banners or

automatically generated error pages that

identify the firewall product we are

using. This could be dangerous if

hackers have found a weakness in your

specific firewall.

1.3 PACKET FILTERING:

KEY: Rejects TCP/IP packets from

unauthorized access hosts and rejects

connection attempts to unauthorized

services.

The main feature of packet filtering is

that it rejects the unauthorized access to

the TCP/IP packets. Filtering operations

are based on set of rules encoded in the

software running in the firewall. The

most common type of packet filtering is

done on IP datagrams. Datagrams are

209

the packets that are transferred through

the datalink layer.

1.4 NETWORK ADDRESS

TRANSLATION:

It translates the IP address of the internal

hosts to hide them from outside

monitoring. Network Address

Translation is also called as IP

masquerading.

1.5 PROXY SERVICES:

Makes high-level application

connections on behalf of internal hosts to

completely break the network layer

connection between internal and external

hosts. The proxy services provide or

establish a connection between the high-

level applications on behalf of the

internal hosts to break the connection

between the internal and external hosts

in the network layer completely.

ADVANTAGES OF USING

FIREWALLS:

The firewall plays a very important role

in the case of security and secures the

system .The basic advantages of the

firewall are listed below.

1. A firewall is a focus for security

decisions

2. A firewall can enforce a security

policy

3. A firewall can log Internet

activity efficiently

4. A firewall limits your exposure

5. A firewall can’t fully protect

against viruses out of a network.

PROBLEMS FIREWALLS

CAN’T SOLVE:

The problem with all proprietary

firewalls is the lack of support for third-

party security software like content

filters and virus scanners.

2. Virtual Private Networks:

With the Point-to-Point Tunneling

Protocol (PPTP) or Layer Two

Tunneling Protocol (L2TP), which are

automatically installed on your

computer, you can securely access

resources on a network by connecting to

a remote access server through the

Internet or other network. The use of

both private and public networks to

create a network connection is called a

virtual private network. Now, after these

initial security works performed by

firewall, VPN performs equally

important functions

2.1 VPN:-

210

Virtual Private Network is a method of

providing secure data and voice

communication across a public network,

usually the Internet.

Virtual Private Network should basically

possess

Encryption

Authentication

Access control

Virtual Private Network security is

established by tunneling.

Tunneling would establish a set of

protocols legs which complaint Internet

components could create their own

channel inside the Internet. This channel

would be protected by authentication

and encryption counter-measures.

2.2 ENCRYPTED

AUTHENTICATION:

Allows users on the public network to

prove their identity to the firewall in

order to gain access to the private

network from external locations.

2.3 ENCRYPTED TUNNELS:

It is used to establish a secure

connection between two private

networks over a public medium like the

Internet. This method allows physically

separated networks to use to the Internet

rather than a leased-line connection to

communicate. Tunneling is also called as

VPN.

VPN (CONNECTION

OVERVIEW)

Advantages of VPN:-

1. Cost advantages:

The Internet is used as a connection

instead of a long distance telephone

number or 1-800 service. Because an

ISP maintains communications hardware

such as modems and ISDN adapters,

your network requires less hardware to

purchase and manage.

2. Outsourcing dial-up networks

You can make a local call to the

Telephone Company or Internet service

provider (ISP), which then connects you

to a remote access server and your

corporate network. It is the telephone

company or ISP that manages the

modems and telephone lines required for

dial-up access.

3. Enhanced security

The connection over the Internet is

encrypted and secure. New

authentication and encryption protocols

are enforced by the remote access server.

211

Sensitive data is hidden from Internet

users, but made securely accessible to

appropriate users through a VPN

5. IP address security

Because the VPN is encrypted, the

addresses you specify are protected, and

the Internet only sees the external IP

address. For organizations with

nonconforming internal IP addresses, the

repercussions of this are substantial, as

no administrative costs are associated

with having to change IP addresses for

remote access via the Internet.

3. INTEGRATING FIREWALL

& VPN:-

(OUR PROPOSED MODEL)

Here is the list of actions occurring, once

VPN and firewall have been integrated.

SECURITY SERVICES:

Firewall use codes and ciphers for two

virtually important purposes.

3.1 Authentication

3.2 Encryption

3.1. AUTHENTICATION:

To provide the identity of the user.

3.1.1. HTTP

AUTHENTICATIONTH

THROUGH A FIREWALL:

The HTTP authentication is done by the

following 4 steps:

Step 1: The user from the Internet

initiates an HTTP request to a specified

Corporate web server.

Step 2: The firewall intercepts the

connection and initiates the

authentication process.

Step 3: If the user authenticates

successfully the firewall completes the

HTTP connection to the corporate web

server.

Step 4: Firewall forwards requests and

response without further intervention.

3.2. ENCRYPTION:

Encryption is a technique by which the

text in the readable format is converted

into an unreadable format. This process

involves with the three things. That is

the input which is nothing but the

plaintext and the key is used to perform

some calculations on the plaintext then

the cipher text is the output which is

obtained by the operation performed on

the plaintext by using the key. To hide

the contents of the data stream.

Encryption is performed using

cryptography.

3.3 FIREWALLS to be used:

The various firewalls which can be used

in our work are as follows

212

WINDOWS NT:

This is an excellent firewall because it is

configured on existing Operating

Systems. Thus it acts as a double-edged

sword.

ALTAVISTA FIREWALL98:

AltaVista firewall is a high-end security

proxy. Their software runs on both

Windows NT and UNIX. The Alta Vista

provides a similar interface to

administrators of either system of

Windows NT and UNIX.

DEDICATED FIREWALLS:

Dedicated firewalls include built in

operating system and firewall software.

These devices consist of have network

interfaces, plenty of RAM and a fast

microprocessor.

With this firewall, we don’t need a

firewall to run on a same operating

system as we use for file on application

services.

3.4 CONNECTING THE VPN:-

There are two ways to create a

connection: By dialing an ISP, or by

connecting directly to the Internet, as

shown in the following examples.

1. In the first method, the VPN

connection first makes a call to an ISP.

After the connection is established, the

connection then makes another call to

the remote access server that establishes

the PPTP or L2TP tunnel. After

authentication, you can access the

corporate network, as shown in the

following illustration.

2. In the second method, a user who is

already connected to the Internet uses a

VPN connection to dial the number for

the remote access server. Examples of

this type of user include a person whose

computer is connected to a local area

network, a cable modem user, or a

subscriber of a service such as ADSL,

where IP connectivity is established

immediately after the user's computer is

turned on.

213

Connecting directly to the

Internet means direct IP access

without going through an ISP.

(For example, some hotels allow

you to use an Ethernet cable to

connect to the Internet.)

If you have an active Winsock

Proxy client, you cannot create a

VPN. A Winsock Proxy client

immediately redirects data to a

configured proxy server before

the data can be processed in the

fashion required by a VPN. To

establish a VPN, you should

disable the Winsock Proxy client.

3.5 INTEGRATION OF VPN

AND FIREWALLS:

Below is the diagrammatic

representation of our proposed model.

Now the question arises as to how the

functions of firewalls and VPN can be

integrated,

Firewall has its own flaws and

VPN has its own drawbacks.

By their integration of VPN and

firewalls all flaws can be

truncated.

VPN fails to provide security if

the operating system of the

machine is not secure. Thus,

PN is mainly based on the O.S.

Thus it is clear that platform for

VPN to perform well is the

operating system.

The firewalls protect the base

operating system from attack.

Thus VPN and firewall

combination is the best security

solution for all intrusions

available.

3.6 FUNCTIONING OF OUR

MODEL:

214

Pre-filtering

It is a typical firewall role.

Allows or denies packets based

on simplistic access control and

packet filtering.

EXAMPLES: Allow IPSec traffic

to be passed to IPSec engine.

IPSec processing

Typical VPN role

It performs NAT traversal.

It drops the packets that do not

match a particular security

policy.

EXAMPLES: Disallow telnet

connections that do not use ESP.

Post-filtering

It is also a firewall role.

Matches the connection state for

more granular filtering.

EXAMPLES: Perform additional

firewall rules on data exiting the IPSec

process.

Packet Processing in Stages

1. Firewall rule sets are applied on

packets exiting tunnel

2. Allows for more refined filtering.

1. Apply rules to packets entering the

Firewall

2. Pass this state onto next stage for

IPSec and L2TP processing

3. Apply firewall action to packets

exiting the IPSec processing stage

What have we achieved?

As a result of combining the VPN and

firewall features, we have achieved a

215

dual protection feature in our new

security system.

1. The OS, which is the base for the

performance of VPN and the protection

for the OS is brought about by firewall,

thus mutually benefiting from each of

their features.

2. We have utilized the basic resources

available for us and have come up with a

trustable yet simple feature. The basic

flaws of the proposed systems have been

discarded by just combining them and

with no other major change.

3. A firewall is different from VPN, but

the two of them work together to help

protect your computer. You might say

that a firewall guards the windows and

doors against strangers or unwanted

programs trying to get in, while VPN

protects against viruses or other security

threats that can try to sneak in through

the front door.

MERITS:-

Most simple method.

Based completely on the

available resources (the system is

modularized).

IPSec security parameters can be

used as filtering criteria.

The features of the new system

are based on proved concepts and

hence, probability of flaws is

minimal.

DEMERITS:-

In spite of all the security

features enhanced, still virus

protection isn’t achieved entirely

and we may have to seek a virus

protection system to achieve this.

Combination firewall/VPN

appliances are not the best

solution in some environments,

such as in companies that

generate revenue from high-

traffic Web sites and rely heavily

on VPN for branch office and

remote access connectivity

because VPN eats a lot of

processing power, potentially

slowing down Web-traffic

response-time.

FUTURE ENHANCEMENTS:-

(What we have to achieve in the

upcoming years)

All the firewalls don’t offer very good

virus protection. Thus Virtual Private

Networks and firewall combination is

the best security solution for all

intrusions available but not for viruses.

The best way to provide solutions for all

the security problems related to the

viruses is given below.

216

Security logs

System logs

Traffic logs

Packet logs

The most practical way to address the

virus problem is through host based

virus protection software. Hence the

combination of Virtual Private Networks

and firewalls with some virus protection

software embedded in it will provide the

best secure network. Most anti-virus

software offer options such as automatic

cleaning, prompt for response, delete if

the virus cannot be repaired. If this is

configured it works in background

without disturbing.

CONCLUSION:

No VPN solution provides effective

security if the operating system of the

machine is not secure. The firewall will

protect the base operating system from

attack. But our proposed system will

truncate the flaws of both VPN and

firewalls. Thus the security can be

strengthened by the combination of VPN

and firewall hence become truculent in

terminating any type of intrusion.

Though the combination will work out

effectively, it will be inefficient in

preventing the intrusion of few viruses.

Thus virus protection software can be

built to enhance the security and build

perfect security.

BIBLIOGRAPHY

E-security and You-Sundeep Oberoi.

Firewalls a complete Guide-

Goncalves,Marcus

Building Internet firewalls-

Elizabeth.d.Zwick

Firewalls-Strebe,Mathe

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

217

Authors:

R.Divya Priya ([email protected])

R.Lavanya ([email protected])

Thiagarajar College of Engineering

Madurai-15.

“AUTOMATIONOF IRRIGATION

USING WIRELESS”

Abstract:

The aim of the present paper is to

review the technical and scientific state

of the art of wireless sensor technologies

and standards for wireless

communications in the Agric-Food

sector. These technologies are very

promising in several fields such as

environmental monitoring, precision

agriculture, cold chain control or

traceability. . The paper focuses on WSN

(Wireless Sensor Networks) and data

acquisition system Presenting the

different systems available, recent

developments and examples of

applications of cultivation.

Today in India inclination towards

automization of irrigation is gaining

momentum due to :

• Automation eliminates manual

operation to open or close valves,

especially in intensive Irrigation process.

• Possibility to change frequency of

irrigation and festination process and

also to optimize these processes.

• Adoption of advanced crop systems

and new technologies, especially new

crops system. That are complex and

difficult to operate manually.

• Use of water from different sources

and increased water and fertilizer use

efficiency.

• System can be operated at night, thus

the day time can be utilized for other

agricultural activities.

• Pump starts and stops exactly when

required, thus optimizing energy

requirements.

Automation:

Automation of irrigation system

refers to operation of the system with no

or minimum manual interventions.

Irrigation automation is well justified

where a large area to be irrigated is

218

divided into small segments called

irrigation blocks and segments are

irrigated in sequence to match the flow

or water available from the water source.

SYSTEMS OF AUTOMATION:

1. Closed Loop Systems:

This type of system requires feedback

from one or more sensors. The

operator develops a general control

strategy. Once the general strategy is

defined, the control system takes over

and makes detailed decisions of when to

apply water and how much water to

apply. Irrigation decisions are made and

actions are carried out based on data

from sensors. In this type of system, the

feedback and control of the system are

done continuously.Closed loop

controllers require data acquisition of

environmental parameters (such as soil

moisture, temperature, radiation, wind-

speed, etc) as well as system parameters

(pressure,

flow, etc.).

2. Real Time Feedback System

Real time feedback is the application if

irrigation based on actual dynamic

demand of the plant itself, plant root

zone effectively reflecting all

environmental factors acting upon the

plant. Operating within controlled

parameters, the plant itself determines

the degree of irrigation required. Various

sensors viz., tensiometers, relative

humidity sensors, rain sensors,

temperature sensors etc control the

irrigation scheduling. These sensors

provide feedback to the controller to

control its operation.

Automatic Systems:

In fully automated systems the human

factor is eliminated and replaced by a

computer specifically programmed to

react appropriately to any changes in the

parameters monitored

by sensors. The automatic functions are

activated by feedback from field units

219

and corrections in the flow parameters

by control of devices in the irrigation

system until the desired performance

level is attained. Automatic systems can

also perform auxiliary functions such as

stopping irrigation in case of rain,

injecting acid to control pH, sounding

alarms, etc. Most control systems

include protection in emergencies such

as loss of the handled liquid due to pipe

burst. They close the main valve of the

whole system or of a branching, when an

unusually high flow rate or an unusual

pressure drop is reported by the

sensors.

System Component Of An Automatic

Irrigation System

Controllers:

This device is the heart of the

automation, which coordinates

operations of the entire system. The

controller is programmed to run various

zones of an area for their required

duration or volume. In some cases

sensors are used to provide feedback to

the controller. In the simplest form,

irrigation controllers are devices which

combine an electronic calendar and

clock and are housed in suitable

enclosure for protection from the

elements.The PLC’s, microprocessors

and computers are now available and

being used extensively.

(PLC-Programmable Logic Controllers).

Electromechanical Controllers:

These types of controllers are

generally very reliable and not very

sensitive to the quality of the power

available. However, because of the

mechanically-based components, they

are limited in the features they provide.

The picture shows the controller for

opening and closing of irrigating pipes.

Electronic Controllers:

These types of systems are more

sensitive to power-line quality than

electromechanical

220

controllers, and may be affected by

spikes, surges and brownouts. These

type of systems may require electrical

suppression devices in order to operate

reliably. Since electronic based these

provide a large number of features at a

relatively low cost.

Sensors:

A sensor is a device placed in the

system that produces an electrical signal

directly related to the parameter that is to

be measured. Sensors are an extremely

important component of the control loop

because they provide the basic data that

drive an automatic control system.

In general, there are two types of

sensors: continuous and discrete.

Continuous sensors produce a

continuous electrical signal, such as a

voltage, current, conductivity,

capacitance, or any other measurable

electrical property. Continuous sensors

are used when just knowing the on/off

state of a sensor is not sufficient. For

example, to measure pressure drop

across a filter or determine tension in the

soil with a tensiometer fitted with a

pressure transducer requires continuous-

type sensors. Discrete sensors are

basically switches (mechanical or

electronic) that indicate whether an on or

off condition exists. Discrete sensors are

useful for indicating thresholds, such as

the opening and closure of devices such

as valves, alarms, etc.

Various types of soil moisture sensors,

weather instrumentation, plant-water

stress or crop canopy temperature are

available and can be used in feedback

mode for irrigation

management.

The picture shows the sensor in the

fields.

221

222

This flowchart explains about the

requirements needed for the irrigation

and the abnormal conditions of the

climatic environment such as drought,

climatic changes, seasonal rainfall,

and many more.

The chart is explained as follows:

First and foremost step is the

analysis of soil condition by

sensors such as tensiometer

to measure the water

percolation, moisture in soil.

Then, the crop requirements

are taken into account and the

analysis is done for the crop.

If suppose the crop is paddy,

we need an adequate supply

of water.

After the analysis, the

climatic condition is taken

into account such as monsoon

season, summer and winter.

The data acquired from the

above analysis are processed

and finally using the wireless

network the operation is

done.

Automatic Metering Valve:

These valves are required only in

volume based irrigation system. The

volume of water

required for the irrigation can be

adjusted in these automatic metering

valves. These valves

can be simple metering valve which

shuts off after delivering preset quantity

of water or

automatic metering valve with pulse

output which provides pulses to the

controller to count

the volume of water.

Valves:

Automated valves are activated either

electrically, hydraulically or

pneumatically and used to switch water

on or off, flush filters, mains and

laterals, sequence water from one field

or segment to other.

Metering pumps:

These pumps are suitable for feeding of

known quantity of fertilizers/chemicals.

223

The capacity of pumps varies from 1.5 to

3.5 litres / hour

APPLICATION OF WIRELESS:

Using wireless sensor networks

within the agricultural industry is

increasingly common. Gravity fed water

systems can be monitored using pressure

transmitters to monitor water tank levels,

pumps can be controlled using wireless

I/O devices, and water use can be

measured and wirelessly transmitted

back to a central control center for

billing. Irrigation automation enables

more efficient water use and reduces

waste.

The above picture shows the root and

nodes of the wireless network.

Wireless sensor networks (WSN)

technologies are increasingly being

implemented for modern precision

agriculture monitoring. The advantages

of WSN in agriculture are for several

reasons; these includes suitability for

distributed data collecting and

monitoring in tough environments,

capable to control an economical way of

climate, irrigation and

nutrient supply to produce the best crop

condition, increase the production

efficiency while decreasing cost and

provide the real time information of the

fields that enable the farmers to adjust

strategies at any time. This paper

presents the preliminary design on the

development of WSN for paddy rice

cropping monitoring application. The

propose WSN system will be able to

communicates each other with lower

power consumption in order to deliver

their real data collection. The main

objective of the new design architecture

is to cater the most important and critical

issue in WSN, that is power

consumption. The design will attempt a

sensor node board systems consist of a

low power microcontroller so called

nano-Watt technology, low-power

semiconductor based sensors, Zigbee™

224

IEEE 802.15.4 wireless transceiver and

solar energy source with optimal power

management

system. The all information’s are

processed using data acquisition system.

Conclusion:

Environmental monitoring in agriculture

is essential for controlling in economical

way of climate, irrigation and nutrient

supply to produce the best crop

condition, increase the production

efficiency while decreasing cost . The

real time information of the fields will

provide a solid base for farmers to adjust

strategies at any time. WSN will

revolutionize the data collection in

agricultural research. However, there

have been few researches on the

applications of WSN for agriculture.

“ Innovation is the mothering for the

Future Technology”

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

225

Key management and distribution for

Authenticating group communication

Krisshnakumar K.S., Department of computer science and engineering,

Jaisakthiraman M., Department of computer science and engineering,

Velammal college of engineering and technology, Madurai-9.

Abstract – In network communication,

security is important to prevent intruder’s

attack, where authentication plays a very

important role in ensuring the

communication with the right participant.

To authenticate a group of users and to

hold confidentiality between them a new

approach of key management is handled in

this paper. This key management is a

group key management used for

cryptography process that provides security

with less computation complexity, which

leads to fast and easy authentication. Here

a tree based key management scheme has

been designed through which key

distribution among the user involves

authentication with less storage complexity

too. here dynamic membership

authentication is described which talks

about communication or transaction

among user and server with a new different

instant key that is formed from a group of

members who are currently participating in

transaction or communication. Releasing

or joining a user from or to the group is

possible here by which the keys get

changed instantly whenever there is an

alteration in the current participation in

transaction or communication. With the

help of users hidden key an ‘Instant key

can be computed As the entire users’

hidden keys is used for the generation of

the instant key we can ensure the

communication between a valid set of

authenticated users. So an intruder cannot

interrupt any communication between the

users as their keys get vied often. Thus we

can provide a very reliable system for the

communication of the people. Keywords: - Authentication, Instant key,

Hidden key, Dynamic membership,

Group key.

I Introduction To strengthen the communication security,

authentication is preferred to authenticate the

users to ensure their validly and hold their

confidentiality. This is based on cryptography

which has many advantages over the other

schemes [4].This paper describes a group key

management that is used for cryptography

process that provides authentication for

providing identification. Here the group key is an instantly generated

key, which is as a function of all the current

group members, termed as an instant key.

This is for group communication over

particular servers‟ access. It is formed by a

virtual tree based approach where each leaf

node denotes a user (who wishes to access the

same server) and each intermediate node

value is evaluated from its subordinate node.

Each user holds a pair of keys namely „short-

term secret key‟ and „hidden key‟.

226

The „instant key is generated from the

collection of the entire user‟s value that

remains valid from the time when the user

joins and until he leaves the tree structure. In this group key management scheme,

releasing a user from the group is possible

which is described in section III.B. Similarly

a member can also join the existing group

which is described in III.C. In this situation

the key gets changed instantly whenever there

is an alteration in current participation of

users for communication. Only after receiving the group key, a user

can further communicate with other

servers. It‟s explained in the section III of

this paper. By ensuring the correctness of

the user, the server will abduct all its

communication information by

encrypting using the group key. Hence only valid users are supported by the

server for the further communication. This method involves less storage space

complexity and less computation

complexity that leads to fast up the process.

II Subset Tree Structure There is some tree structure based scheme

that is used for group communication [2]

[3] [5] [7] as shown in figure: 1. here all the

nodes will be allotted with a value. The leaf

nodes are members M1, M2……Mn. A

member‟s value is obtained from the root

node to its member leaf node. By evaluating

the members in the root nodes, group key is

formed for communication. Then the key is

broadcasted across the tree structure to all

its nodes. In a proposed scheme called Group key

Management Protocol (GKMP) [5] [7], the

entire group shares the same key called as

session-encrypting key (SEK). In some of

the tree-based key distribution scheme [6]

[7], the root value is used as the session key

as well as group key. In these schemes, to

our knowledge, the storage detail about the

user occupies more space [8]. In some techniques [5] [7], it is stated that

at the depth two or more from the root, the

key values are duplicated across the tree

twice. For example, L1 and L2 in figure: 1.

Also it is stated that the deletion of more

than one member may bring the scheme to

halt. This paper proposes the scheme

through which it highlights the

enhancement of this tree based group

communication for authenticating user with

less storage and computation complexity.

III Proposed Tree Based Group Key Generation In this Group key generation each user is

denoted as a leaf node .Each key „u‟

represents secret key „Ku‟(i.e. own key

and a hidden key „Hku‟ as shown in the

figure 2 that remain valid from the time

user joins and until he leaves.

227

A Hidden Key Generation Hidden Key can calculated as Every intermediate node holds a secret

key obtained from its leaf user (parent‟s

value is evaluated from the user). The secret key of the intermediate node

can be generated as

With the help of each users hidden key,

group key can be computed .for example

as in figure: 2 from the user „M1‟ a group

key can be generated.

Using K7 and HK8 the node „3‟

can calculate its‟ key value(i.e. K3)

Then using K3 and HK4, the key

value of K1 can be obtained. Similarly K1 and Hk2 provide a

computed value K0, which is the

group key. A binary tree with „n‟ leaves will have

(2n-1) nodes. i.e. (2n-1) HKs are

considered for the evaluation. From the authorized users „Mi‟ those who

require service will be collected and

formed as a tree structure. Then the key

calculation takes place to compute the

group key, say „K‟. If a new user joins the

group dynamically the key gets changed

as K‟. If a user leaves the group then

instantly it becomes K‟‟.

B Member joins Consider a binary tree that has a valid

members M1, M2, Mn.Among these n

valid users only those who request for a

service on a particular server will be

formed as an individual group. A new

member Mn+1 indicate the service request

message that contains its own HKey

(hidden key) i.e.HK [Mn+1].This message is

distinct from any join messages generated

by the underlying communication system.

The authentication server first determines

the insertion point in the tree. The insertion

point is the shallowest rightmost node,

where the join doesn‟t increase the height

of the key tree. Otherwise if the key tree is

fully balanced, the new member joins to the

root node.

228

Suppose a user M8 wished to access a

particular server the newly requested

user M8 has to be added to the existing

list of users as shown in the fig3. Here, the

insertion point in the tree is node5 where

M4 exists.

Now a binary node is

created from node5, in turn it becomes

the intermediate node where M4 is shifted

to its left most node (seen in the above

figure) and the newly requested user is

joined in the rightmost node i.e. M8.The

server proceeds to update its share and

compute the group key; it can do this

since it‟s know all the necessary HKey.

While addition of a new user the

server has to perform n+ 1 computation [7].

Infact in obtaining the HK[8] the server

starts computations and then it rekeys the

nodes‟5‟,‟2‟ and „0‟ and then process further .All other members update their new

instantly generated group key.

C Member Leaves Once again, we start with n members and

assume that member Mn leaves the group.

The sponsor in this case is the server which

instantly updates the generated keys.

Suppose a user wants to leave from this

group all the keys from the leaf node to the

root most must be updated in order to

prevent the access to the future data. If member M7 wants to departure from

the service, then intermediate node 6 will

have a change and shuffle with 13 and 14

as shown in the figure 5. The server has to perform (n-1)

computations under deletion of an existing

user [7]. Now the key S of the intermediate

nodes „6‟,‟2‟ and „0‟ must be renewed as

shown in the figure 6 after the release of

M7 from the current group. Hence the key

gets dynamically convert, due to the

changes in the participation of user access,

with less computation. The server uses

instant key whenever there is an alteration

in the group key communication.

229

IV Instant Key Communication

The dynamically generated instant

key from the group of valid users is used to

encrypt the authentication details for

further communications in figure 7.

Consider a user „U‟ with identity

‟IDu‟ and an authenticating server „AS‟

with the identity „IDas‟ shares a secret

codeword „cw‟.consider two large prime

numbers p and q, where p=2q+1, and

perform a secure one way hash function „h‟ In the proposed scheme following

procedure takes place:

a. Initial phase

Initially the user requests the AS by

sending its H value.

After receiving H, AS checks the „cw‟.

Then it generates a random number „R1‟

sends to U where („i‟

already obtained from U)

b. Middle phase U checks the received message

whether it equals the sent content

which is i. Then U sends its HK along

with R1 to AS encrypted as Kas. After receiving the HK of every valid

user, AS will form the tree structure

for grouping the key (GK) as per

request for „Srv‟. Also it creates identification content

„id-msg‟ comprises of the following

Where „Ku,srv‟ denotes the key used

for cryptography, process of user and

data server communication and

„Kas,srv‟ denotes the key for

communicating AS and server.

c. Main phase AS constructs and sends the

confirmation message to individual U

by encrypting through the newly

generated group key.

230

„TS‟ denotes”Time Sampling” at

which the key was issued. Here the entire content was encrypted

using GK, which can be decrypted

only by the authorized users. The

detail content can only decrypt by the

respective user because a distinct key

(Kuj, AS) for each user encrypts it. Now, the user passes the „id-msg‟ to the

server. It is obtained from AS that cannot

be revealed even by the user. The server

decrypts its (Server-User) Key using [id-msg].

V Discussion The discussions and comparisons

emphasize the benefits of the proposed

group key scheme as given below: In the group key management protocol

(GKMP)[5][7],the entire group shares

the same key called session-encrypting

key (SEK), where as in the proposed

Instant key method, the instant key is

common to all users for initial stage but

a distinct key for each user is provided. In some of the tree-based key

distribution scheme [6] [7], the root

value is used as the session key as well

as the group key. But here the

computation of the hidden key forms

the group key, through which the

session key for individual user is sent. The instant key gets varied all the

time when a node arrives or leaves the

group. It keeps on changing the key so

that an intruder cannot enter into the

communication of the group. So it

leads to high security.

As the instant key is new for all changes

in the tree path, impersonation of the

authorized user is avoided so it confirms

only the valid user‟s transmission. As the newly generated key is obtained

from current user‟s hidden key, the

valid currently operating users only can

access it. Hence authentication is

successfully ensured. It leads to assure

secure communication. Whenever a node joins/leaves, the server

has to perform (n+1)/ (n-1) computations

respectively [7].Also in some tree-based

communications, all other members

update their tree accordingly and

compute the new group key. It n=might

appear wasteful to broadcast the entire

tree to all members [2] [3] [6]. But in this

proposed scheme, only a particular part

of the tree is necessary for computation

whenever the tree structure is altered.

Consequently it minimizes the storage

requirements for the user information on

server. In some authentication scheme (like

certification scheme), verifiers spend

more time to extract a certificate path

from network of certificates and do

repeated process to verify and

authenticate a user. In this scheme

verification time is very less. As a unique session key is provided to

the individual user, impersonation is

also not possible. In certain tree-based group key scheme,

the intermediated nodes are duplicated

across tree several times (at least

twice)[7].where as in this scheme, even

while generating the instant key, the

required path alone chosen without

disturbing the other intermediate nodes

so the duplication is avoided.

231

VI Conclusion The proposed scheme is a tree based key

management where an instantly generated

key is computed as a function of all current

group members present in communication.

Thus only valid users can communicate with

each other and the presence of users is

ensured without much difficulty.

Authentication is successfully processed

during communication. The time and storage

complexity is reduced. Hence the proposed

tree based key management scheme ensures

the right participant in communication with

less time and authenticates the users to

maintain confidentiality. REFERENCES

[1] David A. McGrew and Alan T.

Sherman, “Key Establishment in

Large Dynamic Groups Using

One- Way Function Trees"

Cryptographic Technologies

Group, TIS Labs at Network

Associates, May 20, 1998.

[2] Yongdae, Perrig and Tsudik,

"Group Key Agreement efficient in

communication", IEEE transactions

on computers, Oct 2003.

[3] Sang won Leel, Yongdae Kim2,

Kwangjo Kiml "An Efficient Tree-

based Group Key Agreement using

Bilinear Map", Information and

Communications University (ICU),

58-4, Hwaam-Dong, Yuseong-gu,

Daej eon, 305-732, Korea.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

232

COMPARISON OF CRYPTOGRAPHIC ALGORITHMS FOR SECURE

DATA TRANSMISSION IN WIRELESS SENSOR NETWORK

R.Vigneswaran

R.Vinod Mogan

Pre-Final B.Tech(It)

[email protected]

[email protected]

Abstract:

This paper proposes an secure data

transmission over the wireless sensor network

environment using various symmetric

cryptographic algorithms which provides a

guarantee to the transmitted data to be secure, no

encryption and decryption take place at cluster

head and comparison is made based on the

performance of different (AES, DES & Blowfish)

cryptographic algorithms.

Security is a critical requirement in data

transmission over the wireless sensor network

environment. A dishonest intermediate node can

read, modify or drop the data as well as send false

values to mislead the base station. Generally data

transmitted by the sensor node should be

decrypted at cluster head .The aggregated data is

then encrypted before being transmitted to the

base station. This technique is vulnerable from

security perspective because decryption of data

requires the cluster head to obtain the symmetric

key that may result in insecure data transmission.

Therefore sensor node sends encrypted data to the

cluster head and cluster head just forward the

encryption data from the sensor nodes to base

station and base station can decrypt the

transmitted data and obtain the plaintext.

Keywords: wireless sensor network,

encryption, decryptions and security.

Introduction:

Sensor network are dense wireless

network of small, and low cost sensors, which

collects, and disseminate (spread widely) the data

233

from remote location. It facilitate monitoring [4]

and controlling of physical environment from

remote location with better accuracy, sensor

network can be described as a collection of sensor

nodes which co-ordinate to perform some specific

function.

Sensor nodes are small, inexpensive low

power distributed devices that are capable of

processing the sensed data locally and

communicated them through wireless network.

Each sensor node is capable with only a limited

amount of processing but when co-ordinate with

the information from a large number of other

nodes, they have the ability to measure a physical

environment. Such sensor network are expected to

be widely deployed in a vast variety of

environment for commercial, civil and military

application such as surveillance, vehicle tracking,

climate and habitat monitoring, intelligence,

medical and acoustic data gathering The key

limitation of wireless sensor network are the

storage, power and processing. These limitations

and the specific architecture of sensor nodes call

for energy efficient and secure communication

protocol. The foremost challenge in sensor

networks is to maximize the lifetime of sensor

nodes due to the fact that it is not feasible to

replace the batteries of thousand of sensor nodes.

To obtain the estimated location

information of sensors, deployment of the sensor

nodes over the network area must be done by

following a procedure. Assume the sensor

network area is a square; the following is the

deployment procedure of our key distribution

scheme.

Sensor nodes are divided equally in to an

„n‟deployment groups

Sensor deployment area is divided equally

into „n‟deployment grids.

The center points of these „n‟grids is

determined as deployment points.

A group of sensor is dropped from each

deployment point

Figure1. Represent the sensor network area with

deployment pints, each deployment point is

represented by small circle while gray shaded

square represent a deployment grid. Deployment

grids are denoted by notation n (i, j).

Figure (1) also has neighboring grids that are

represented by White Square (si, j). This

neighboring area of shaded deployment grids

234

Figure No. 1 Sensor network area with sub-

square and deployment points

Sensor network is hierarchical architecture

where data is routed from sensor nodes to base

station through cluster heads. Base station

interface sensor network to the outside network.

Sensor nodes are assumed to immobile and also

they do not have a specific architecture when

deployed over a specific geographic area.

However, these nodes organize themselves into

clusters, based on self-organizing cluster to

handle the communication between the cluster to

handle the communication between the cluster

nodes and the base station.

Sensor nodes are battery powered and

their lifetime is limited. Therefore, cluster heads

are dynamically chosen initially. In order to have

uniform power consumption among all the sensor

nodes, cluster heads are then selected based on the

remaining battery power. To maximize the

lifetime, sensor nodes are in active mode when

they are put to sleep mode. Base station is

assumed to have sufficient power and memory to

communicate securely with all sensor nodes and

also with the external wired network.

Sensor nodes are assigned a secret key (ki)

and a unique ID number that identifies itself in the

network. As wireless transmission is not

completely trustable, assigning secret keys to

sensor nodes in wireless environment is not

secure. Hence Ki and sensor IDs are assigned

during the manufacturing phase. Base station is

then given all the ID numbers and Ki used in the

network before the deployment of network.

Having a complete list of sensors in the base

station protect sensor network from malicious

sensor nodes. Whenever base station broadcasts a

new Kb all the sensor nodes have to re-generate

their new secret session keys (Kt, b). The built in

keys in sensor nodes avoid the distribution if

secret keys in the wireless environment as well as

providing substantial security.

SECURE DATA TRANSMISSION

ALGORITHM FOR WIRELESS

SENSOR NETWORKS

The security protocol achieves secure data

transmission on wireless sensor network by

implementing the following five phases in two

algorithms: Algorithm A and B are implemented

in the sensor nodes and in the base station

respectively,

1) Broadcasting session key by base station is

performed in algorithm B.

235

2) Generation of cryptographic key in sensor

nodes is performed in algorithm A.

3) Transmission of encrypted data from sensor

nodes to cluster heads is performed in algorithm

A.

4) Appending the ID# to data and then forwarding

it to higher level cluster heads are performed in

algorithm A

5) Decryption and authentication of data by the

base station is performed in algorithm B.

The base station, periodically

broadcasts new session key to maintain data

freshness. Sensor nodes receive broadcasted

session key Kb and computes their nodes –

specific secret session key (Ki, b) by XORing

with Ki with Kb.Ki, b is used for all the

consequent data encryption and decryption during

that session by both algorithms. Since each sensor

node calculates Ki,b using its unique built in key,

encryption data with Ki,b also provides data

authentication in the proposed architecture.

Changing encryption keys time-to –time

guarantees data freshness in the sensor network,

moreover it helps to maintain confidentiality of

transmitted data by preventing the use of the same

secret key at all times. During the data

transmission, each sensor nodes appends its ID#

and time stamp to the messages to verify data

freshness. Additional security in the data

communication is obtained by sending the

encrypted data .The cluster head, when receives

the data from the nodes, appends its own ID#

before forwarding the data to the base station.

This helps the station in locating the origin of the

data.

Algorithm A: Implemented in sensor nodes

1). If sensor node I want to send data to its cluster

head, go to next step, and otherwise exit the

algorithm.

2). Sensor node I requests the cluster head to send

the current session key Kb.

3). Sensor node I XORs the current session key

(Kb) with it‟s built in key Ki to compute the

encryption key Ki, b.

4). Sensor node I encrypts the data with Ki, b and

appends its ID# and the time stamp to the

encrypted data and then sends them to the cluster

head.

5). Cluster head receives the data, appends its own

ID#, and then sends them to the higher –level

cluster head or the base station. Go to step 1.

236

Note: Cluster head appends its own ID# to the

data in order to help the base station to locate the

originating sensor node.

Algorithm B: Implemented in base station

1. Check if there is any need to broadcast the

session key Kb all the sensor nodes. If so,

broadcast Kb to all sensor nodes.

2. If there is no need for a new session key then

check if there is any incoming data from the

cluster heads. If there is no data being sent to the

base station, then go to step 1.

3.If there is any data coming to the base station

then compute the encryption key Ki,b using the

ID# of the node and the time stamp within there

data .base station ten uses the Ki,b to decrypt the

data.

4.Check if the current encryption key Ki,b has

decrypted the data perfectly .This leads to check

the creditability of the time stamp and the ID# .If

the decrypted data is not perfect discard the data

and go to step 6.

5. Process the decrypted data and obtain the

message sent by the sensor nodes.

6. Decides whether to request to all sensor nodes

for retransmission of data .If not necessary then

go back to step 1.

7. If a request is necessary the request to the

sensor nodes to retransmit the data. When this

session is finished go back to step 1.

In the above algorithms A and B we use prototype

sensor nodes of smart Dust .As shown table 1

these sensor nodes are resources constraint in

every sense.

IMPLEMENTATION

Considering restricted resource the cry to

graphic primitives used in sensor networks must

chosen carefully in term of their code size and

energy consumption therefore, we evaluated

several cryptographic algorithms to use for

encryption and decryption. It is observed from the

analysis that code space can be saved by using

same cipher for both encryption and decryption.

TABLE 1.CHARACTERISTICS OF

SMART DUST SENSORS

CPU 8-bit, 4 MHz

Storage 8K instruction flash

512 byte RAM

512 bytes

EEPROM

Communication 916 MHz radio

237

Bandwidth 10 kilobits per

second

Operating system TinyOS

OS code space 3500 bytes

Available code space 4500 bytes

The number of clock needed by the

processor to compute the security function

primarily determines the amount of computational

energy consumed by a security function on a

given microprocessor. The number of clocks

necessary to perform the security function mainly

depends on the efficiency of the cryptographic

algorithms.

In the above methods of data transmission the

security algorithms faces the some disadvantages.

This is Limited number of Storage available and it

takes additional memory space and takes more

delay time. In base station we get less number of

throughputs. Security function determined in

microprocessor needs more of clock signal. Due

to this disadvantage we cannot get secure data in

base station correctly .In order to overcome these

shortcomings and to reduce interference in data

transmission we introduce various cryptographic

algorithms implemented using “Network

simulator (NS-2)”and find the performance results

such as throughput, delay time

Network security:

Security is a critical requirement in data

transmission [3] over the wireless sensor network

environment. A dishonest inter mediate node can

read, modify or drop the data as well as send false

values to mislead base station. Asymmetric

cryptographic algorithms are not suitable to

provide security on wireless sensor network since

they require high computation power, and storage

resources. Therefore symmetric key cryptographic

algorithms are employed to support security

because of limited key length and memory

available on the sensor nodes. In symmetric key

encryption or secret key encryption, only one key

is used to encrypt and decrypt data and the key

should be distributed before transmission between

two entities. It is also known to be very efficient

since the key size can be small. Generally data

transmitted by the sensor nodes should be

decrypted at cluster head.

The data was encrypted before being

transmitted to the base station. This technique is

vulnerable from security perspective because

decryption of data requires the cluster head to

obtain the symmetric key that may result in

insecure data transmission.

238

Figure No. 2 Encryption Model

Therefore sensor node send encrypted data

to the cluster head and cluster head just forward

the encrypted data from the sensor nodes to base

station and base station can decrypt the

transmitted data and obtain the plaintext.

Performance between symmetric cryptography

algorithms are proposed based on throughput as

well as end-to-end delay using „Network

simulator (NS-2)‟.

Throughput is the amount of information that is

successfully transmitted from the transmitting end

to receiving end at the instant of time taken to

receive the last packet.

Cryptographic Algorithms:

1. Blow fish:

a) Blowfish is a symmetric block cipher that can

be used as drop in replacement for DES or IDEA

[4].

b). It takes variable length key, from 32 bits to

128 bits. The block size is 64 bits.

c). It is ideal for both domestic and exportable

use.

d) It is much faster than DES and IDEA,

unpatebted and royalty free, no license are

required.

e). It iterates a simple encryption function by 16

times.

f). In DES, 56 bits key size is vulnerable to brute

force attack, and recent advance in differential

cryptanalysis and linear cryptanalysis indicate that

DES is vulnerable to other attacks as well.

g). Data encryption occurs via a 16round each

round consist of key dependent permutation and a

key and data dependent substitution .All operation

are xores and addition on 32 bit word. The only

additional operation is fore indexed array data

lookups per round.

2. Data Encryption Standard (DES)

In data encryption standard key [4 consists of 64

binary digits („0‟s or 1‟s) of which 56 bits are

randomly generated and used directly by the

algorithms .The others 8‟bits, which are not used

by the algorithms, are used for error detection.

239

The algorithm is designed to encrypt and decrypt

blocks of data consisting of 64 bits under control

of 64-bit key. The unique key chosen for use in a

particular application makes the results of

encrypting data using the algorithm unique.

Selection of a different key causes the cipher that

is produced for any given set of inputs to be

different. The cryptographic security of the data

depends on the security provided for the key used

to encipher and decipher the data. The algorithm

is designed to encipher and decipher blocks of

data consisting of 64 bits under control of a64–bit

key**. Deciphering must be accomplished by

using the same key used for enciphering, but with

the schedule of addressing the key bits altered so

that the deciphering process is the reverse of the

enciphering process. A block to be enciphered is

subjected to an initial permutation IP, then to a

complex key- dependent computation and finally

to a permutation which is the inverse of the initial

permutation IP inverse. The key –dependent

computation can be simply defined in terms of a

function f called the cipher function, and a

function Ks called the schedule.

3. Advance encryption standard (AES):

AES [9] must be symmetric block cipher with a

block length of 128 bits and support for key

lengths of 128,192,256, bits. AES must have high

computational efficiency, so as to be usable in

high-speed application links. Analysis of AES

compared to DES the amount of time requirement

for the operation is limited. The input to the

encryption and decryption algorithm is a single

128 –bit block. This block is copied into the state

array, which is modified at each stage of

encryption and decryption. After the final stage,

state is copied to an output matrix. Similarly 128-

bit key is depicted as a square matrix of bytes.

This key is then expanded into an array of key

schedule is 44 words for the 128 bit key, ordering

the bytes within a matrix is by column for

example, the first four bytes of a 128-bit plaintext

input encryption cipher occupy the first column

and the second four byte occupy second column,

and so as on.

Overall AES structure:

1. AES do not use a festal structure but process

the entire data block in parallel during each round

using substitution and permutation.

2. The key that is provided as input is expanded

into an array of forty-four 32-bit worlds [i]. Four

distinct words serve as each round.

3. Four different stages are used, one of

permutation and three of substitution.

Substitution bytes: Uses an S-Box to perform a

byte-by-byte substitution of the block.

Shift rows: A simple permutation

240

Mix columns: A substitution that makes use of

arithmetic over GF (2^8)

Add round key: A simple bit wise XOR of the

current block with apportion of the expanded key.

4. For both encryption and decryption, the cipher

begins with an add round key stage, followed by

nine rounds that each includes all four stages.

5. Only the Add Round Key stage makes use of

the key. For this reason, the cipher begins ends

with an Add Round Key stage.

6. Cipher as the alternating operation of XOR

encryption (Add Round Key) of the block,

followed by scrambling of the block (the other

three stages), and followed by XOR encryption,

and so on. This scheme is efficient highly secure.

7. Each stage is easily reversible .for the substitute

byte, shift Row, and Mix columns a stage, an

inverse function is used in the decryption

algorithm. For the Add round key stage, the

inverse is achieved by XORing the same round

key to the block, using the result that A XORs A

XORs B = B.

8. As with most block ciphers, the decryption

algorithm makes use of the expanded key in

reverse order. However the decryption algorithm

is not identical to the encryption algorithm.

9. The final round of the both encryption and

decryption consists of only three stages again.

Again, this is a consequence of the particular

structure of AES and is required to make the

cipher reversible.

Rijndael (AES) was designed to have the

following characteristics:

Resistance against all attacks.

Speed and code compactness on a wide

range of platforms.

Design simplicity.

Simulation Results:

Fig 3.Load Vs Throughput graph for

cryptographic algorithms

From the above three curves red co lour curve

is load versus throughput graph of AES algorithm,

241

blue co lour curve is load versus throughput graph

of Blowfish algorithm and green co lour curve is

load versus throughput graph of DES algorithm.

From above three curve show that Blowfish is

better than DES and AES is better than Blowfish.

So, AES highest throughput compare than three

curves.

Fig 4.load Vs Time delay graph for

cryptographic algorithms

fig 3 & 4 depict that AES consumes less

Delay and 33% more throughput compared with

DES and 16% more throughput compared with

Blowfish,

From the above three curves red co lour curve

is load versus time delay graph of the AES

algorithm, blue co lour curve is load versus graph

of Blowfish algorithm and green co lour curve is

load versus time delay graph of DES algorithm.

From above three curve show that Blowfish is

better than DES and AES is better than Blowfish.

So, AES lowest time delay compare than three.

Conclusion:

In this project we have designed a secure

data transmission in wireless sensor networks by

using symmetric key cryptographic algorithms. It

has been proposed in this work that the sensor

node send encrypted data to the cluster head and

cluster head just forward it to the base station. The

base station can decrypt the transmitted data and

obtain the plaintext. From the simulation results it

can be seen that AES algorithm consumes less

time delay and give more throughputs compared

with DES and Blowfish algorithms. Therefore

AES algorithm is better for secure data

transmission

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

242

FREQUENCY SYNCHRONIZATON IN 4G WIRELESS COMMUNICATION

BHUVANESHWARAN.C , VASANTH KUMAR.M

III Year UG students, Department of Electrical and Electronics Engineering

Thiagarajar College of Engineering, Madurai-625015

Ph.no:9791249011, 9043844059

Email id: [email protected] , [email protected]

Abstract

Third Generation Partner Project

(3GPP) is considering Long Term

Evolution for both radio interface and

network characteristics. These systems

support reliable high data communication

over time dispersive channels with limited

spectrum and Inter Symbol Interference

(ISI). Orthogonal Frequency Division

Multiple Access (OFDMA) is used as a

multiple access technology in the

downlink of LTI systems and OFDM as

data transmission technology. The OFDM

is highly sensitive to timing and frequency

synchronization errors causing loss of

orthogonality among sub-carriers, because

orthogonality allows simultaneous

transmission on a lot of sub-carriers in a

tight frequency space without interference

with each other. This paper addresses

algorithm for frequency synchronization

for downlink LTE using defined

synchronization sequence. The detection

algorithm along with simulation is

presented.

Introduction

Long Term Evolution (LTE) is the

fourth generation of radio technologies

designed to increase the capacity and

speed of the mobile telephone network. It

is standard being specified by 3GPP within

release 8. In Orthogonal Frequency

Division Multiple Access (OFDMA) used

for downlink transmission and for uplink

Single Carrier Frequency Division

Multiple Access (SC-FDMA) has been

specified by 3GPP. It offers favourable

features such as high spectral efficiency,

robust performance in frequency selective

channel conditions, simple receiver

architecture , etc

However it is well known that

OFDM systems are more prone to time

and frequency synchronization errors,

hence accurate synchronization is needed

for interference free data reception.

Furthermore, when an user equipment

(UE) operates in a cellular system, it needs

to establish connection as fast as possible

with the best serving base station. So time

and frequency synchronization is the

primary criterion to be considered for any

LTE systems.

The frequency synchronization for

3GPP LTE has been investigated in this

paper and also a detection algorithm for

frequency offset is described.

LTE frame structure

Fig.1. LTE FRAME STRUCURE

243

LTE radio frames are 10ms in

duration. Each frame has ten sub-frames

and each sub-frame consists of two slots

each of 0.5ms duration.Each slots consists

of six or seven OFDM symbols depending

on whether normal or extended cyclic

prefix used.

The number of sub-carriers allocated

depends on the available bandwidth of the

system. The LTE specification defines

bandwidth from 1.25MHZ to 20MHZ. The

physical resource block (PRB) is defined

as time-frequency grid consisting of 12

sub-carriers in each slot (0.5ms). Resource

element is the smallest element in the time

frequency grid.

Fig.2.DOWNLINK RESOURCE GRID

0FDM system model

OFDM is a combination of

modulation and multiplexing.

Multiplexing generally refers to

independent signals, those produced by

different sources. In OFDM the signal

itself is first split into independent

channels, modulated by data and then re-

multiplexed to create OFDM carrier. The

modulation scheme may be any of the

following method, BPSK,QPSK,8PSK,32-

QAM or whatever.

Fig.3.OFDM system model

The OFDM signal is generated at

baseband by taking Inverse Fast Fourier

Transform (IFFT) of Binary Phase Shift

Keying (BPSK) modulated signal. The

data x(k) are modulated on N sub-carriers

by IFFT and last L samples are copied and

put as a preamble (cyclic prefix) to form

OFDM frame s(k). This data is transmitted

over the channel. The reason for addition

of cyclic prefix is to mitigate the effects of

fading, inter symbol interference and

increase bandwidth.

The transmitted signal s(k) is

affected by complex, Additive White

Gaussian Noise (AWGN) n(k). The

uncertainty in carrier frequency occurs in

the receiver that may affect the s(k) is due

244

to difference in the local oscillator in the

transmitter and the receiver, and is

modelled as a complex multiplicative

distortion with a factor of the

received data. Where ε denotes the

difference in the transmitter and receiver

oscillators relative to the inter-carrier

spacing known as the frequency offset.

Hence the received data is given by

r(k)= [s(k)+n(k)]

(we considered only frequency offset)

The data y(k) is obtained from r(k)

by discarding last L samples (cyclic prefix)

and demodulating the n remaining samples

of each frame by means of a FFT. Since

the symbol error mainly depends on ε, our

aim is to estimate this parameter from the

received data r(k).

Effect of frequency offset

Frequency synchronization is

necessary to preserve orthogonality

between the sub-carriers. OFDM are more

susceptible to frequency offset errors. So a

robust frequency algorithm is to be

developed. The estimation of frequency is

done by the synchronization signals.

LTE synchronization signals

A dedicated synchronization signal

is specified for LTE by 3GPP. The

sequence d(n) used for the primary

synchronization signal is generated from a

frequency domain ZADOFF-CHU

sequence

Where the Zadoff-CHU root index u can

take the values:25,29,34.

The synchronization sequence is

mapped on to 62 sub-carriers located

symmetrically around the DC-carrier.

Frequency Synchronization Algorithm

The objective of synchronization is to

estimate carrier frequency offset (CFO).

Since ε denotes the frequency offset.

Let s(n) will be obtained after taking

IFFT,

r(n)= s(n)* exp((j*2*Pi* ε)/N)

To calculate frequency offset,multiply

the received signal with conjugate of ifft of

the sequence and sum the product which is

given as

y=r(n) * [conj(d(n))]

Let θ be the angle of y,

.

The frequency offset can be derived as

ε= [y*N] / [(N+1 )*pi]

Simulations

Here we used Monte Carlo

simulations for evaluating the performance

of the presented algorithm. The 1.25 MHZ

system is considered with 128 point

FFT/IFFT, 9 sample Cyclic Prefix (CP)

and a 15 KHZ sub-carrier spacing.

Simulations were performed over

sufficient number of sub-frame based

transmissions over independent channel

realization for SNR = 0 to 10 dB.

In figure the root mean square error

(RMSE) of the OFDM symbol frequency

estimation is plotted against SNR.

Conclusion

By using the proposed algorithm

the frequency offset for the LTE systems

can be determined easily. In this algorithm

the estimation of frequency offset is

performed with the primary

synchronization signal.

Proceedings of the Third National Conference on RTICT 2010 Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

245

NETWORK SECURITY (BINARY LEVEL ENCRYPTION)

D.Ashok, T.Arunraj,

B-Tech Information Technology, K.S.RANGASAMY COLLEGE OF TECHNOLOGY K.S.R. KALVI NAGAR, TIRUCHENGODE

[email protected]

[email protected]

ABSTRACT

Cryptography – The art of converting the

data to unintelligible form (Encryption) to make it

secure to be transferred over the network and

retrieves the data back (Decryption), only if the

receiver is the intended one i.e., the one who knows

the key. One of the main principle’s of

cryptography is the Principle of Weakest Link

which states that “Security can be no stronger than

its weakest link”.

Kerckhoff’s principle is that “All

algorithms must be public; only the keys are

secret”. This shows the vitality of the encryption

algorithm. In all encryption algorithms, the plain

text is modified based on the key. This gives room

for intruders to break the key. Conventional

Encryption Algorithms suffer from various

problems making them susceptible to various forms

of attack including Frequency analysis, Brute force

attack, etc.

Using keys in English may appear to have

some inherent pattern and is hence possible to break

such a code. Using random numbers provides an

increased level of security during transmission. Key

tables are popularly used for encryption. Once the

key table has been identified by any of the

cryptanalytic means, all the messages being sent

between the source and the destination can be easily

decrypted by the intruder. In our proposed

algorithm, we use binary random numbers as keys

for encryption. The use of binary random numbers

makes cracking the code virtually impossible. We

also propose to use a different encryption table each

time the messages are encoded and sent. The use of

binary digits and a varying encryption table

provides an increased level of security in message

transfer. We also created software for our proposed

algorithm.

246

CRYPTOGRAPHY:

The messages to be encrypted, known as

plain text, are transformed to cipher text by a

function that is parameterized by a key. This art of

devising ciphers is called as cryptography.

CRYPTANALYSIS:

This deals with the breaking of a cipher to

recover information or forging encrypted

information that will be accepted as authentic.

There are two approaches to attack a conventional

encryption scheme.

Cryptanalysis: Cryptanalytic attacks rely

on the nature of the algorithm plus the

general knowledge of the characteristics of

the plain text. This type of attack exploits

characteristics of the algorithm to attempt to

deduce a specific plain text or to deduce the

key being used. If the attack succeeds in

deducing the key, the effect is catastrophic.

All future and past messages encrypted with

that key are compromised.

Brute Force Attack: The intruder tries

every possible key on a piece of cipher text

until an intelligible translation into plain text

is obtained.

GENERAL PLAINTEXT – KEY RELATION:

The key is a value independent of the

plaintext.

C = E (P, K)

P = D (C, K)

C - Cipher P - Plaintext K – Key

E - Encryption algorithm D - Decryption

algorithm

SECURITY GOALS:

The basic goals to be ensured for a secure

communication are as mentioned below

Confidentiality – protection of transmitted

data from passive attacks.

Integrity – assurance that the data received

are exactly as sent by an authorized entity

Availability – the property of a system or

resource being accessible and usable upon

demand by an authorized entity.

RANDOM NUMBER GENERATION:

Random numbers play an important role in

encryption. The random numbers are combined

with the plain text before transmission. But the

generation of random numbers is a daunting task.

The generation of random numbers cannot be truly

random and follow some well-defined statistical

sense. They are also highly unpredictable. Sources

of true random number generators are few and far

between. The decoding procedure is also highly

complex.

RSA ALGORITHM:

247

The RSA scheme is a block cipher

developed by Rivest, Shamir and Adelman. The

plain text and the cipher text values are less than

some value n. The Encryption and Decryption are

of the following form:

C = Pe mod n

P = Cd mod n

C – cipher text P – plain text n – a

large number

e – encryption key d – decryption key

In our proposed algorithm, we use this

method to transmit the key and encryption table.

DES ALGORITHM:

This is a scheme based on the Data

Encryption Standard adopted by the National

Bureau of Standards in 1977. It is one of the best

algorithms proposed so far. The DES algorithm

follows the following five steps: an initial

permutation, a complex permutation and

substitution, a simple permutation that switches the

two halves of the data, a complex permutation –

substitution again and finally a permutation that is

the inverse of the first permutation.

Our algorithm is based on this principle of

permutation and substitution and incorporates the

additional security of binary manipulations.

KEYWORDS USED:

Plain text – Text to be encrypted

Cipher text – Encrypted text

Cryptanalysis –The process used by a

cryptanalyst or cryptographer to break an

algorithm.

Cryptanalyst – One who tries to retrieve the

information illegitimately.

CORE OF THE PROPOSED METHOD:

A key table and an encryption table contain

codes to convert the English alphabet to binary.

First, convert the entire plain text and key to binary.

Reverse the every alphabet of the key. Add the

plain text to the reversed key using the ex-or

mechanism. Take the binary value of every letter of

the key, divide it into two, reverse the left half alone

and perform ex-or addition with the plain text.

Finally, reverse the right half of the original key

alone and perform ex-or addition with the plain

text. Repeat the above process 26 times to ensure

maximum security.

PROCEDURE:

AT THE TRANSMITTING ENDS:

Build an encryption table:

Choose any random number less

than 512(since we have assumed 8 bit

codes). Assign it to the first alphabet (i.e.

A). Assume any other random number

considerably less than the first. Generate the

next number by adding the above mentioned

random numbers. This is assigned to the

next alphabet (i.e., B). Add the latter to the

last generated number to generate the next

248

alphabet (i.e. C). This procedure is followed

till all the alphabets have binary numbers.

Building a key table:

The key table is generated in a

manner similar to the encryption table.

Ensure that the random number chosen for

the key table and the encryption table are

different.

The following steps should

be repeated 26 times

Reverse the binary value of

every alphabet of the key

Add the plain text and the

reversed key using ex-or

addition.

Consider the binary value of

every letter of the key. Split the

key into two halves. Reverse the

left half of the key.

Add that value to the text

resulting after the first step.

Consider the initially taken key.

Split the key into two halves.

Reverse the right half of the key

this time around.

Add that value to the text

resulting after that third step.

Once the above procedure is

complete, the key table must be

rotated.

Transmitting the cipher:

The first number of the

encryption table and the key

table are concatenated together.

This results in a large number.

This number is encoded using

the RSA algorithm.

This number is sent first,

followed by the encrypted key

and then the resulting cipher text

in binary.

IMPLEMENTATION OF ENCRYPTION:

Encryption table:

A 00000101 K 01001011 U 10010001 B 00001100 L 01010010 V 10011000 C 00010011 M 01011001 W 10011111 D 00011010 N 01100000 X 10100110 E 00100001 O 01100111 Y 10101101 F 00101000 P 01101110 Z 10110100 G 00101111 Q 01110101 H 00110110 R 01111100 I 00111101 S 10000011 J 01000100 T 10001010

Key table:

A 00001011 K 01010001 U 10010111 B 00010010 L 01011000 V 10011110 C 00011001 M 01011111 W 10100101 D 00100000 N 01100110 X 10101100 E 00100111 O 01101101 Y 10110011 F 00101110 P 01110100 Z 10111010 G 00110101 Q 01111011 H 00111100 R 10000010 I 01000011 S 10001001 J 01001010 T 10010000

The plain text is BEANS and the key used is

SNAKE

P: 0000 1100 0010 0001 0000 0101 0110 0000

1000 0011

K: 1000 1001 0110 0110 0000 1011 0101 0001

0010 0111

249

Step 1:

i. Reverse the key

K: 1001 0001 0110 0110 1101 0000 1000

1010 1110 0100

ii. Add it to the plain text

P: 0000 1100 0010 0001 0000 0101 0110

0000 1000 0011

K: 1001 0001 0110 0110 1101 0000 1000

1010 1110 0100

1001 1101 0100 0111 1101 0101

1110 1010 0110 0111

iii. Reverse the left half of the

original key

K: 0001 1001 0110 0110 0000 1011 1010

0001 0100 0111

iv. Add it to the P resulting from

ii.

P: 1001 1101 0100 0111 1101 0101 1110

1010 0110 0111

K: 0001 1001 0110 0110 0000 1011 1010

0001 0100 0111

1001 0100 0010 0001 1101 1110

0100 1011 0010 0000

v. Reverse the right half of the

original key

K: 1000 1001 0110 0110 0000 1011 0101

0001 0010 0111

vi. Add it to the P resulting from

iv.

P: 1001 0100 0010 0001 1101 1110 0100

1011 0010 0000

K: 1000 1001 0110 0110 0000 1011 0101

0001 0010 0111

0001 1101 0100 0111 1101 0101

0001 1010 0000 0111

vii. Rotate the entire key table

such that a = b, b = c, and so on till

z = a.

Resulting key table:

A 00010010 K 01011000 U 10011110 B 00011001 L 01011111 V 10100101 C 00100000 M 01100110 W 10101100 D 00100111 N 01101101 X 10110011 E 00101110 O 01110100 Y 10111010 F 00110101 P 01111011 Z 00001011 G 00111100 Q 10000010 H 01000011 R 10001001 I 01001010 S 10010000 J 01010001 T 10010111

The above steps are repeated 25 times. The

resulting binary sequence is,

0001 1101 1101 1000 0110 0001

1001 0111 0010 1010

The above binary sequence is transmitted

without any of the spaces.

AT THE RECIEVER ENDS:

The RSA encoded number is received.

Decode the number and retrieve the

numbers used.

250

Generate the encryption and the key tables

appropriately.

The encrypted key is received next.

Decrypt the received key.

Use the same procedure to decipher the

received cipher text.

IMPLEMENTATION OF DECRYPTION:

The steps that are followed for encryption

are followed for the decryption process as well.

Since we use ex-or operations on the binary

numbers, the key table and the encryption table

used at the transmitter can be used at the receiver.

Step 1:

i. Reverse the key

K: 1001 0001 0110 0110 1101 0000 1000

1010 1110 0100

ii. Add it to the plain text

C: 0001 1101 1101 1000 0110 0001 1001

0111 0010 1010

K: 1001 0001 0110 0110 1101 0000 1000

1010 1110 0100

1000 1100 1011 1110 1011

0001 0001 1101 1100 1110

iii. Reverse the left half of the

original key

K: 0001 1001 0110 0110 0000 1011 1010

0001 0100 0111

iv. Add it to the C resulting from

ii.

C: 1000 1100 1011 1110 1011 0001 0001

1101 1100 1110

K: 0001 1001 0110 0110 0000 1011 1010

0001 0100 0111

1001 0101 1101 1000 1011

1010 1011 1100 1000 1001

v. Reverse the right half of the

original key

K: 1000 1001 0110 0110 0000 1011 0101

0001 0010 0111

vi. Add it to the C resulting from

iv.

C: 1001 0101 1101 1000 1011 1010 1011

1100 1000 1001

K: 1000 1001 0110 0110 0000 1011 0101

0001 0010 0111

0001 1101 1011 1110 1011

0001 1110 1101 1010 1110

vii. Rotate the entire key table

such that a = b, b = c, and so on till

z = a.

The following steps are repeated and at step

26 we end up with the plain text value

P: 0000 1100 0010 0001 0000 0101 0110 0000

1000 0011

Referring to the table we get the plain text message

BEANS

STRENGTHS:

The following are the strengths of our

proposed algorithm:

251

Binary manipulations are

performed and hence can be

easily done using hardware.

Two levels of keys used, hence

increased level of security.

Dynamically generated tables.

Algorithm can be extended to 16/

24/ 32 bits depending on the

amount of security required.

Both encryption and decryption

use the same algorithm

SOFTWARE IMPLEMENTATION:

P1 := binary_ptext(P);

for i := 1 to 26 do

K1 := binary_key(K);

for j:= 1 to m do

K2 := reverse binary value of

every alphabet of the key

K1;

end for

P2 := P1 (+) K2;

for j:= 1 to m do

K3 := reverse left half of

binary value of every alphabet

of the key K1;

end for

P3 := P2 (+) K3;

for j:= 1 to m do

K4 := reverse right half of

binary value of every

alphabet of the key K1;

end for

P4 := P3 (+) K4;

rotate the key table once;

P1 := P4;

end for

Note: - (+) denotes Ex-or addition

HARDWARE IMPLEMENTATION:

COMPLEXITY OF THE ALGORITHM:

On using software techniques for the

encryption and decryption process, we have found

the complexity of our algorithm to be of the order

of m*n.

This can be further reduced by the use of

rapidly developing hardware. The complexity in

this case is of the order of m.

Where, m indicates the number of alphabets and n

indicates the number of bits used.

252

NOTE:

The transfer of the key is by using any

existing key transfer mechanism. The transfer of

keys should be effective to ensure that

confidentiality is maintained. Cipher cannot be

converted back to English alphabets. This is

because the resulting binary values are not part of

the encryption table. To ensure an error free

transmission, some error checking and correcting

mechanisms are appended to the cipher.

FUTURE DEVELOPMENTS:

To devise our own key transfer mechanism.

To convert our algorithm to an asymmetric

mode of encryption for increased security.

REFERNCES:

1. Stallings, W. Cryptography and Network

Security: Principles and Practice, Third

edition. Upper Saddle River, NJ: Prentice

Hall, 1999.

2. Scheneier, B. Applied Cryptography. New

York: Wiley, 1996.

3. Stallings, W. Network Security Essentials:

Applications and Standards. Delhi, India:

Pearson Education Asia, Reprint 2001.

4. Pfleeger, Charles and Pfleeger, Shari. P., Elementary

Cryptography, Third edition. Delhi, India: Pearson

Education Asia.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

253

Improving Broadcasting Efficiency in MANET by

Intelligent Flooding Algorithm

Madala V Satyanaryanaya (M.Tech C.S.E)

Guide:

Dr. M.A. DoraiRanga Swamy,

QIS College of Engg. & Tech.

Abstract:

Broadcasting is a common operation in a network to resolve many issues in

Mobile Ad hoc Networks (MANET) in particular due to host mobility. Such operations

are expected more frequently, eg. such as finding a route to particular host, and sending

alert signals, etc. In this paper we mainly consider the three factors - bandwidth

utilization, computational and space complexity, and power utilization in the network as

low as possible. We consider sender-based broad casting algorithms, specified by Liu et

al and Mojid Khabbazian. In both sender-based broadcasting algorithms, they maintained

list of forwarding nodes attached in the message and selection of subset of neighbors. In

each host it improves space and computational complexity. To overcome these factors,

we proposed a simple and efficient algorithm, Intelligent Flooding Algorithm (IFA),

which efficiently utilizes bandwidth as low as possible and reduces space and

computational complexity by reducing the number of broadcasting host, and redundant

rebroadcast in the network. Using simulation, we confirm these results and show that the

number of broadcasts in our proposed IFA can be even less than one of the best known

approximations for the minimum number of required broadcasts.

* * *

Key words: Mobile Ad hoc Networks, wireless networks, broadcasting, and

flooding.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

254

Implementation of SE Decoding Algorithm in

FPGA for MIMO Detection

S. Sharmila Amirtham1, S. Karthie

2

1 M.E. (Applied Electronics) Dept of ECE, SSN College of Engineering, Kalavakkam, (TN) 2Assistant Professor Dept. of ECE, SSN College of Engineering, kalavakkam, (TN)

[email protected],

[email protected]

Abstract

Multiple-Input-Multiple-Output (MIMO) systems use

multiple antennas in both transmitter and receiver ends for

higher spectrum efficiency. The hardware implementation of

MIMO detection becomes a challenging task as the

computational complexity increases. MIMO sphere decoding

algorithm, the Schnorr-Euchner (SE) algorithm is taken for

analysis. The SE decoding algorithm is implemented on an

FPGA platform and are evaluated for decoding rates, BER

performance, Power consumptions, and area utilizations. FPGA

based implementation of this algorithm is advantageous

compared to DSP based sphere decoding algorithm

implementation.

Key Words: FPGA, MIMO detection, lattice point search,

sphere decoding.

I. INTRODUCTION

The use of multiple input-multiple output (MIMO)

technology in wireless communication systems enables

high-rate data transfers and improved link quality through

the use of multiple antennas at both transmitter and

receiver. It has become a key technology to achieve the bit rates that will be available in next-generation wireless

communication systems, combining spatial multiplexing

and space-time coding techniques. In addition, the

prototyping of those MIMO systems has become

increasingly important in recent years to validate the

enhancements advanced by analytical results. For that

purpose, field-programmable gate arrays (FPGAs), with

their high level of parallelism and embedded multipliers,

represent a suitable prototyping platform.

The information theory for MIMO systems has been

well studied on performance parameters such as data rate and bit error rate (BER). The layered space-time receiver

structures and coding schemes have allowed the MIMO

systems to approach the theoretical capacities on a

multiantenna channel. On the receiver end, one of the key

functions is to perform channel decoding to recover the

original data stream corresponding to each of the

transmitted antennas from the receiving signal vector and

estimated channel information. Both lattice theory and

coding theory are applied in the design of MIMO detection

algorithms. In a multiple antenna channel environment,

each of the transmitted signal vectors is aligned on the

modulated constellation points. Therefore, a multilayered lattice is formed with a set of finite points and the MIMO

detection is essentially an algorithm to search for the

closest lattice point to the received vector. There are two

typical classes of comprehensive search algorithms for a

lattice without an exploitable structure. One is the Pohst

strategy that examines lattice points lying inside a

hypersphere. The lattice decoding algorithm developed by

Viterbo and Boutros is based on the Pohst strategy.

Another class of lattice search strategy is suggested by

Schnorr and Euchner, based on examining the points inside

the aforementioned hypersphere in zig zag order of lattice layers with nondecreasing distance from the received

signal vector. A representative lattice decoding algorithm

based on Schnorr–Euchner (SE) strategy is applied by

Agrell et al. Both lattice search algorithms solve the

maximum-likelihood (ML) detection problem. Both

algorithms are considered the most promising approaches

for MIMO detection, and are also commonly referred as

sphere decoders since the algorithms search for the closest

lattice point within a hypersphere.

While it would require more operations on a standard

DSP it is much more suitable for a VLSI implementation.

It completely avoids the numerically critical operation of analytically finding an initial branch for the enumeration

procedure. Instead, it examines all branches that originate

from the current node in parallel and finds the smallest

one. This allows to operate directly on arbitrary complex

constellations without increasing the dimension of the

problem (i.e. the depth of the tree). The architecture is

designed to visit one node in each clock cycle. Its main

advantage in terms of throughput results from the fact that

no node is ever visited twice. The basic principle of SDA

is to avoid the exponentially complex exhaustive search in

the signal constellations, by applying a sphere constraint (only the constellation points within the sphere would be

considered) and transform the ML detection problem into a

tree search and pruning process. Regular SDA conducts a

depth-first search in the tree while the K-Best lattice

decoding algorithm, a variant of SDA, does a breadth-first

tree search. The latter approach, however, has performance

degradation unless K is sufficiently large.

Field-programmable gate-array (FPGA) devices are

widely used in signal processing, communications, and

network applications because of their reconfigurability and

support of parallelism. FPGA has at least three advantages

over a DSP processor; the inherent parallelism of an FPGA is equipped for vector processing; it has reduced

instruction overhead; the processing capacity is scalable if

the FPGA resource is available. The development cycle of

the FPGA design is usually longer than the DSP

implementation. But once an efficient architecture is

255

developed and the parallel implementation is explored,

FPGA is able to significantly improve the processing speed

because of its intrinsic density advantage. FPGA also has

several advantages over an ASIC implementation: an

FPGA device is reconfigurable to accommodate system

configuration changes even in run-time; it has significantly reduced prototyping latency comparing to ASIC; it is a

cost-effective solution to meet the low volume short cycle

product requirement.

II. SPHERE DECODING ALGORITHM

The main idea in SD is to reduce the number of

candidate vector symbols to be considered in the search,

without accidentally excluding the ML solution. This goal is achieved by constraining the search to only those points

Hs that lie inside a hyper sphere with radius around the

received point. The corresponding inequality is referred to

as the sphere constraint (SC).

d(s) ≤ C2 with d(s) ‖Hs – y‖

2 . (1)

The fundamental idea is to reduce the number of

candidate vector symbols that need to be considered in the

search for the ML solution. To this end, the search is constrained to only those candidate vector symbols s for

which Hs lies inside a hypersphere with radius c around

the received point y as illustrated in Figure 1 The

corresponding inequality is given by

‖y- Hs‖2 < C2 (2)

and the above equation will be referred to as the sphere

constraint (SC). Complexity reduction through tree

pruning is enabled by realizing that the SC can be applied

to identify admissible nodes on all levels of the tree

according to

di (s(i)

) < C2 (3)

because it is known that if the PED of any node within the

search tree violates the (partial) SC, all of its children and

eventually also the corresponding leaves will also violate

the SC.

Fig 1: Illustration of the sphere constraint

Considering an MIMO system with M transmit and N

receive antennas, the received signal y is given by,

y= Hs+n (4)

where s is the transmitted signal vector and n is additive

white Gaussian noise vector. H is the M×N channel matrix

that can be assumed as known from perfect channel

estimation and synchronization. For selected modulation

scheme, each element of the transmit vector s is a constellation point and the channel matrix H generates a

lattice. The ML decoding algorithm is to find the minimal

distance between the received point and the examining

lattice point that

s = arg TM

u min ‖y- sH‖2 (5)

where s is the decoded vector. The entries of s are chosen

from a complex constellation Ω. The set of all possible

transmitted vector symbols is denoted by ΩM.

III. SE DECODING ALGORITHM

We consider a spatial multiplexing MIMO system

with M transmit and N receive antennas. The transmitter

sends M spatial streams. Assume the transmitted symbol is

taken from complex constellation Ω. The transmission of

vector s over MIMO channels can be modelled as y =

Hs+n, where y is an N×1 receive signal vector. H is the

M×N channel matrix that can be assumed as known from

perfect channel estimation and synchronization.

Depth-first SD algorithm is divided into two parts, the first part is the partial Euclidean distance (PED)

calculation and the second part is the tree traversal. The

lattice generation matrix H is factorized into a lower

triangular matrix R and an orthonormal matrix Q using KZ

reduction or LLL reduction where H= R-1Q.The closest

lattice point problem is formulated as

s = arg TM

u min ‖yQT- sR-1‖2 (6)

The index uk is calculated and examined in two steps as

ek = yQTR

uk = [ekk], [ekk] ± 1, [ekk] ± 2,….. (7)

The orthogonal distance dk to the K-dimensional layer can

be found as

dk = (ekk – uk) / rkk (8)

where rkk is the diagonal element of R. Thus, the partial

Euclidean distance (PED) up to the K-dimensional layer is

calculated as

PEDk =

M

ki

id 2 (9)

If the PEDk is less than the current best distance dbest, it is

stored and the search procedure expands to (K-1)

dimensional sublayer with an update

ek-1, i = ek,i – dkrk,i, where i = 1, 2,…,k-1. (10)

Conversely, if the PEDk is greater than dbest, the

searching procedure steps back by one dimensional layer,

followed by an update of the examining index in zig-zag

order, which leads to a nondecreasing distance from the examining index to the current layer.

The procedure exits when the search moves down to

the bottom layer without finding a shorter distance.

256

Therefore, the best lattice point found so far becomes the

output.

The step-by-step procedures of the SE algorithm are

illustrated in the following.

1) Preprocessing: Perform QR factorization on channel

matrix H and inversion of the triangular matrix R. Initialize dimensional index K = M find the bounded index uk and

the distance dk from y to the current sublayer M , and set

dbest = ∞.

2) FSM: Updates PEDk using dk ; let dnew = PEDk.

if dnew < dbest and k > 1, go to state A;

if dnew < dbest and k = 1, go to state B;

if dnew ≥ dbest and k > 1, go to state C;

3) State A: Expand the search into K-1 sublayer, find

bounded uk and distance dk, and go to FSM.

4) State B: Record current best distance dbest = dnew and the

lattice point u ; set k = 2 find bounded uk and distance dk

and go to FSM.

5) State C:Exit if K = MT Otherwise, move the search into

K-1 sublayer, find bounded uk and distance dk , and go to

FSM.

Because there is no bound constraint in the SE algorithm, it

does not need to calculate and to update the bound in each

layer, thus the time-consuming square root operations are

avoided. Second, the chance of early identifying the correct

layer is maximized using the nondecreasing order of

investigation. Furthermore, the lattice point search starts from the Babai point, so no initial radius is required.

IV. EXPERIMENTAL RESULTS

The DSP implementation of MIMO sphere decoding

SE algorithm is discussed below.

A. Decoding Rate Performance

The data rate for the MIMO sphere decoder with M

transmit and N receive antennas is determined by

R =

stateavgc

fMb

dim

(11)

where is the system frequency of the circuits in megahertz,

bits per dimension bdim is determined by the modulation

scheme, and ηstate is the number of state visits required to

decode a receive vector at certain SNR. cavg is the average

number of clock cycles per state visit that is calculated as

cavg = stateistatei CP ,,( ) (12)

where Pi, state denotes the statistic percentage for visiting

state i, which can be obtained from high-level simulation.

Ci,state is the number of clock cycles used for state i

captured from the FPGA circuit simulation. A DSP

processor TMS320C6201 is chosen which requires 94

cycles, the decoding rate is around 3.66 Mbps. The number of state visits of the DSP implementation is significantly

larger than that of the FPGA implementation because DSP

does not support the state level parallelism. Similarly, the

DSP-based decoders cannot implement the current

execution of preprocessing and decoding It does not

support the parallel processing of real/image parts of

decoding algorithms either

A. B. BER Performance

The BER performance is evaluated using different

lattice generation matrices with Gaussian distribution and a

large number of source data at different SNRs with

additive white Gaussian noise (AWGN). The FPGA

implementations apply the proposed complex

transformation method and decode the real and imaginary

parts seperately. The software simulations directly conduct

the lattice search using complex numbers. The BER

performance of the FPGA implementations is then compared to the software simulation.

B. C. Power and Area Estimation

The power consumption of the TMS320C6201 DSP is

based on the typical values for high activity operations

(75% high and 25% low). The area of the DSP is larger

than the FPGA design. One obvious reason is the

difference in process technology. The most important

reason is that the area of FPGA is calculated based on the

number of slices mapped for a particular algorithm. But the

die size of the entire DSP is counted as the silicon area,

which is application- independent. In particular, some elements on the DSP device may not actively used for the

MIMO decoding algorithm, such as phase-locked loop

(PLL) and direct memory access (DMA) controller.

FPGA implementation differ from implementations on

DSPs through their potential for massively parallel

processing and the availability of customized operations

and operation sequences that can be executed in a single

cycle. The potential of an algorithm to exploit these

properties is crucial to guarantee an efficient high-

throughput implementation. The FPGA implementation of SE algorithm is expected to get a decoding rate above 50

Mbps and upto 80 Mbps which is about 20 times faster

than DSP implementation. The power consumption is

comparable for both FPGA and DSP implementation. The

area utilization in case of DSP implementation is about

95mm2 and for FPGA implementation we expect this to be

around 25-35mm2 which is very less compared to that of

DSP implementation.

The simulation result of the SE algorithm is as shown

in fig 2. First the pre processing unit of sphere decoding

algorithm is considered that is generation of channel matrix and then the best distance is calculated from the

received vector y and the examining lattice plane. The

algorithm proceeds following the steps in SE decoding

algorithm as explained above. Finally the closest lattice

point to the received vector y is calculated.

CONCLUSION

The system architecture and analysis of sphere

decoding SE algorithm for MIMO detection is presented

in this paper. The SE algorithm has lower circuit

complexity, smaller silicon area, and fewer lattice search

iterations.The MIMO sphere decoder which is of great

attraction towards the Wireless communication based

applications, its implementation in FPGA and

optimization of its architecture gains more interest in the

field of research. Power consumptions of the FPGA and DSP implementations are comparable, but the areas of

the FPGA designs are smaller. The throughput

advantage of the FPGA is due to its rich computational

resources and the parallel processing. This work can be

extended towards reducing the power and area utilization

and to achieve faster area throughput.

REFERENCES

[1]. B. M. Hochwald and S. Ten Brink, “Achieving near-capacity on a

multiple-antenna channel,” IEEE Trans. Commun., vol. 51, no. 3, pp.

389–399, Mar. 2003.

[2]. M. O. Damen, A. Chkeif, and J.-C. Belfiore, “Lattice code decoder

for space-time codes,” IEEE Comm. Let., pp. 161–163, May 2000.

[3]. D. Gesbert, M. Shafi, S. Da-shan, P. J. Smith, and A. Naguib,

“From theory to practice: An overview of mimo space-time coded

wireless systems,” IEEE J. Sel. Areas Commun., vol. 21, no. 3, pp.

281–302, Mar. 2003.

[4]. M. O. Damen, H. El Gamal, and G. Caire, “On maximum-

likelihood detection and the search for the closest lattice point,” IEEE

Trans. Inf.Theory, vol. 49, no. 10, pp. 2389–2402, Oct. 2003.

[5]. E. Viterbo and J. Boutros, “A universal lattice code decoder for

fading channels,” IEEE Trans. Inf. Theory, vol. 45, no. 5, pp. 1639–

1642, May 1999.

[6]. M.V. Clark, M. Shafi, W.K. Kennedy and L.J. Greenstein,

“Optimum Linear Diversity Receivers for Mobile Communications”,

IEEE Trans. Veh. Technol., Vol. 43, No. 1, pp. 47–56, 1994.

[7]. E. Agrell, T. Eriksson, A. Vardy, and K. Zeger, “Closest point

search in lattices,” IEEE Trans. Inf. Theory, vol. 48, no. 8, pp. 2201–

2214, Aug. 2002.

[8]. D. Garrett, L. Davis, S. ten Brink, B. Hochwald, and G. Knagge,

“Silicon complexity for maximum likelihood MIMO detection using

spherical decoding,” IEEE J. Solid-State Circuits, vol. 39, no. 9, pp.

1544–1552, Sep. 2004.

Fig 2: Simulation Result of SE Decoding Algorithm

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

257

Wireless Networks

Accident Response Server

“An Automatic Accident Notification System” NOORUL IRFANA.S

Department of Computer Science and

Engineering (3rd

year)

Anand Institute of Higher Technology

[email protected]

Abstract

“Technology does not drive change, it

Enables change.” The primary purpose of Technology is its

implementation in day-to-day life where in it

could enhance the lifestyle as well as provide better safety and performance to its end users.As

it may be understood, it is thus required to

provide systems that would cater to the unforeseen and unfortunate events that happen

in the modern day world. Providing timely and

accurate information is another key aspect in

reacting to situations of that would be catastrophic. With the number of vehicle users

in the country going manifold and with statistics

of accidents increasing at an alarming rate, systems should be devised that would call for

attention instantaneously reducing the severity

of any such calamities. It is sometimes referred

to as the golden first hour after any accident that takes place and is very precious in savings lives

of the victims. Another important factor to be

considered is accidents at highways are completely isolated from nearby hospitals,

rescue sites, etc. It is therefore required to

Provide an automated system that will help provide assistance in a minimal time. This

project, titled, Accident response server provides

an effective and reliable method that could be

used in highways to monitor and report location of accidents and take necessary action further.

The first and foremost requirement is to

accurately locate and pinpoint the accident so that help could be called for. For this purpose, a

database server is maintained online that stores

the list of area names with nearby help-centers contact numbers that should be alerted.

An electronic system would be installed in the

vehicle connected to the centralized system. The

RAMESH KRISHNAN.R

Department of Computer Science and

Engineering (3rd

year)

Anand Institute of Higher Technology

[email protected] electronic system has GPS and GSM to

communicate with the centralized database

system. When an accident occurs, the sensor

senses the vibration then sends a pulse to the system which is in vehicle side. Then the same

system passing the coordinates and vehicle

details to it. The centralized system then identifies the corresponding telephone number

and mail id and automatically sends the

screenshot of the accident to alert the area help-centers about the vehicle details , location and

notifies by call or SMS.

Keywords :- Accident detection, GPS based vehicle

Tracking, Vibration Sensors, GSM Modem,

Information Transfer to local help center.

1.Introduction

An accident is a disaster which is specific,

identifiable, unexpected, unusual and unintended external event which occurs in a particular time

and place, without apparent or deliberate cause

but with marked effects. It implies a generally negative probabilistic outcome which may have

been avoided or prevented had circumstances

leading up to the accident been recognized, and acted upon, prior to its occurrence. The first one

hour is the golden hour and that can make all the

difference. The aim is to reach out quickly to the

victims with help, upping the chances of their survival after an accident. Serious injuries can

result in disability, fatalities and life-long

psychological, emotional and economic damage to loved ones.Our system provides the

automatics response to such accidents. The

working of our project is divided in to the following sections:

1.1. GSM communication

GSM modem receives vehicle details and coordinate position of vehicle from GPS

258

module. It transmits messages to centralized

server for accident alerting. It is controlled by micro controller by interfacing with RS-232

1.2. GPS tracking

The GPS module calculates the geographical position of the vehicle. This helps in detecting

the location/position, velocity of our system.The

module output data like global positioning system fixed data, geographic position–

latitude/longitude are passed to GSM modem.

1.3.Microcontroller interfacing - RS- 232

It acts as an interface to GSM modem and micro

controller and to centralized server and GSM

Modem.

1.4 Vibration Sensing

The sensor used to sense the accident and sends as pulse to the microcontroller .The pulse is

transmitted as data to the GSM modem.

1.5. Accident notification

PC uses RS-232 protocol to forward call to

nearest help center using GSM Modem.

2. General Terminologies [1]Control Room: A control room is a room

serving as an operations centre where a facility

or service can be monitored and controlled. [2]Microcontroller: microcontroller (also MCU

or μC) is a functional computer system-on-

achip. It contains a processor core, memory, and

programmable input/output peripherals. [3]Google Earth: Google Earth displays satellite

images of varying resolution of the Earth's

surface, allowing users to visually see things like cities and houses from a bird's eye view.

[4]SMS: Short Message Service (SMS) is a

communications protocol allowing the interchange of short text messages between

mobile telephone devices.

[5]GSM: GSM (Global System for Mobile

communications: originally from Groupe Special Mobile) is the most popular standard for

mobile phones in the world.

. [6]GPS: Global Positioning System: a

navigational system involving satellites and

computers that can determine the latitude and longitude of a receiver.

[7]Vibration Sensor: It is a type of device

which can be activated by vibrations in a pre determined zone.

[8]AT Commands: AT commands are

instructions used to control a modem. AT is the

abbreviation of Attention.

3. Existing system

The following are the approaches followed

today, regarding accident reporting. They are, D Initially people used to make a call to the

help center. Anyone can save a life, but

people are not taking the initiative at the right time.

Accident reporting system with instant

messaging to default caller. This takes a

long time and it is of no use. Accident reporting system with default calling. This

system makes the situation worse and confusing.

4. Proposed system

Architecture of the Accident Response

Server

The architecture of the accident response server

is shown,

Fig 1 – The Architecture of the Accident

Response Server

Accident monitoring system with a database server maintained online that stores the list of

area names with nearby help-centers contact

numbers that should be alerted. The GPS- GSM

based accident notification system is such a state of the art automobile accessory. In case of any

accident, the accident notification subsystem

automatically makes a call to the police control room and also to some predefined telephone

numbers such as ambulance service; highway

security team etc.The whole system must be installed in the vehicle system with caution.

The main board of the system must be installed

in the middle position of the vehicle in a secure

location.

259

It provides us with the following advantages

It enables us to respond immediately when an accident occurs.

It provides the screenshot of the accident

captured and send to the help center.

The help centers are notified by call or SMS. It has an enhancement of blocking the

notification by engaging a button for minor

vibrations. We have our system designed with four

modules and the following components.

Module 1: GSM Communication

Module 2: GPS Tracking

Module 3: Vibration sensing

Module 4: Accident notification

4.1.Microcontroller The AT89C51 is a low-power, highperformance

CMOS 8-bit microcomputer with 4 Kbytes of

Flash Programmable and Erasable Read Only Memory (PEROM). The device is manufactured

using Atmel’s high density nonvolatile memory

technology and is compatible with the industry standard MCS- 51O instruction set and pinout.

The on-chip Flash allows the program memory

to be reprogrammed in-system or by a

conventional nonvolatile memory programmer. By combining a versatile 8-bit CPU with Flash

on a monolithic chip, the Atmel AT89C51 is a

powerful microcomputer which provides a highly flexible and cost effective solution to

many embedded control applications.The

AT89C51 provides the following standard

features: 4 Kbytes of Flash, 128 bytes of RAM, 32 I/O lines, two 16-bit timer/counters, a five

vector two-level interrupt architecture, a full

duplex serial port, on-chip oscillator and clock circuitry.

4.1.1 AT89C51 Instruction Set

ACALL:ACALL unconditionally calls a subroutine at the indicated code address.

ACALL pushes the address of the instruction

that follows ACALL onto the stack, least

significant- byte first, most-significant-byte second. The Program Counter is then updated so

that program execution continues at the

indicated address. The new value for the Program Counter is calculated by replacing the

least significant-byte of the Program Counter

with the second byte of the ACALL instruction, and replacing bits 0-2 of the most significant-

byte of the Program Counter with 3 bits that

indicate the page. Bits 3-7 of the most-

significant-byte of the Program Counter remain unchanged. Since only 11 bits of the

Program Counter are affected by ACALL, calls

may only be made to routines located within the

same 2k block as the first byte that follows ACALL.

CJNE:CJNE compares the value of operand1

and operand2 and branches to the indicated relative address if operand1 and operand2 are

not equal. If the two operands.

DJNZ:DJNZ decrements the value of register by 1. If the initial value of register is 0,

decrementing the value will cause it to reset to

255 (0xFF Hex). If the new value of register is

not 0 the program will branch to the address indicated by relative addr. If the new value of

register is 0 program flow continues with the

instruction following the DJNZ instruction the value "rolls over" from 0 to 255.

JNB:JNB will branch to the address indicated

by reladdress if the indicated bit is not set. If the bit is set program execution continues with the

instruction following the JNB instruction.

MOV:MOV copies the value of operand2 into

operand1. The value of operand2 is not affected. Both operand1 and operand2 must be in Internal

RAM. No flags are affected unless the

instruction is moving the value of a bit into the carry bit in which case the carry bit is affected or

unless the instruction is moving a value into the

PSW register (which contains all the program

flags. RET:RET is used to return from a subroutine

previously called by LCALL or ACALL.

Program execution continues at the address that is calculated by popping the topmost 2 bytes off

the stack. The mostsignificant- byte is popped

off the stack first, followed by the least-significant-byte.

SETB:Sets the specified bit.

4.2. GSM Modem

A GSM modem is a wireless modem that works with a GSM wireless network. A wireless

modem behaves like a dial-up modem. The main

difference between them is that a dial-up modem sends and receives data through a fixed

telephone line while a wireless modem sends

and receives data through radio waves. A GSM modem can be an external device or a

PC Card / PCMCIA Card. Typically, an external

260

GSM modem is connected to a computer

through a serial cable or a USB cable. A GSM modem in the form of a PC Card / PCMCIA

Card is designed for use with a laptop computer.

It should be inserted into one of the

PC Card / PCMCIA Card slots of a laptop computer. Like a GSM mobile phone, a GSM

modem requires a SIM card from a wireless

carrier in order to operate. Both GSM modems and dial-up modems support a common set of

standard AT commands. You can use a GSM

modem just like a dial-up modem. In addition to the standard AT commands, GSM modems

support an extended set of AT commands. These

extended AT commands are defined in the GSM

standards.

4.3.AT commands

AT commands are instructions used to control a

modem. AT is the abbreviation of ATtention. Every command line starts with "AT" or "at".

That's why modem commands are called AT

commands. Many of the commands that are used to control wired dialup modems, such as ATD

(Dial), ATA (Answer), ATH (Hook control) and

ATO (Return to online data state), are also

supported by GSM/GPRS modems and mobile phones. Besides this common AT command set,

GSM/GPRS modems and mobile phones

support an AT command set that is specific to the GSM technology, which includes SMS-

related commands like AT+CMGS (Send SMS

message), AT+CMSS (Send SMS message from

storage), AT+CMGL (List SMS messages) and AT+CMGR (Read SMS messages).Note that the

starting "AT" is the prefix that informs the

modem about the start of a command line. It is not part of the AT command name. For example,

D is the actual AT command name in ATD and

+CMGS is the actual AT command name in AT+CMGS. However, some books and web

sites use them interchangeably as the name of an

AT command.

4.4.GPS The global positioning system (GPS) was

developed by the U.S government for the

department of Defense. It is essentially a U.S military system, it offers navigation services to

civilians; however, at present, there is no law

which mandates the service to be made available for commercial application. Position fix is

obtained through passive receivers by the

triangulation method; where in estimated ranges from four satellites are used to derive the

position of a point. Ranges from three satellites

can provide the latitude and longitude of a point

on the earth; the addition of a fourth satellite can provide a user’s altitude and correct receiver

clock error. It is possible to derive the velocity

of the user and precise time information originating from onboard atomic clocks, which

have a drift rate of 1sec per 70,000 years

There are cesium atomic clocks aboard each satellite.

4.5.GPS Receivers

In 1980, only one commercial GPS receiver was

available on the market, at a price of several hundred thousand U.S. dollars. This, however,

has changed considerably as more than 500

different GPS receivers are available in today’s market.Commercial GPS receivers may be

divided into four types, according to their

receiving capabilities. These are: singlefrequency code receivers, single-frequency

carrier-smoothed code receivers,

singlefrequency

code and carrier receivers, and dualfrequency receivers. Single-frequency receivers access the

L1

frequency only, while dualfrequency receivers access

both the L1 and the L2 frequencies. GPS

receivers

can also be categorized according to their number of

tracking channels, which varies from 1 to 12

channels. A good GPS receiver would be multi channel, with each channel dedicated to

continuously

tracking a particular satellite. Presently, most GPS

receivers have 9 to 12 independent (or parallel)

channels. Features such as cost, ease of use,

power consumption, size and weight, internal and/or

external datastorage capabilities, interfacing

capabilities are to be considered when selecting a

GPS receiver.

4.6.GPS Antennas The antenna receives the GPS satellite signals

and

261

passes them to the receiver. The GPS signals are

spread spectrum signals in the 1575 MHz range and

do not penetrate conductive or opaque surfaces.

Therefore, the antenna must be located outdoors

with a clear view of the sky. The Lassen SQ GPS

receiver

requires an active antenna. The received GPS signals

are very low power, approximately -130 dBm, at

the surface of the earth. Trimble's active antennas

include

a preamplifier that filters and amplifies the GPS

signals before delivery to the receiver.

5.Working Of the system

5.1.Module 1: GSM Communication

GSM modem sends the message as “ACCIDENT OCCURRED” to the centralized

server using AT commands as AT commands

+CMGS can be used to send SMS messages from a computer / PC

AT+CMGS="91234567"<CR> Sending text

messages is easy.<Ctrl+z> GSM modem is

controlled by using microcontroller which is interfaced with RS- 232.RS-232 and

microcontroller is interfaced using MAX-232.

5.2.Module 2: GPS Tracking Each GPS satellite transmits radio signals that

enable the GPS receivers to calculate where its

(or your vehicles) location on the Earth and convert the calculations into geodetic latitude,

longitude and velocity. A receiver needs signals

from at least three GPS satellites to pinpoint

your vehicle’s position. GPS Receivers commonly used in most Vehicle tracking

systems can only receive data from GPS

Satellites. They cannot communicate back with

GPS or any other satellite. A system based on GPS can only calculate its location but cannot

send it to central control room. In order to do

this they normally use GSM-GPRS Cellular

networks connectivity using additional GSM modem/module. GPS Receiver can only Receive

data and cannot send data to Satellite (s). GPS

satellites do not know the position of a GPS Receiver. GPS Receiver calculate its position

using data from 3-4 satellites and no single

satellite know the calculations done by GPS receiver or its position. GPS Satellite service is

freely available throughout the world, anyone

anywhere can receive GPS data by buying any

off-the-shelf GPS receivers. GPS satellites are different satellites and can only send small and

week radio signals to earth.GPS Satellite signals

are weak and can be received normally with GPS antenna (external or integrated with GPS

receiver) facing open sky. GPS signals cannot be

received inside the home, building, garage, bridges. Even clouds and Trees can prevent GPS

signals from reaching GPS receiver. Hence no

GPS receiver can guarantee performance.

Certain advanced and sensitive GPS receivers can receive signals in above situations but

performance is still not satisfactory. GPS

receiver only gives Latitude, Longitude and Velocity calculated. To know your location you

need highly accurate map. If maps are not

accurate you will have much high error in

location. Normally accuracy of Maps is considered more important then accuracy of

GPS Receiver. GSM-GPRS is most suitable two

way communication system available to complement GPS receiver and send location

information to Mobile phone, Office or Internet.

GSM-GPRS Modem/Receiver can allow you to receive GPS position by SMS or Data.

5.3.Module 3: Vibration sensing

Vibration sensors are sensors for measuring,

displaying and analyzing linear velocity, displacement and proximity, or else acceleration.

They can be used on a stand-alone basis, or in

conjunction with a data acquisition system. Vibration sensors are available in many forms.

They can be raw sensing elements, packaged

transducers, or as a sensor system or instrument, incorporating features such as totalizing, local or

remote display and data recording. When the

262

vibration occurs it is sensed as pulse and send to

the micro-controller which is taken as the input to it.

5.4.Module 4: Accident notification

The control room maintains a centralized server.

The server maintains a database of names of cities all over the world and the rescue and the

help center number with the mail-id of those

centers.. MySql database is used to store the data. JDBC connectivity is used to retrieve the

data from the server and forward the call or sms

to the help centers. With the received vehicle details, the screen shot of vehicle location is

captured and sent to the help center from the

centralized server.

5.5. Reason for utilizing MySQL

5.5.1.MySQL is Cross-Platform

One great advantage of using MySQL is its

crossplatform capabilities. You can develop your database on a Windows laptop and deploy

on Windows Server 2003, a Linux server, a IBM

mainframe, or an Apple XServe, just to name a few potential platforms. This gives you a lot of

versatility when choosing server hardware. You

can even set up replication using a master on a

Windows platform with Linux slaves. It's incredibly easy to move between platforms: on

most platforms you can simply copy the data

and configuration files between servers and you are ready to go.

5.5.2.MySQL is Fast

An independent study by Ziff Davis found

MySQL to be one of the top performers in a group that included DB2, Oracle, ASE, and SQL

Server 2000. MySQL is used by a variety of

corporations that demand performance and stability including Yahoo!, Slashdot, Cisco, and

Sabre. MySQL can help achieve the highest

performance possible with your available hardware, helping to cut costs by increasing time

between server upgrades.

5.5.3.MySQL is Free

MySQL is Open Source software.As per its GPL license, you are free to redistribute those

changes as long as your software is also Open

Source. If you do not wish to make your software Open Source, you are free to do so as

long as you do not distribute your application

externally. If you adhere to the requirements of the GPL, MySQL is free for you to use at no

cost. If you wish to distribute your closed source

application externally, you will find that the cost

of a MySQL commercial license is extremely low (MySQL licenses start at only $249 US).

MySQL AB also offers well priced commercial

support that is significantly less expensive than

some of its counterparts.

6. Component Details (Hardware and

Software Requirements)

Table 1 – Table depicting the various hardware

and software requirements of Accident Response Server

7.Implementation – Approach and

Details

7.1.Module 1: GSM Communication

Microcontroller embedded with assembly

language program is interfaced with the GSM

modem using the RS-232. RS232 data is bi-polar.... +3 TO +12 volts indicates an "ON or 0-

state (SPACE) condition" while A -3 to -12 volts

indicates an "OFF" 1- state (MARK) condition.... Modern computer equipment

ignores the negative level and accepts a zero

voltage level as the "OFF" state. In fact, the "ON" state may be achieved with lesser positive

potential. This means circuits powered by 5

VDC are capable of driving RS232 circuits

directly, however, the overall range that the RS232 signal may be transmitted/ received

may be dramatically reduced. To make RS-232

TTL compatible, MAX-232 is used to interface Microcontroller and RS-232. The MAX232

device is a dual driver/receiver that includes a

263

capacitive voltage generator to supply EIA-232

voltage levels from a single 5- V supply. Each receiver converts EIA-232 inputs to 5-V

TTL/CMOS levels. These receivers have a

typical threshold of 1.3 V and a typical

hysteresis of 0.5 V, and can accept .30- V inputs. Each driver converts TTL/CMOS input levels

into EIA-232 levels. The driver, receiver, and

voltage-generator functions are available as cells in the Texas Instruments LinASIC. library.

RS-232 MAX-232 interfacing pin configuration

GSM Comminucation includes the following

Vehicle to Centralized server

• GSM modem receives vehicle details and coordinate position of vehicle through GPS

module.

• It transmits vehicle ID and location map to centralized server for accident alerting.

• It is controlled by micro controller by

interfacing with RS-232.

Centralized server to Help center

• GSM modem transmits the message and screen

shot to the help center.

7.2.Module 2:GPS Tracking GPS module tracks the position of vehicle using

GPS tracking unit installed in the vehicle and

sends the vehicle details to the GSM modem

Fig 8 – A Sample GPS Tracking unit

7.3.Module 3:Vibration Sensing The sensor placed in the vehicle senses the

accident and in response sends the pulse to the

microcontroller which acts as an input to it.

Vibration Sensor (diaphragm based

Condenser

Microphone)

A microphone, sometimes referred to as a mic

is an acoustic-to-electric transducer or sensor

that converts sound into an electrical signal.

Microphones are used in many applications such as telephones, tape recorders, hearing aids,

motion picture production, live and recorded

audio engineering, in radio and television

broadcasting and in computers for recording voice, VoIP, and for non-acoustic purposes such

as ultrasonic checking. The most common

design today uses a thin membrane which vibrates in response to sound pressure. This

movement is subsequently translated into an

electrical signal. Most microphones in use today for audio use electromagnetic induction

(dynamic microphone), capacitance change

(condenser microphone, pictured right), or

piezoelectric generation to produce the signal from mechanical vibration. A capacitor has

two plates with a voltage between them. In the

condenser mic, one of these plates is made of very light material and acts as the diaphragm.

The diaphragm vibrates when struck by sound

waves, changing the distance between the two

plates and therefore changing the capacitance. Specifically, when the plates are closer together,

capacitance increases and a charge current

occurs. When the plates are further apart, capacitance decreases and a discharge current

occurs.A voltage is required across the capacitor

for this to work.This voltage is supplied either by a battery in the mic or by external phantom

power.

264

7.4.Module 4:Call forwarding

The server placed in the control room stores the details of millions of help centers and the rescue

centers. JDBC connectivity is used to retrieve

the data and the server forwards the call to centers. The snapshot of the accident is taken

and it is mailed to the police station to get the

exact view of the accident.

Compilation of java file:

Snapshot:

Mail sending:

8. Conclusion

Thus the Accident Response Server would act as a benchmark rendering effective and quick

responses for reporting accidents. Also, it could

be fitted onto existing vehicles, requiring no

special setups. . Being Students of Technology we strongly feel that the Accident Response

Server would be a landmark of both

Technological as well as Social excellence .If our project could help to save precious human

lives by reporting accidents faster ,then the

success of our project would have been

achieved.

9. Future Research

In future, introducing a still faster responding

system would increase the efficiency. Till now GSM is the best known communication.GSM

could be replaced by some other technology. As

it is very costly. The sensors used in the system could be replaced by better detecting devices for

increasing the capability of sensing We are

working in this project to implement it as a best

real time project. Various steps are taken to use this system in forest areas where the possibility

of signal coverage is very less.

10. References

[1] Accident Prevention Stratergies : Causation

Model and Research directions Panagiotis Mitropoulos , Gregory A. Howell, and Tariq S.

Abdelhamid

265

[2] Accident notification system for vehicle,

Yamagishi, Junichi (Tokyo 111-0034, JP) [3]http://www.gsm-modem.de/gsmfaq.Html

[4]http://www.econsystems.com/gpscomm

module.asp

[5]http://www.java2s.com [6]http://www.developershome.com/sms/at

CommandsIntro.asp

[7] http://www.mysql.com/

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

266

SECURE DYNAMIC ROUTING IN NETWORKS

Marimuthu G,Saranya V,Thiripurasundari S & Akilandeeswari M

B.Tech, Information Technology,

Bannari Amman Institute of Technology, Sathyamangalam.

Email: [email protected] and [email protected]

Contact no: +91 9597100912

ABSTRACT

Security has become one of the

major issues for data communication over

wired and wireless networks. Different from

the past work on the designs of

cryptography algorithms and system

infrastructures, we propose a “Secure

Dynamic Routing Algorithm” that could

randomize delivery paths for data

transmission. It adopts the method of

multiple paths between source and

destination to increase the throughput of the

data transmission. It provides considerably a

minimum path similarity of two consecutive

transmitted packets. The algorithm is easy to

implement and compatible with popular

routing protocols. It could be used in

cryptography-based system designs to

further improve the data transmission over

networks. Thus it focuses on the security of

data transmission with very minimal

interception in multiple paths.

1. EXISTING SYSTEM

In the past decades, various security-

enhanced measures have been proposed to

improve the security of data transmission

over public networks. Existing work on

security-enhanced data transmission

includes the designs of cryptography

algorithms and system infrastructures and

security-enhanced routing methods. Their

common objectives are o ften to defeat

various threats over the Internet, including

eavesdropping, spoofing, session hijacking,

etc.

Among many well-known designs

for cryptography based systems, the IP

Security (IPSec) and the Secure Socket

Layer (SSL) are popularly supported and

implemented in many systems and

platforms. Although IPSec and SSL do

greatly improve the security level for data

transmission, they unavoidably introduce

substantial overheads especially on

gateway/host performance and effective

network bandwidth.

For example, the data transmission

overhead is 5 cycles/byte over an Intel

267

Pentium II with the Linux IP stack alone,

and the overhead increases to 58 cycles/byte

when Advanced Encryption Standard (AES)

is adopted for encryption/decryption for

IPSec .

1.1 Limitations of Existing System

Number of Retransmission is high

Cost is high to implement a

cryptographic technique

Eavesdropping is easily occur

2. PROPOSED SYSTEM

The security-enhanced data

transmission is to dynamically route packets

between each source and its destination so

that the chance for system break-in, due to

successful interception of consecutive

packets for a session, is slim. The intention

of security-enhanced routing is different

from the adopting of multiple paths between

a source and a destination to increase the

throughput of data transmission. The

proposed system is a secure routing protocol

to improve the security of end-to-end data

transmission based on multiple path

deliveries.

The set of multiple paths between

each source and its destination is determined

in an online fashion, and extra control

message exchanging is needed. A set of

paths is discovered for each source and its

destination based on message flooding.

Thus, a mass of control messages is needed.

The system proposed a traffic

dispersion scheme to reduce the probability

of eavesdropped information along the used

paths provided that the set of data delivery

paths is discovered in advance. Although

excellent research results have been

proposed for security-enhanced dynamic

routing, many of them rely on the discovery

of multiple paths either in an online or

offline fashion. For those online path

searching approaches, the discovery of

multiple paths involves a significant number

of control signals over the Internet.

On the other hand, the discovery of

paths in an offline fashion might not be

suitable to networks with a dynamic

changing configuration. Therefore, we will

propose a dynamic routing algorithm to

provide security enhanced data delivery

without introducing any extra control

messages. Let us discuss about the DDRA

algorithm which we used.

2.1 DDRA ALGORITHM

Consider the delivery of a packet

with the destination „t‟ at a node „N‟. In

order to minimize the probability that

packets are eavesdropped over a specific

link, a randomization process for packet

deliveries shown in Procedure 1 is adopted.

268

In this process, the previous next hop „B‟ for

the source node „S‟ is identified in the first

step of the process. Then, the process

randomly picks up a neighboring node in N

excluding B as the nexthop for the current

packet transmission. The exclusion of B for

the nexthop selection avoids transmitting

two consecutive packets in the same link,

and the randomized pickup prevents

attackers from easily predicting routing

paths for the coming transmitted packets.

Procedure 1:

Randomizedselector(S,t,pkt)

1: Let H be the used nexthop for the

previous packet delivery for

the source node S.

2: if H belongs to Candidate nodes(C) then

3: if |C|>1 then

4: Randomly choose a node x from |C-H| as

a nexthop, and send the packet pkt to the

node x.

5: H x, and update the routing table of N.

6: else

Send the packet pkt to H.

7: end if

8: else

Randomly choose a node y from C

as a nexthop,

and send the packet pkt to the node

y.

9: H y, and update the routing table of N.

10: end if

Before the current packet is sent to

its destination node, we must randomly pick

up a neighboring node excluding the used

node for the previous packet.

Once a neighboring node is selected, we

need to determine whether the selected

neighboring node for the current packet is

the same as the one used by the previous

packet. If the path is same as that of the

previous one, we should select the

alternative path.

3. MODULES

The following are the modules used ,

Network Structure Formation

Random Path Selection

Message Transmission

3.1Network Structure Formation

In this module, topology structure is

constructed using the existing nodes. The

nodes are connected to form a network. The

connection is made through sockets. Here

we use mesh topology because of its

unstructured nature.

269

The input is to be given as number of

nodes. And each node details such as IP

address, and port no are given. The link

information between the nodes obtained and

checked with existing network in the

database. If the link is already available, we

go for other links. Otherwise the link is

made between the nodes.

3.2 Random Path Selection

Random path algorithm is applied to

select a random path between the nodes

from the routing table. New route is found

using secure dynamic routing algorithm.

Routing history is updated in order to select

the path for message transmission.

The node details are given in the

login form. Then the destination is selected.

The nexthops are chosen from the database.

Once a neighboring node is selected,

we need to determine whether the selected

neighboring node for the current packet is

the same as the one used by the previous

packet. If the path is same as that of the

previous one, we should select the alternate

path.

3.3 Message Transmission

Random path is selected using the

routing table. Then the message is

Yes

No

Data Store

270

transmitted from source to destination. For

each transmission we choose a random path,

in that only the packet is transmitted.

From the above modules, the random

path for sending of information is selected

between the source and destination. As on

selected path the corresponding packets are

sent to the destination. The whole message

can be seen at the destination in a separate

window.

4. CONCLUSION AND FUTURE

WORK

This project has proposed a security-

enhanced dynamic routing algorithm based

on distributed routing information widely

supported in existing networks. A series of

simulation experiments were conducted to

show the capability of the proposed

algorithm. Thus we have achieved data

transmission through multiple paths which

ensures security by using dynamic routing.

Our future work is based on the

designs of cryptography algorithms and

system infrastructures with less overhead.

Our security enhanced dynamic routing

could be used with cryptography-based

system designs to further improve the

security of data transmission over networks.

It also involves with the implementation of

the algorithm in wireless networks.

5. REFERENCES

1. Chin-Fu Kuo, Ai-Chun Pang, and Sheng-

Kun Chan “Dynamic Routing with Security

Considerations,” 48 IEEE Transactions on

Parallel and Distributed Systems, vol. 20,

no. 1, january 2009

2. S. Bohacek, K. Obraczka, J. Lee, and C.

Lim,“Enhancing Security via Stochastic

Routing,” Proc. 11th Int‟l Conf. Computer

Comm. and Networks (ICCCN), 2002.

Data Store

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

271

Abstract – Considering the problem of keeping all types of

sensitive data and algorithms contained in a mobile agent from

discovery and exploitation by a malicious host. The paper

illustrates a novel distributed protocol for multi agent

environments to improve the communication security in packet-

switched networks. To enrich the overall system security the

approach makes use of distribution and varieties of available

encryption and some other traditional methods such as digital

signature. In this approach the encrypted private key and the

message are broken into different parts and is then encrypted

carrying by different agents which makes it difficult for

malicious entities to mine the private key for message

encryption, while the private key for the encrypted key is

allocated on the predetermined destination nodes. On the other

hand, all of the previously proposed encryption algorithms can

be applied in the proposed approach that deteriorates the key

discovery process. To improve the overall system security, the

paper makes use of Advanced Encryption Standard (AES) as

the encryption base for message encryption. The paper also

presents some evaluation discussions presenting time overhead

analysis and crack probability. And the destination address is

divided into two agents for carrying the network and host

identification. Thus the paper ensures the security policies not

only for the source and its information but the destination host

also.

Keywords- Mobile Agents; Distributed Cryptography;

Cryptographic Protocols; MA-MK (Multi Agent – Multi Key);

Multi Agent Environment; Double Cryptography Approach;

Internet Security; Communication Security;

I. INTRODUCTION

Because of dispersed number of system users, security is one of

the most challenging issues particularly in networked

environments. In traditional security methods, discovering the

determined private key is enough for message decryption that can

be done through malicious attacks to the network nodes or listening

to communication links.

1 K. Jeyakumar, is currently pursuing his Final year B.Tech - Information

Technology in Anna University Coimbatore. ([email protected]).

2

N.E.Vengatesh, is currently pursuing his Final year B.Tech - Information

Technology in Anna University Coimbatore. ([email protected]).

Since mobile agents have the features like self-learning and

mobility, we enable it as a carrier wave for the information

transmission. For a surprise, still there is no killer application for

mobile agents, even though we should consider the security issues

like masquerading, denial of service, unauthorized access,

eavesdropping, alteration and repudiation. Our proposed solution

avoids all kinds of security issues related to the mobile agents.

In traditional way, it is not possible to transfer the code and the

execution state. The most advantage of agent based communication

is that it can transfer the data, code and as well as the execution

state. The transfer of state together with data and code is based on

its mobility whether the strong mobility or weak mobility

respectively.

The paper is organized as follows:

Since we are using AES algorithm, the next section explains this

algorithm. Then we will talk about basic concepts of multi-key and

multi agent cryptography. The proposed multi agent-multi key

(MA-MK) protocol is then explained and the related simulation

results are presented. Finally, the conclusion section ends the paper.

II. ADVANCED ENCRYPTION STANDARD

According to malicious view, cryptography can be discussed

according to the following points:

How to access to the system.

How to recognize the encryption algorithm.

How to access to the private key.

Accessing to the private key can be considered as the end point of

a malicious process. The proposed approach provides the system

security, and makes the masquerader not to recognize the

encryption algorithm easily but not in the existing concepts and

improves private key security using the following strategies:

Encrypting the private key and data using an encryption

algorithm (AES algorithm is used in this paper).

Breaking the encrypted private key and data into different

units.

Encrypting the different parts carrying by the mobile

agents.

Breaking the destination address into two agents for the

identification of network and host.

K.Jeyakumar 1, N.E.Vengatesh

2

MA-MK Cryptographic Approach for Mobile Agent Security with

Mitigation of System and Information Oriented Attacks

272

AES encryption algorithm is a powerful algorithm introduced by

Daemen and Rijmen [1]. The AES algorithm runs four functions

namely: SubBytes, ShiftRows, MixColumns and AddRoundKey to

complete the encryption process. All of the different classes of AES

use these four functions. AES-128 runs this sequence for 10 times.

AES-192 runs for 12 and AES-256 runs the functions for 14 times.

Other details about AES encryption algorithm can be found in [2].

The following section illustrates a multi-agent model, which can be

considered as a main concept of this paper.

III. MULTIAGENTMODEL – BASICS

In this section the basic concepts of multi agent systems are

described according to multi agent entities and reputation. On a

network environment, a mobile agent is defined according to the

following concepts [3]:

The internal functionality of the mobile agent.

The state of the mobile agent.

The platform and its characteristics on which the agent

acts on.

The environment and its possibilities used by agents to be

transferred over the network.

IV. MULTI AGENT –MULTI KEY

CRYPTOGRAPHIC PROTOCOL

In multi agent systems malicious behaviors should be well

studied for better security provisions. A malicious attack against a

multi agent system can be considered as one of the following points

[4]:

1. Mobile agent security against other malicious mobile

agents.

2. Mobile agent security against malicious hosts.

3. Host security against malicious agents.

The proposed approach focuses on all the three concepts. Our

algorithm improves the overall security by splitting the encrypted

private key into several parts at the original node and then again

encrypted together with the data towards the destination. By

reassembling them at the destination node we can have the original

message. The n segregated parts are transferred towards the desired

destinations using n mobile agents plus 2 mobile agents for

transferring the destination side identification.

Splitting of the message and key is accomplished through a Split

function, which can be explained as:

ψi = Spliti (ψ) (1)

The above function is run many times according to the need for

splitting the original encrypted message and once for splitting the

private key (ψi corresponds to the ith part of the encrypted message).

The source node generates the required number of mobile agents

and equips each agent with the predetermined part of the message

and the key. It is supposed that the ith agent carries the ith part of the

message and the ith part of the encrypted key.

It should be noted that at each host four different states may be

determined for the agent. Fig. 1 describes these four states together

with the related relations. The following text, expresses completely

the relations mentioned in fig. 1:

1. If the previous agent has not accommodated on the host yet

the agent will kill itself. Such a host can be considered as

an unreliable host.

2. The previous agent is still working on the host. The

arriving agent should wait until complete action of the

previous agent.

3. The previous agent has left the host; the arriving agent can

run the internal function.

4. The running agent on the host might know the status of

succeeding agent.

Finally at the destination node a Join function acts conversely to

reassemble the encrypted message:

ψ = Join (ψi) (2)

It should be noted that the integration of the ψi is according to

their initial segregation orders. The key is also reassembles similar

to the message according to (1).

Figure 1: Four states of each agent on a network

host

The result of this process is the encrypted key that should be

decrypted using the initialized private key and obtained key. Finally

ψ decrypted by private key and destination host will obtain original

message.

The next section focuses on the evaluation of the proposed

algorithm regarding malicious hosts and we compare our protocol

with similar existing protocols.

V. EVALUATION

A. Comparing Protocol

273

About mobile agent security in multi agent environment, before

this, three protocols proposed: first ACCK that proposed by

Algesheimer, Cachin, Camenisch and Karjoth[5], second Xu

protocol[4], and third Multi-Agent Multi-Key protocol[6].

ACCK protocol requires the participation of a trusted third party

who does not collude with either the originator or any host.

Although this may be a reasonable assumption in some

circumstances, but our work improves upon the ACCK protocol by

eliminating this trusted third party. Xu protocol was used Yao

Encrypted circuits [7] for encrypting and transmitting the message,

but in this protocol if malicious host cheat a mobile agent then it

can discover the message. Though the Multi-Agent Multi-Key

protocol [6] overcomes these two protocol’s requirements it does

not concentrate on the alleviation of the system and information

oriented attacks.

But our proposed MA-MK Cryptographic protocol ensures the

mobile agent security together with the mitigation of system and

information oriented attacks because it is having the distributed

structure of mobile agent, the usage of more than two encryption

algorithms and the segregated sink identification.

B. Algorithm evaluation against malicious hosts

Malicious entities can be considered namely: malicious agents

and malicious hosts. First we analyze the time overhead of the

algorithm. It is evident that the cryptography process is similar to

the traditional algorithms since this process is accomplished only

once (before splitting and after reassembling). Therefore, we only

calculate the authentication processes that can be:

T = m + h -1 (3)

Where m is the number of mobile agents and h is the number of

destination hosts. Fig. 2 shows the diagram for relation (3)

according to number of mobile agents.

We also evaluated the proposed technique according to the

message crack probability. It is supposed that P is the crack

probability for one mobile agent. According to the proposed

approach, all the agents should be cracked to encrypt the message.

Therefore, for h mobile agents we have:

Pm = Ph (12) (4)

Where Pm is the probability for cracking the total message. Fig. 3

illustrates the above relation.

Fig 2: Time Overhead Analysis

10

15

20

25

30

35

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20

Number of Agents

Tim

e O

verh

ead

Un

it

Fig 2: Time Overhead Analysis

Fig 3: Crack and Access Probability Analysis

VI. CONCLUSION

According to the open nature of networked environments and

security challenge of such systems, we present a software-only, MA-

MK Multi Agent – Multi Key protocol for protecting mobile agents

against tampering by malicious hosts and we proposed a distributed

algorithm for cryptography to increase the overall security

including the mitigation of system and information oriented attacks.

On the other hand, the algorithm uses two private keys one for

message encryption and one for encrypting the first segregated

private key. The distributed nature that eliminates the single point

of failure and types of encryption prepares an appropriate

infrastructure for today critical areas.

REFERENCES

[1] Rosenthal Joachim, “A Polynomial Description of the Rijndael

Advanced Encryption Standard”, Journal of Algebra and Its

Applications, Vol. 2(2), 2003, pp. 223-236.

[2] National Institute of Standards and Technology, “Announcing

the ADVANCED ENCRYPTION STANDARD (AES),” Federal

Information Processing Standards Publication, no. 197, Nov. 2001.

[3] Dasgupta Dipankar and Brian Hal, “Mobile Security Agents for

Network Traffic Analysis”, in Proc. proceedings of DARPA

Information Survivability Conference, Anaheim California, June

2001.

[4] Xu Ke, “Mobile Agent Security Through Multi-Agent

Cryptographic Protocols”, PhD Thesis, Department of Computer

Science and Engineering, University of North Texas, May 2004.

[5] Joy Algesheimer, Christian Cachin, Jan Camenisch, and G Unter

Karjoth, “Cryptographic security for mobile code”, in Proc. IEEE

Symposium on Security and Privacy, May 2001, pp. 2-11.

[6] Abolfazl Esfandi, and Ali Movaghar Rahimabadi, “Mobile Agent

Security in Multi agent Environments Using a Multi agent-Multi

key Approach”, in Proc. IEEE, 2009, pp. 438-441.

[7] Andrew Chi-Chih Yao, "How to generate and exchange secrets",

in Proc. 27th IEEE Symposium on Foundations of Computer

Science (FOCS), 1986, pp. 162-167.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

274

NEW STEGANOGRAPHY METHOD FOR SECURE DATA

COMMUNICATION

T.Thambidurai

Demonstrator, P.G. Department of Software Engineering

M.Kumarasamy College of Engineering

Thalavapalayam,Karur-639113,Tamilnadu,India E-Mail: [email protected], Contact No:9444851747

K.Nagaraj

Student, II-M.Sc Software Engineering

P.G. Department of Software Engineering

M.Kumarasamy College of Engineering

Thalavapalayam,Karur-639113,Tamilnadu,India E-Mail: [email protected], Contact No:9688628175

1.Abstract :

Steganography is the art of passing

information in a manner that the very

existence of the message is unknown. The

goal of steganography is to avoid drawing

suspicion to the transmission of a hidden

message. If suspicion is raised, then this goal

is defeated. Steganalysis is the art of

discovering and rendering useless such

covert messages. In this paper,we identify

characteristics in current steganography

software that direct the steganalyst to the

existence of a hidden message and introduce

the ground work of a tool for automatically

detecting the existence of hidden messages in

images. Today sending information through

internet is very difficult and somebody can

hack the secret information and send secret

information to the receiver with some

modifications and receiver can receive and

understand the secret information wrongly,

so inorder to avoid this problem we

introduced a new method to send our secret

information through internet(send

information from sender to receiver).During

our research work we find out this new

method using steganography and we also

discussed about our new algorithm and we

explained our new method with example.

2.Introduction :

An overview of current steganography

software and methods applied to digital

images . Hiding information, where

electronic media are used as such carriers,

requires alterations of the media properties

which may introduce some form of

degradation. If applied to images that

degradation, at times, may be visible to the

275

human eye and point to signatures of the

steganographic methods and tools used.

These signatures may actually broadcast the

existence of the embedded message, thus

defeating the purpose of steganography,

which is hiding the existence of a message.

Two aspects of attacks on steganography are

detection and destruction of the embedded

message. Any image can be manipulated

with the intent of destroying some hidden

information whether an embedded message

exists or not.

3.Detecting Hidden Information :

Steganography tools typically hide relatively

large blocks of information where

watermarking tools place less information in

an image, but the watermark is distributed

redundantly throughout the entire image . In

any case, these methods insert information

and manipulate the images in ways as to

remain invisible to the human eye. However,

any manipulation to the image introduces

some amount of distortion and degradation

of some aspect in the "original" image's

properties. The tools vary in their approaches

for hiding information. Without knowing

which tool is used and which, if any,

stegokey is used, detecting the hidden

information may become quite complex.

However, some of the tools produce stego-

images with characteristics that act as

signatures for the steganography method or

tool used.

4.History & Importance

Steganography was first used by the

Greeks to send secret message. One of the

first documents mentioning steganography is

from Histories of Herodotus. The Greek

historian, Herodotus wrote how documents

with strategic information were transferred

during a battle. Text was written on tablets

that were then covered with wax to hide the

original message. The messenger could

transport the undetectable information

hidden on the tablets. Upon delivery the wax

would be melted and the message would

appear. Other stories documented by

Herodotus indicated it was a common

practice to shave the heads of slaves and

tattoo messages. After the hair had grown

back, the slave would sent on their way with

the message that was undetectable until their

hair was shaved off. This worked was long as

the non-intended recipient did not have the

key; which was the clue to shave the heads of

message carrier. Before the tragic event of

September 11,2001 USA Today reported that

terrorists were using the internet to transmit

hidden communications. During the recent

U.S Embassy bombing case in Africa,

several documents came to light in the media

that suggest Osama bin Laden and his

associates have been using steganography to

hide terrorist target plans inside pornography

and MP3 other files that are freely distributed

over the internet. These claims proved to be

false according to research done by Nives

proves, using the web crawler that

downloaded over two million images from

Ebay’s auction site, not a single message was

acquired. Even there has been a major shift

of focus toward identifying the use of this

technology because of the ability to conceal

strategic and demanding information on the

276

internet . Since September 11,the emphasis

on detecting hidden communications has

again become a hot interest area to law

enforcement and counter intelligence

agencies. They are interested in

understanding these technologies and their

weakness , so as to detect and trace allegedly

hidden messages in communications that the

al-Qaeda terrorist cell could use to continue

their terrorist attack on the united states. No

evidence was found so far but

acknowledgment was given in the public

media that this kind of information hiding

could be used. There are still frequent reports

in the new media that the same tactics are

being used. Such is the importance of

steganography in the today’s world. So we

have to develop tools, which could counter

the technologies used improperly.

5.Algorithm Based Stegnography

5.1 Proposed Algorithm For Encryption

Most Steganographic methods also encrypt

the message so even if the presence of the

message is detected, deciphering the message

will still be required. Steganography to

complementary to cryptography because it

adds an extra Maximum capacity = entropy

of PXb Xa:

H(X/X=Xa)=-PX/X( X/X)*log2P

X/X( X/X)layer of security. Therefore

both Steganography and cryptography were

used in this study. The proposed algorithm

explained step by step below:

1.Select a numeric key (This key may be

equal to MxN. where M and N denote the

total number of the pixels in the horizontal

and the vertical dimensions of the image)

2.Initialize the random permutation of the

integers algorithm (RIP) by using the

selected numeric key. The RIP (m:n)

corresponds to random permutations of

integers between m and n (m and n are both

integers)

3.Obtain a message scattering matrix by

using the RIP (1;MxN)

4.Replace the positions of the pixels of the

message by using the elements of RIP

(1;MxN) in order to obtain the scattered

message matrix (SM).

5.Run step-2 and Step-3 in order to obtain

pixel data scattering in binary bits matrix.

(PC), PC =RIP (K:8)

6.Take the ith pixel from the Cover Image

and convert its gray value to 8-bit binary (P)

Take that ith clement of PC corresponding

to the position of the bit which will be

replaced in P with the binary value of the

ith pixel of the SM.

7. Repeat step 6 for each pixel of the

scattered message.

5.2 Proposed Algorithm for the Decryption.

The proposed Algorithm for the Decryption

of the encrypted steganography algorithm is

given below step by step.

277

1. Select the numeric key used at

encryption phase.

2. Initialize the random permutation of

the integers algorithm (RIP) by using the

selected numeric key.

3. Obtain the message scattering

matrix by using the RIP (1;M x N).

4. Take the pixel data scattering in

binary bits matrix (PC), PC = RIP (k; 8).

5. Take the ith pixel from the

encrypted image and convert its gray value to 8 bit binary (P). Take the ith element

of PC corresponding to the position of the

bit which will be taken as the

6.ith pixel of the original binary converted

message (T) in P.

7.Replace the position of the ith pixel in T

with the ith element of PC.

The above algorithm is implemented

in MATLAB. The output of the

encrypting program is the PNG file

containing the embedded secret text. We

go in for PNG format with the view

of reducing the size of the image for

easy and efficient transmission. The

decryption program outputs the text file

filtering the image from text.

6.Implementation

The algorithm has been implemented in

MATLAB. Here is the procedure involved

in hiding a text file.

1) Give finger print to the system

2) Give finger print to the system(for

authentication &signature

3) Select the first source Image

4) Apply the above source image to hide

and seek method

5) Embed the First source image

6) Select the second source image

7) Apply the above source image to hide

and seek method

8) Embed the second source image

9) Give relation ship to person which you

want to send message

10) Embed these two source image into

one single source image.

11) connect your image with neural

network

12) Apply the above source image to hide

and seek method

13) Select the text file.

14) Set a password for the file.

15) The program outputs a PNG image with the hidden text.

The program outputs a PNG image Here is

the procedure for recovering the hidden

text files.

Here is the procedure for recovering the

hidden text files.

1)Select the PNG file to be recovered.

2) Give finger print to the system

3)Give finger print to the system(for

authentication &signature)

4)Select the first source Image

5)Apply the above source image to hide and

seek method

6) Find the Embedded the First source

image

7)Select the second source image

278

8)Apply the above source image to hide and

seek method

9)Find the Embedded the second source

image

10)connect your image with neural network

11)Give relation ship to person which the

want sender already send’s

message

12)Select the text file.

13)Set a password for the file.

14)The program outputs the hidden text

file in the same directory.

15)If the password is incorrect, the user

gets only an gets only an encrypted file,

which he can not understand.

For Example :

1. For example I have the following(this is

the password for the system)

2. Again give the same finger print to the

system

3) Select the first source Image

Imagesource:http://images.wikio.com/ima

ges/s/562/super-hind.jpg

4) Apply the above source image to hide and

seek method

5)Embed the first source image

6)Select the Second source image

Imagesource:https://newsline.llnl.gov/artic

les/2008/aug/images/super_earth_planet.jp

g&imgrefurl=https://newsline.llnl.gov/arti

cles/2008/aug/08.22.08_super.php&usg=__

nqaQ3NfYAencIcMdWqlnPq4vt9Y=&h=

342&w=528&sz=19&hl=en&start=13&um

=1&itbs=1&tbnid=ELEG7XXdR7slOM:

&tbnh=86&tbnw=132&prev=/images%3F

q%3Dsuper%2Bimages%26um%3D1%2

6hl%3Den%26rls%3DGAPB,GAPB:2005

-09,GAPB:en%26tbs%3Disch:1

279

7)Apply the above source image to hide and

seek method

8)Embed the Second source image

9)Give relation ship to person which you want to send message .For example if we are

three members in the system named

dines,dur, KrishIf I want to send send

message(i.e dur) to dinesh .just give the arrow mark to the persion (dinesh)which you

want to send.

10)Embed these two source image into one

single source image.

The following source image is taken as final

source image and it is used for processing .

11)connect your image with neural network

12)Apply the above source image to hide and

seek method

13)select text file:

D:du.doc

type the contents

Hi how are u.

14)set the password

15)The program outputs a PNG image with

the hidden text.

The reverse process can be done for receiver

side

7)Related Work

This paper provided an introduction to

steganalysis and identified weaknesses and

visible signs of steganography. This work is

but a fraction of the steganalysis approach.

To date a general detection technique as

applied to digital image steganography has

not been devised and methods beyond visual

analysis are being explored. Too many

images exist to be reviewed manually for

hidden messages. We have introduced some

weaknesses of steganographic software that

point to the possible existence of hidden

messages. Detection of these "signatures"

can be automated into tools for detecting

steganography .

8.Comments and Conclusion

Steganography transmits secrets through

apparently innocuous covers in an effort to

conceal the existence of a secret. Digital

image steganography and its derivatives are

XXXXXX

280

growing in use and application. In areas

where cryptography and strong encryption

are being outlawed, citizens are looking at

steganography to circumvent such policies

and pass messages covertly. Commercial

applications of steganography in the form of

digital watermarks and a stego-image which

seems innocent enough may, upon further

investigation, actually broadcast the

existence of embedded information digital

fingerprinting are currently being used to

track the copyright and ownership of

electronic media .This system will definitely

give more security and unbreakable method

9.References

[1].Anderson, R., : Information hiding: first

international workshop, Cambridge, UK.

Lecture Notes in Computer Science, Vol.

1174. Springer-Verlag, Berlin Heidelberg

New York (1996)

[2].Anderson, R., Petitcolas, F.: On the

Limits of Steganography, IEEE Journal on

Selected Areas in Communications, Vol. 16,

No. 4, May (1998) 474-481.

[3]. D.Artz. Digital Steganography; hiding

Data within Data, IEEE Internet Computing,

Vol.5

[4]. N.F. Johnson, S. Jajodia, Steganalysis :

The Investigation of Hidden Information

Proc. IEEE Information Technology Conf.

[5]. R.J. Anderson, F.A.P. Petit Colas, on

the Limits of Steganography. IEEE

Journal of Selected Areas in

Communications, Vol.16.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

281

AUDIO STEGANOGRAPHY BY REVERSE FORWARD

ALGORITHM USING WAVELET BASED FUSION

P.Arthi G.Kavitha

B.Tech-IT, III Year, B.Tech-IT, III Year,

[email protected] [email protected]

Abstract: Digital Audio

Steganography exploits the use of a host

audio to hide another audio of same size

into it in such a way that it becomes

invisible to a human observer. This

paper presents a Reverse Forward

algorithm based on the wavelet based

fusion technique. In this method the

wavelet decomposition of both the cover

audio and the secret audio are merged

into a single fused result using an

embedding factor containing the

concatenation of the dimensions of the

secret audio coefficient matrix.

Experimental results showed the high

invisibility of the proposed algorithm as

well as the large hiding capacity it

provides.

Keywords, steganography, cover audio,

stego audio, forward reverse scheme,

integer wavelets.

1. Introduction

The existence of the Internet has been

considered to be one of the major events

of the past years; information became

available on-line, all users who have a

computer can easily connect to the

Internet and search for the information

they want to find. This increasing

dependency on digital media has created

a strong need to create new techniques

for protecting these materials from

illegal usage. One of those techniques

that have been in practical use for a very

long time is Encryption. The basic

service that cryptography offers is the

ability of transmitting information

between persons in a way that prevents a

third party from reading it correctly.

Although, encryption protects content

during the transmission of the data from

the sender to receiver, after receipt and

282

subsequent decryption, the data is no

longer protected and is in the clear. That

is what makes steganography to

compliment encryption. Digital

Steganography exploits the use of a

host(container) data to hide or embed a

piece of information that is hidden

directly in media content, in such a way

that it is imperceptible to a human

observer, but easily detected by a

computer. The principal advantage of

this is that the content is inseparable

from the hidden message .

2. Earlier work on Steganography

The scientific study of steganography

began in 1983 when Simmons stated the

prisoner's

Problem. Some techniques are more

suited while dealing with small amounts

of data, while others to large amounts.

Some techniques are highly resistant to

geometric modifications, while others

are more resistant to non-geometric

modifications, e.g., filtering. Current

methods for the embedding of messages

fall into two main categories: High bit-

rate data hiding and low bit-rate data

hiding, where bit-rate means the amount

of data that can be embedded as a

portion of the size of the cover. High bit-

rate methods are usually designed to

have minimal impact upon the

perception of the host audio, but they do

not tend to be immune to audio

modifications. In return, there is an

expectation that relatively large amounts

of data can be encoded. All high bit-rate

methods can be made more robust

through the use of error-correction

coding, at the expense of data rate. So,

high bit-rate codes are only appropriate

where it is reasonable to expect that a

great deal of control will be maintained

over the audios. The most common and

simplest form of high bit-rate encoding

is the least significant bit insertion

(LSB). The advantages of the LSB

method include the ease of

implementation and high message

payload.

Other techniques include embedding the

message by modulating coefficients in a

transform domain, such as the Discrete-

Cosine Transform (DCT), Discrete

Fourier Transform, or Wavelet

Transform. The transformation can be

applied to the entire audio or to its

subparts. The embedding process is done

by modifying some coefficients that are

selected according to the type of

protection needed. If we want the

message to be imperceptible then the

283

high range of frequency spectrum is

Chosen, but if we want the message to

be robust then the low range of

frequency spectrum is selected. Usually,

the coefficients to be modified belong to

the medium range of frequency

spectrum, so that a tradeoff between

perceptual invisibility and robustness is

achieved.

3. Integer-to-Integer Wavelet

Transforms

The wavelet domain is growing up very

quickly. Wavelets have been effectively

utilized as a powerful tool in many

diverse fields, including approximation

theory; signal processing, physics,

astronomy, and image processing .A one

dimensional discrete wavelet transform

is a repeated filter bank algorithm . The

input is convolved with a high pass filter

and a low pass filter. The result of the

latter convolution is a smoothed version

of the input, while the high frequency

part is captured by the first convolution.

The reconstruction involves a

convolution with the syntheses filters

and the results of these convolutions are

added. In two dimensions, we first apply

one step of the one dimensional

transform to all rows. Then, we repeat

the same for all columns. In the next

step, we proceed with the coefficients

that result from a convolution in both

directions. These steps result in four

classes of coefficients: the (HH)

coefficients represent diagonal features

of the audio, whereas (HG and GH)

reflect vertical and horizontal

information. At the coarsest level, we

also keep low pass coefficients (LL). We

can do the same decomposition on the

LL quadrant up to log2(min (height,

width)).

Since the discrete wavelet transform

allows independent processing of the

resulting components without significant

perceptible interaction between them,

hence it is expected to make the process

of imperceptible embedding more

effective. However, the used wavelet

filters have floating point coefficients.

Thus, when the input data consist of

sequences of integers (as in the case for

images), the resulting filtered outputs no

longer consist of integers, which doesn't

allow perfect reconstruction of the

original audio. However, with the

introduction of Wavelet transforms that

map integers to integers we are able to

characterize the output completely with

integers.

One example of wavelet transforms that

284

map integers to integers is the S-

transform. Its smooth (s) and detail (d)

outputs for an index n are given. Note

that the smooth and the detail outputs are

the results of the application of the high-

pass and the low-pass filters

respectively. At the first sight it seems

that the rounding-off in this definition of

s(n) discards some information.

However, the sum and the difference of

two integers are either both odd or both

even. We can thus safely omit the last bit

of the sum since it equals to the last bit

of the difference. The S-transform is

thus reversible..

Note that the transforms are not

computed using integer arithmetic, since

the

Computations are still done with floating

point numbers. However, the result is

guaranteed to be integer due to the use of

the floor function and hence the

invertetibility is preserved.

4. The Proposed Algorithm

The proposed algorithm

employs the concept of wavelet based

fusion. It involves merging of the

wavelet decomposition of the

normalized versions of both the cover

audio and the secret audio into a single

fused result. In a normalized audio the

pixel components take on values that

span a range between 0.0 and 1.0 instead

of the integer range of [0,255]. Hence,

the corresponding wavelet coefficients

will also range between 0.0 and

1.0.Before the embedding process takes

place we need first to apply a pre-

processing step on the cover audio. This

step actually ensures that the embedded

message will be recovered correctly. The

extraction process involves subtracting

the original cover audio from the stego

audio in the wavelet domain to get the

coefficients of the secret audio. Then the

embedded audio. Then the embedded

audio is retrieved by applying IIWT.

4.1 The Embedding Module

The main idea of the proposed algorithm

is the reverse forward technique

employing wavelet based fusion. Fusion

or more specifically, data fusion refers to

the processing and synergistic

combination of information from various

knowledge sources and sensors to

provide better understanding of the

situation under consideration. As we can

see in figure1, the fusion process takes

place in the wavelet domain between the

IWT of the cover audio and the IWT of

the secret data. To be more specific we

mean by the “secret data” another audio.

285

That is, the proposed algorithm takes

advantage of the format similarity

between the cover and the secret.

However, we know that the ordinary

wavelet filters have floating point

coefficients i.e, when the input data

consist of sequences of integers, the

resulting filtered outputs no longer

consist of integers. We applied a

normalization operation on the cover

audio, i.e, to make the corresponding

wavelet coefficients to range between

0.0 and 1.0. This normalization

operation applies on the cover audio as

well as the secret audio. The step

concerning the wavelet-based fusion,

actually merges the wavelet

decomposition of both the cover audio

and secret audio into a single fused using

the following equation:

f’(x,y)= f(x,y) + α g(x,y)

where f’ is the modified IWT

coefficient, f is the original IWT

coefficient, g is the message coefficient,

and α is the embedding strength.

As you can notice, this operation may

cause the resultant embedded

coefficients to go out of the original

normalized range. That is, according to

the above equation, if it happens that the

value of f is 1, then the value of f’ will

go out of range depending on the sign of

the coefficient g and vice versa. Hence,

we first need to perform cover

adjustment before the embedding

process takes place. Then we apply the

2D IWT on the normalized versions of

the audio. We also propose a new idea

for the computation of the secret key

used to extract the stego audio. That is

instead of handling the key as an input

parameter, the key is derived from the

concatenation of the dimensions from

the secret audio coefficient matrices.

And hence in the extraction process the

key would be used instead of embedding

a header in the stego audio. Then the

technique of reverse forward

algorithm is employed to fuse the

normalized coefficients of both the cover

audio and the secret audio to obtain the

stego audio that need to be transmitted to

preserve the secret audio data. As per

this algorithm, the normalized

coefficients of the secret audio is merged

with that of the host audio in the reverse

order i.e. the normalized wavelet

transformed coefficients of the secret

audio in the reverse order is merged with

that of the host audio coefficients in the

forward (original) order so that the stego

audio thus obtained becomes highly

286

imperceptible for a casual observer thus

avoiding the attack or an attempt to

discover the presence of any secret data

or information.

4.2 The Extraction Module

As shown in figure 2(b), the extraction

process reverses the embedding

operation starting from applying the

IWT on the stego audio, then selecting

the embedded coefficients using the

reverse forward scheme as in the

embedding process.The next step

involves extracting the embedded

message coefficients from fused

normalized coefficients by subtracting

the original cover audio from the stego

audio in the wavelet domain.

Furthermore, the extracted normalized

coefficients are transformed into audio

format by applying IIWT. Obviously,

the algorithm discussed above is a cover

screw scheme since the extraction

process requires the cover audio to get

the message coefficients by subtracting

the original IWT coefficients from the

modified stego coefficients

.

5. Hiding capacity

In high bit-rate data hiding we have two

primary objectives: the technique should

provide the maximum possible payload

and the embedded data must be

imperceptible to the observer. We stress

on the fact that steganography is not

meant to be robust. Any modifications to

the file, such as conversions between file

types is expected to remove the hidden

data from the file. Fundamentally, data

payload of a steganographic scheme can

be defined as the amount of information

[3] it can hide within the cover media.

As with any method of storing data, this

can be expressed as a number of

wavelets, which indicates the max

message size that might be inserted into

287

an audio.

PAYLOAD PERCENTAGE = 100%

6. Experimental Results

Usually the invisibility of the hidden

message is measured in terms of the

Signal-to-Noise Ratio (SNR).

Fig: Experimental result of the proposed method

on audios.

(a) Cover audio entitled cover.wav.

(b) Secret audio entitled secret.wav (c) Stego

audio entitled stego.wav

7. Conclusions

This paper presents a reverse forward

steganographic scheme that is based on

the idea of merging wavelet

decomposition of both the cover audio

and the secret audio are merged into a

single fused result using an embedding

strength factor. Experimental results

showed that applying the idea of the

wavelet based fusion provides a better

performance than any existing method.

8. References

1) “High capacity image steganography

using wavelet based fusion ” by

M.Fahmy Tolba,M, Al-Said Ghonemy,

Ismail Abdoul-Hameed Taha, Ain

Shams university, Cairo ,Egypt.

2) Neil F.Johnson and Sushil Jajodia,

“Steganography:Seeing the unseen”,

IEEE computer, February 1998,pp 26-

34.

3) “Image adaptive steganography using

wavelets” by K.B.Raja, S.Akshatha,

Lalit patnaik, University Visvesvaraya

college of engg., bangalore.

4)

www.watermarkingworld.org/WMMLA

rchive/0509/msg00004.html - 22k

FEATURE SELECTION FOR MICROARRAY DATASETS USING SVM &

ANOVA

Janani.G1

A.Bharathi2

A.M.Natrajan3

PG Scholar1, Research Scholar

2, Supervisor

3

Bannari Amman Institute of Technology,

Sathyamangalam, Tamil Nadu

[email protected],

[email protected], [email protected]

3

Abstract: This project highlights the work in

making use of an accurate classifier and

feature selection approach for improved

cancer dataset classification. Developing an

accurate classifier for high dimensional

microarray datasets is a challenging task

due to availability of small sample size.

Therefore, it is important to determine a set

of relevant genes that classify the data well.

Traditionally, gene selection method often

selects the top ranked genes according to

their discriminatory power. Often these

genes will be correlated with each other

resulting in redundancy. In this work,

ANOVA with SVM has been proposed to

identify a set of relevant genes that classify

the data more accurately and along with

these two methods K-NN is used for filling

the missing values. The proposed method is

expected to provide better results in

comparison to the results found in the

literature in terms of both classification

accuracy and number of genes selected.

Keywords-- Cancer classification, Gene

Selection, Support Vector Machine , Genetic

Algorithm, Analysis of Variance.

I. INTRODUCTION

Data mining is defined as the

process of extracting information or

knowledge from large amount of data. It is

widely being used in a number of fields like

Bioinformatics, retail industry, finance,

telecommunication, etc. In recent year’s

bioinformatics is becoming one of the more

and more notable areas in research field

since it allows us to analyze data of an

organism in order to diagnose various

diseases like cancer. All human diseases are

accompanied by specific changes in gene

expression in the gene expression.

Bioinformatics community has generated

much interest in classification of patient

samples based on gene expression for

disease diagnosis and treatment. Cancer is

the second leading cause of death.

Classification is one of the data mining tasks

which allow arranging the data in a

predefined group. Classification of human

sample is a crucial aspect for the diagnosis

and treatment of cancer. Cancer

classification plays an important role in

cancer treatment.

From the classification

point of view it is well known that when the

number of samples is much smaller than the

Proceedings of the Third National Conference on RTICT 2010Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

288

number of features, classification methods

may lead to over fitting. Moreover high

dimensional data requires inevitably large

processing time. So for analyzing

microarray data, it is necessary to reduce the

data dimensionality by selecting a subset of

genes (features) that are relevant for

classification. In order to deal with these

challenges there is necessity to select the top

most genes for better classification accuracy.

The main challenges in using microarray

data are

1) Overfitting, which occurs when the

number of samples is much smaller than the

number of features;

2) Redundant data, since microarray data

are generally highly multidimensional that

leads to noisy data.

Alon[2] et al used two

way clustering for analyzing the analyzing

a data set consisting of the expression

patterns of different cell types identify

families of genes and tissues based on

expression patterns in the data set. Chris

et al [3] proposed a minimum

redundancy – maximum relevance

(MRMR) feature selection framework.

Genes selected via MRMR provide a

more balanced coverage of the space and

capture broader characteristics of

phenotypes. They lead to significantly

improved class predictions in extensive

experiments on 5 gene expression data

sets. Edmundo [6] made use of hybrid

approach of SVM and GA with LOOCV

validation for better accuracy but the

computation time taken was high and it did

not give small set of genes. Jaeger [8] used a

gene informative technique in which the top

k genes are selected using a t statistics

method the problem here is correlation was

high. Jin [9] proposed SGANN which is a

speciated GA with ANN, this system used a

small sized set of gene and obtained better

performance. Mohd [12] used a hybrid

approach of GA+SVM I and GA+SVM II it

yielded better accuracy but did not avoid

over fitting.

In this proposed system, the

number of genes selected, M is set by

human intuition with trial-and-error. There

are also studies on setting M based on

certain assumption on data distributions.

These M genes may be correlated among

themselves, which may lead to redundancy

in the feature set. Also certain genes may be

noisy which may decrease classification

accuracy. So this set of selected genes is

further reduced with the help of genetic

algorithm combined with SVM. ANOVA is

proposed which selects the smallest

important set of genes that provides

maximum accuracy. The proposed method is

experimentally assessed on two well known

datasets (Lymphoma and SRBCT).

II.METHODOLOGY

Each micro array is a

silicon chip on which gene probes are

aligned in a grid pattern. One microarray can

be used to measure all the human genes.

However, the amount of data in each

microarray is too overwhelming for manual

analysis, since a single sample contains

measurement for around 10,000 genes. Due

to this excessive amount of information,

efficiently producing result is done. By

using machine learning techniques the

computer can be trained to recognize

patterns that classify the microarray. One of

the goals of microarray data analysis is

cancer classification. Cancer classification

plays an important role in cancer treatment

and drug discovery. Cancer Classification,

given a new tissue, predict if it is healthy or

not; or categorize it into a correct class.

There are two tasks in cancer classification.

They are class prediction and class

discovery. Class discovery refers to the

process of dividing samples into groups with

similar behavior or properties. Class

prediction refers to the assignment of

particular tumor sample to already-defined

classes, which could reflect current states or

future outcomes. Given a set of known

classes, determine the correct class for a new

patient.

289

In order to deal with

microarray data set which are highly

redundant and to classify the data sets more

accurately there are two steps to be

followed. They are to select the top most

genes from entire gene subsets and then

apply the classifier to selected set of top

most genes and finally obtain the selected

set of genes.

In the proposed system the

entire dataset is taken to that the gene

ranking method is applied. Entropy based

gene ranking is used from which the

Topmost M genes are obtained. The number

of genes selected, M, is set by human

intuition with trial and error. These M genes

maybe correlated among themselves which

may lead to redundancy in the feature set.

Also certain genes maybe noisy which may

lead to decreased classification accuracy. So

this set of gene is further reduced with help

of ANOVA with SVM.

SVM was originally designed

for binary classification. Since SVM is a

binary classifier it cannot be effectively used

for more than two classes in order to solve

this problem the Multi-class SVM is used.

The most common way to build a Multi-

class SVM is by constructing and combining

several binary classifiers. The representative

ensemble schemes are One-Against-All and

One-Versus-One. The outputs of the

classifier are aggregated to make a final

decision. Error correcting output code is a

general representation of One-Against-All or

One-Versus-One formulation, which uses

error-correcting codes for encoding outputs.

Radial basis function (RBF)

and Cross-validation is a way to predict the

fit of a model to a hypothetical validation set

when an explicit validation set is not

available. The five fold cross validation is

used for calculating the classification

accuracy. The 5 fold cross validation was

carried out in the training data set. The CV

was used for all of the data sets and selects

the smallest CV error.

Figure 1: Gene Selection in Proposed

System

ANOVA is a powerful

statistical technique that is used to compare

the means of more than two groups. One

way ANOVA is a part of the ANOVA

family. When we are comparing the means

of more than two populations based on a

single treatment factor, then it said to be one

way ANOVA. The equation used for one

way ANOVA is as follows: yij = m + ai + eij,

where this equation indicates that the jth

data value, from level i, is the sum of three

components: the common value (grand

mean), the level effect (the deviation of each

level mean from the grand mean), and the

residual.

III EXPERIMENTAL RESULTS

The topmost genes were

selected using the ranking method called

Entropy based ranking scheme. SVM

classifier is also implemented using

Entire gene

set

Gene

Ranking

Method

Topmost M

genes

ANOVA

with SVM

Selected set of

genes

290

MATLAB 7.4. Kernel chosen here is RBF

kernel (K (x→,

y→)

= exp(-γ||x→

-y→

|| 2) in the

multi-class SVM classifier. The data sets

that were taken are Lymphoma, SRBCT.

The number of genes used for lymphoma

dataset is 4026 and he number of samples

used are 62 with 3 classes that is 42 samples

from DLBCL, 9 samples from FL and 11

samples from CLL. For that of SRBCT

dataset the number of genes used are 2308

and number of samples used are 83 with 4

classes NB, RMS, NHL and EWS. ANOVA

is to be implemented for feature selection. In

the experiments the original partition of the

datasets into training and test sets is used

whenever information about the data split is

available. In the absence of separate test set,

cross validation is used for calculating the

classification accuracy.

IV CONCLUSION

In the proposed system one

way ANOVA in conjunction with the SVM

with entropy gene ranking method produced

92.75% accuracy for lymphoma dataset and

99.32% accuracy for SRBCT dataset with 5

fold cross validation. ANOVA and CV are

highly effective ranking schemes, along with

these SVM is used which is a good

classifier.

5 13 21 29 37 450

10

20

30

40

50

60

70

80

90

100

No of Features

Accura

cy

All parameters

Genetic + SVM

Anova + SVM

Figure 1: Classification accuracy - SRBCT

5 13 21 29 37 450

10

20

30

40

50

60

70

80

90

100

No of Features

Accura

cy

All parameters

Genetic + SVM

Anova + SVM

Figure2:Classification accuracy- Lymphoma

REFERENCE

[1] R. K. Agrawal, and Rajni Bala (2007) ‘A

Hybrid Approach for Selection of Relevant

Features for Microarray Datasets’ World

Academy of Science, Engineering and

Technology 29 2007. pp. 281-288.

[2] Alon, U. (1999) ‘Broad patterns of

gene expression revealed by clustering

analysis of tumor and normal colon tissues

probed by oligonucleotide arrays’

Proceedings. of National. Academy of

Science. Vol. 96, pp. 6745–6750.

[4] Chris, D. and Hanchuan, P. (2004)

‘Minimum redundancy feature selection

from microarray gene expression data’ IEEE

Computer Society Bioinformatics Conf

pp.1-8.

[5] Donna, K S. and Tamayo, P (2000)

‘Class prediction and discovery using gene

expression data’ ACM pp 263-271.

[6] Edmundo, B H. (2006) ‘A hybrid

GA/SVM approach for gene selection and

classification of microarray data’

EvoWorkshops 2006, LNCS 3907, pp. 34–

291

44, 2006. Springer-Verlag Berlin

Heidelberg.

[7] Golub, T R. and Slonim, D K. (1999),

‘Molecular classification of cancer: class

discovery and class prediction by gene

expression monitoring’ SCIENCE VOL 286

15 pp. 531-537.

[8] J. Jaeger, J. and Sengupta, P. (2003)

‘Improved gene selection for classification

of microarrays’ Proceedings of Pacific

Symposium on Biocomputing, vol 8, pp53-

64.

[9] Jin, H H. Sung, B C. (2005) ’Efficient

huge-scale feature selection with speciated

genetic algorithm’ Science direct Pattern

Recognition Letters 143–150.

[10] Lei, X. (2004) ‘Is cross-validation valid

for small-sample microarray classification?’

Topics in Bioinformatics.

[11] Momiao, X. and Wuju, Li. (2001)

‘Feature (Gene) selection in gene

expression-based tumor classification’

Molecular Genetics and Metabolism 73, pp

239–247.

[12] Mohd, S M. and Safaai, D. (2008) ‘An

approach using hybrid methods to select

informative genes from microarray data for

cancer classification’ Second Asia

International Conference on Modelling &

Simulation, IEEE Computer Society, pp

603-608.

[13] Mohammed, L A. (2005) ’Feature

selection of DNA microarray data’

Proceedings of the tenth ACM SIGKDD

international conference on Knowledge

discovery and data mining, pp.737 – 742.

[14] Sridhar, R. and Tamayo, P (2001)

’Multiclass cancer diagnosis using tumor

gene expression signatures’ PNAS, vol. 98,

no. 26, pp 15149–15154.

292

9-10 April 2010

Digital Fragile Watermarking using Integer Wavelet Transform V. Kavitha, C. Palanisamy,and Amitabh Wahi

Department of Information Technology

Bannari Amman Institute of Technology, Sathyamangalam- 638 401

[email protected], [email protected] and [email protected]

ABSTRACT

A new method for speeding up the integer wavelet transforms constructed by the lifting scheme

is proposed. This work describes the fragile watermarking technique using wavelets. In our

approach we change a few wavelet coefficients to achieve fragile watermarking. This lifting

algorithm hides data into one or more middle bit-plane(s) of the integer wavelet transform

coefficients in the LH, HL and HH frequency sub bands. It can embed more data into the bit

planes and also has the necessary imperceptibility requirement. As a result, our method can save

the decomposition/reconstruction time and experiments are carried out using Mat lab.

Keywords: Watermarking, Wavelets, lifting, LH, HL, HH.

1. INTRODUCTION

Recently, some distortion less

watermarking techniques have reported

in the literature. The first method [1] is

carried out in the image spatial domain.

With the focus on the space domain,

several fragile watermarking methods

that utilize the least significant bit (LSB)

of image data were developed. For

example, a technique that inserts a

checksum determined by the seven most

significant bits into the LSBs of selected

pixels was proposed in [2].

Authentication techniques are required

in applications where verification of

integrity and authenticity of an image is

essential [3], [4]. A watermark is said to

be fragile if the watermark hidden within

the host signal is destroyed as soon as

the watermarked signal undergoes any

manipulation. When a fragile watermark

is present in a signal, we can infer, with

a high probability, that the signal has not

been altered. The attacker is no longer

interested in making the watermarks

unreadable.

The lifting scheme is new method for

constructing integer wavelet transform

[5]. Recently, biorthogonal wavelets

constructed by the lifting scheme have

been identified as very promising filters

for lossless/lossy image compression

applications [6]. By making use of

similarities between the high- and low-

pass filters, the lifting scheme reduces

the computation complexity by a factor

of two compared with traditional

wavelet transform algorithms. Actually,

disturbing this type of watermark is easy

because of its fragility. So the

researchers developed a hybrid

authentication watermarking consisting

of a fragile and a robust watermark [7].

The goal of the attackers is, conversely,

producing a fake but legally

watermarked signal. This host media

forgery can be reached by either making

undetectable modifications on the

watermarked signal or inserting a fake

watermark into a desirable signal.

In our paper we used the integer wavelet

based fragile watermarking. The main

aim of our approach is to reduce

distortion in the watermarked image and

can split the very small watermarked

block size can embed the information.

The integer wavelet based fragile

Proceedings of the Third National Conference on RTICT 2010Bannari Amman Insitute of Technology , Sathyamangalam 638 401

293

watermarking also improves the

robustness of the message. Fragile

watermarking authentication has an

interesting variety of functionalities

including tampering localization and

discrimination between malicious and

non-malicious manipulations.

The rest of the paper is organized as

follows. Section 2 discusses the exiting

Data hiding approaches. The proposed

method is presented in section 3.

Experiments and results are presented in

Section 4. Section 5 concludes the paper.

2. EXISTING APPROACH

When we transform an image block

consisting of integer-valued pixels into

wavelet domain using a floating point

wavelet transform and the values of the

wavelet coefficients are changed during

watermark embedding, the

corresponding watermarked image block

will not have integer values. The original

image cannot be reconstructed from the

watermarked image.

The horizontal and vertical and detailed

bands are to embed the watermark data.

These bands are separated out and used

to embed the watermark data. The

watermark data is embedded directly in

bytes as the difference between the

neighboring pixels pairs in [8].

The expandable pixel pairs are only

selected for embedding. If a pair is not

suitable then the next two pixels are

checked and so on till all bytes are

embedded. But still this method is

having the problem of overflow.

3. PROPOSED APPROACH

Lifting scheme is an effective

method to improve the processing speed

of DWT. Integer wavelet transform

allows to construct lossless wavelet

transform which is important for fragile

watermarking. By lifting scheme, we can

construct integer wavelet transform.

Wavelet transform is a time domain

localized analysis method with the

window’s size fixed and forms

convertible. Also there is quite good

frequency differentiated rate in its low

frequency part. It can distill the

information from signal effectively. The

basic idea of discrete wavelet transform

(DWT) in image process is to multi-

differentiated decompose the image into

sub-image of different spatial domain

and independent frequency component.

Then transform the coefficient of sub-

image.

A two-dimensional image after three-

times DWT decomposed can be shown

as Fig.1. Here L represents low-pass

filter, H represents high-pass filter. An

original image can be decomposed of

frequency components of HL1, LH1,

and HH1. The low-frequency component

information also can be decomposed into

sub-level frequency component

information of LL2, HL2, LH2 and

HH2. By doing this the original image

can be decomposed for n level wavelet

transformation.

The lossless watermarking processed in

the wavelet domain is attracted. The

high compression rate obtained by de-

correlation in wavelet domain is for

embedding high-capacity data. The

watermarked image with suitable

embedding data obtained by multi-

resolution is for imperceptible in vision.

Origina

l Image

Original

Image 294

Fig.1 Embedding process

Fig.2 Extraction process

Watermarked

image Integer wavelet

decomposition

Recovering

Algorithm Key image

Inverse

Integer wavelet

decomposition

Recovered

Original

image

Pre

processing

LWT

Integer wavelet

decomposition

Watermark

Embedding

algorithm Inverse

integer

wavelet

transform

Key image

Watermarked

Image

295

Steps in Embedding

1. Let R= P1 , P2 ,......Pn represent the set of all

neighboring pixel pairs in one

direction and Pi be a single pair

for all i=1 to n.

2. let W = W1 ,W2 ,......Wm represent the watermark data

where each 0,1,2,.....,9

Wi are used to represent ASCII

characters. The same digits can

represent Text or image

watermark.

In this above steps are indicating the

embedding process for watermarking

image.

P is represented as each pixel for

original image. Where W is represented

as each pixel for key image. The

information of low frequency component

is an image close to the original image.

Most signal information of Original

image is in this frequency component.

The frequency components of LH, HL

and HH respectively, represent the level

detail, the upright detail and the diagonal

detail of the original image. For testing

the performance of this algorithm, the

experiments are simulated with the

MATLAB. In the following

experiments, the gray-level image with

size of 256 * 256 “Lena” is used as host

image to embed watermark.

3. EXPERIMENTS AND RESULTS

4.1 DWT Based Watermarking

Fig.3 Before watermarking Lena

image

Watermarked Image

Fig.4 DWT based watermarked

image

Fig.5 DWT based recovered image

Here the standard Lena image is used to

hide the information. When the

embedding message is used in the

transformed domain, the robustness is

high and the distortion is very low.

Compare with other methods,

embedding technique using wavelets

performs better. This is because the

resolution is very high and compared the

Recovered Watermark

296

image distortion is very low. Here the

key is represented as the text which to be

embedded with the image. The Lena

image before and after watermarking are

as shown in figures 4 and 5 respectively.

4.2 IWT Based Watermarking

Fig.6 IWT based watermarked image

Fig.7 IWT based recovered image

Table 1: PSNR values of the embedded

watermarked image

The above Table1 shows the different

types of transforms embedding and it

shows the peak signal to noise ratio

(PSNR) and that will take the time for

embedding an image. So the above

analysis, final result is achieved by

integer wavelet transform by using

lifting scheme. In addition, the security

can be enhanced by incorporating a

private embedding key to code the map

associating the embedded bits to the

selected wavelet watermarking blocks.

By using the integer wavelet

transformation it is avoid the overflow of

the data compared with the existing

method. The result shows the distortion

less watermarked image. The results are

presented in Table.1 Indicates the types

of transformation like LSB, DCT, DWT,

and IWT this compares the results of

peak signal to noise ratio value.

As a result finally we got good PSNR

value in integer wavelet transform. To

verify the advantage of the packed

integer wavelet transform, we compared

its performance with that of the original

integer wavelet transform for image

compression. It produces many small

high-frequency coefficients by the high-

pass filter. On the other hand, the minor

errors are generated by the computation.

In the integer computation, the

magnitude does not even increase

Consequently, overflow of the

magnitude is not a concern in our

applications. In general, multiple

coefficients can be packed into one

integer provided that the width of the

data-path is sufficient.

Transforms PSNR value in

db

Elapsed

time in (s)

LSB 1.2212e+005 10.5625

DCT 2.4061e+003 11.6563

DWT 3.1402e+003 10.6563

LWT 33.153037 35.5625

Retrieved Watermark

Watermarked Image

297

5. CONCLUSION

In this paper, we analyze the security of

fragile image authentication schemes

based on watermarking that can

accurately localize modifications. We

start with a comparison of PSNR Values

and Elapsed time using DWT based

algorithm and proposed algorithm. The

experiments are carried out using

Matlab. In Future work we are going to

use the moments for rotation and

cropping for watermarked image.

REFERENCES

1. C. W. Honsinger, P. Jones, M.

Rabbani, and J. C. Stoffel,

“Lossless recovery of an original

image containing embedded

data,” US Patent application,

Docket No: 77102/E−D (1999).

2. S. Walton, “Information

authentication for a slippery new

age,” Dr.Dobbs J., vol. 20, no. 4,

pp. 18–26, Apr. 1995.

3. B. B. Zhu, M. D. Swanson, and

A. H. Tewfik, “When seeing isn’t

believing,” IEEE Signal Proces.

Mag., vol. 21, no. 2, pp. 40–49,

Mar. 2004.

4. M. U. Celik, G. Sharma, E.

Saber, and A. M. Tekalp, “A

hierarchical image authentication

watermark with improved

localization and security,” in

Proc. IEEE Int. Conf. Image

Processing, Oct. 2001, vol. 2, pp.

502–505.

5. W. Sweldens, “The lifting

scheme: a custom-design

construction of biorthogonal

wavelets,” Appl. Comput.

Harmon. Anal., vol. 3, no. 2, pp.

186–200, 1996.

6. I. Daubechies and W. Sweldens,

“Factoring Wavelet Transforms

into Lifting Steps,” Tech. Rep.,

Lucent Technologies, Bell Labs.,

Holmdel, NJ, 1996.

7. E. Koch and J. Zhao, “Towards

robust and hidden image

copyright labeling,” in Proc.

IEEE Workshop on Nonlinear

Signal and Image Processing,

Jun. 1995, pp. 452–455.

8. S. Kurshid Jinna1, Dr. L.

Ganesan , “Lossless Image

Watermarking using Lifting

Wavelet Transform” ,

International Journal of Recent

Trends in Engineering, Vol 2,

No. 1, November 2009.

298

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

299

Testing Polymorphism in Object Oriented Systems

at Dynamic Phases for Improving Software Quality

R.VinodKumar, G.Annapoorani

Department of Computer Science Engineering

Anna University Tiruchirappalli,Tamilnadu, India [email protected]

[email protected]

Abstract— With object-oriented testing should centre on object

class, generic classes and super classes, inheritance and

polymorphism, instead of subprograms. In object-oriented

software test, the feature of inheritance and polymorphism

produces the new obstacle during dynamic testing.

Polymorphism becomes a problem in dealing with object-

oriented software testing. Although program faults are widely

studied, there are many aspects of faults that still do not

understand, particularly about object-oriented software. The

important goal during testing is to cause failures and there by

detect faults, a full understanding of the characteristics of faults

is crucial to several research areas. The power of inheritance and

polymorphism brings to the expressiveness of programming

languages also brings a number of new anomalies and fault

types. In this application various issues and problems that are

associated with testing polymorphic behaviour of objects in

object oriented systems is discussed. The application of these

techniques can result in an increased ability to find faults and to

create overall higher quality software.

Keywords—Polymorphism, Inheritance, Software Testing,

Software Quality, Method Overriding, Metrics, Overloading and Object Orientated Systems.

I. INTRODUCTION

Testing of object-oriented software presents a variety of

new challenges due to features such as inheritance,

polymorphism, dynamic binding and object state. Programs

contain complex interactions among sets of collaborating

objects from different classes. These interactions are greatly complicated by object-oriented features such as polymorphism,

which allows the binding of an object reference to objects of

different classes. While this is a powerful mechanism for

producing compact and extensible code, it creates numerous

fault opportunities.

Polymorphism is common in object-oriented software for

example; polymorphic bindings are often used instead of case

statements. However, code that uses polymorphism can be

hard to understand and therefore fault-prone. For example,

understanding all possible interactions between a message

sender object and a message receiver object under all possible

bindings for these objects can be challenging for programmers.

The sender object of a message may fail to meet all

preconditions for all possible bindings of the receiver object.

A subclass in an inheritance hierarchy may violate the

contract of its super class’s clients that send polymorphic

messages to this hierarchy may experience inconsistent

behaviour. For example, an inherited method may be incorrect

in the context of the subclass or an overriding method may

have pre conditions and post conditions different from the

ones for the overridden method. In deep inheritance

hierarchies, it is easy to forget to override methods for lower-

level subclasses.

There are a number of testing issues that are unique to object-oriented software. Several researchers have asserted

that some traditional testing techniques are not effective for

object-oriented software and that traditional software testing

methods test the wrong things. Specifically, methods tend to

be smaller and less complex, so path-based testing techniques

are less applicable [1][3][5]. Additionally, inheritance and

polymorphism brings undesirability to the software. The

execution path is no longer a function of the class's static

declared type, but a Intra-method testing, in which tests are

constructed for individual methods.

Inter-method testing, in which multiple methods within a class are tested in concert. Intra-class testing, in which tests

are constructed for a single class, usually as sequences of calls

to methods within the class. Inter-class function of the

dynamic type that is not known until run-time. A class is

usually regarded as the basic unit of OO testing. There are

three levels of testing. Testing in which more than one class is

tested at the same time. Gallagher and Offutt had added this

level.

Polymorphism is an important feature of object oriented

software which allows an object reference to bind with the

objects of other classes. Method overriding is a way of giving a method one name that is shared up and down in a class

hierarchy, with each class in the hierarchy implementing the

method in a way appropriate to itself.

The decision as to which method is to be used is left till

run time using a technique called Dynamic method lookup. If

the subclass includes a method with the same "signature" as a

method in the super class the super class will not be inherited.

300

It is observed that the subclass method overrides the super

class methods.

Dynamic binding allows the same statement to execute

different method bodies. Which one is executed depends on

the current type of the object that is used in the method call.

These features make testing more difficult because the exact data type and implementation cannot be determined statically

and the control flow of the object oriented program is less

transparent.

Various techniques can be found in testing literature for

testing of polymorphic relationships. Most approaches require

testing that exercises all possible bindings of certain

polymorphic references. Each possible binding of a

polymorphic component requires a separate test. It may be

hard to find all such bindings, increasing the chance of errors

and posing an obstacle to reaching coverage goals. This may

also complicate integration planning, in that many server

classes may need to be integrated before a client class can be tested.

II. COVERAGE CRITERIA FOR POLYMORPHISM

Various techniques for testing of polymorphic interactions

have been proposed in previous work. These approaches

require testing that exercises all possible polymorphic

bindings for certain elements of the tested software. Binder [4]

points out that “just as we would not have high confidence in

code for which only a small fraction of the statements or

branches have been exercised, high confidence is not

warranted for a client of a polymorphic server unless all the

message bindings generated by the client are exercised”. These requirements can be encoded as coverage criteria for

testing of polymorphism. There is existing evidence that such

criteria are better suited for detecting object-oriented faults

than the traditional statement and branch coverage criteria.

A program-based coverage criterion is a structural test

adequacy criterion that defines testing requirements in terms

of the coverage of particular elements in the structure of the

tested software. Such coverage criteria can be used to evaluate

the adequacy of the performed testing and can also provide

valuable guidelines for additional testing. In this paper, we

focus on two program-based coverage criteria for testing of

polymorphism [4][6]. The all-receiver-classes criterion requires exercising of all possible classes of the receiver

object at a call site. The all-target-methods criterion requires

exercising of all possible bindings between a call site and the

methods that may be invoked by that site. Some existing

approaches explicitly define coverage requirements based on

these criteria, while, in other approaches, the coverage of

receiver classes and/or target methods is part of more general

coverage requirements that take into account polymorphism .

For example, in addition to all-receiver-classes, proposes

coverage of all possible classes for the senders and the

parameters of a message.

III. CLASS ANALYSIS FOR COVERAGE TOOLS

The use of coverage criteria is essentially impossible

without tools that automatically measure the coverage

achieved during testing. A coverage tool analyses the tested

software to determine which elements need to be covered,

inserts instrumentation for runtime tracking, executes the test

suite, And reports the degree of coverage and the elements

that have not been covered.

To determine which software elements need to be covered, a coverage tool has to use some form of source code

analysis. Such an analysis computes the elements for which

coverage should be tracked and determines the kind and

location of the necessary code instrumentation. For simple

criteria such as statement and branch coverage, the necessary

source code analysis is trivial; however, the all-receiver-

classes and all-target-methods criteria require more complex

analysis [2][4]. To compute the all-receiver-classes and all-

target-methods coverage requirements, a tool needs to

determine the possible classes of the receiver object and the

possible target methods for each call site. In the simplest case,

this can be done by examining the class hierarchy i.e., by considering all classes in the sub tree rooted at the declared

type of the receiver object. It appears that previous work on

testing of polymorphism uses this approach (or minor

variations of it) to determine the possible receiver classes and

target methods at polymorphic calls.

Some of the existing work on static analysis for object-

oriented languages shows that using the class hierarchy to

determine possible receiver classes may be overly

conservative i.e., not all subclasses may be actually feasible.

Such imprecision has serious consequences for coverage tools

because the reported coverage metrics become hard to interpret, is the low coverage due to inadequate testing or is it

due to infeasible coverage requirements. This problem

seriously compromises the usefulness of the coverage metrics.

In addition, the person who creates new test cases may spend

significant time and effort trying to determine the appropriate

test cases before realizing that it is impossible to achieve the

required coverage.

This situation is unacceptable because the time and

attention of a human tester can be very costly compared to

computing time. To address above mentioned short comes,

propose using class analysis to compute the coverage

requirements. Class analysis is a static analysis that determines an overestimate of the classes of all objects to

which a given reference variable may point. While initially

developed in the context of optimizing compilers for object-

oriented languages, class analysis also has a variety of

applications in software engineering tools. In a coverage tool

for testing of polymorphism, class analysis can be used to

determine which are the classes of objects that variable x may

refer to at call site x.m(),from this information, it is trivial to

compute the all-receiver-classes and all-target-methods

criteria for the call site. There is a large body of work on

various class analyses with different tradeoffs between cost and precision. However, there has been no previous work on

using these analyses for the purposes of testing of

polymorphism.

IV.DYNAMIC METRICS

301

Dynamic polymorphic metrics measure the various aspect

of the polymorphism behavior in the programs [7].

a) Call Site Polymorphic Value (CSPV) metric count total

number of different call sites executed. This measurement

does not include static invoke instructions, but does count virtual method calls with a single receiver.

b) Invoke Density Polymorphic Value (IDVP) metric count

number of invoke Virtual and invoke Interface calls per kbc

executed. This metric estimates the importance of invoke byte

codes relative to other instructions in the program, indicating

the relevance of optimizing invokes.

c) Receiver Polymorphic Bin (RPB) metric shows the

percentage of all call sites that have one, two and more than

two different receiver types. The metric is dynamic, since we

measure the number of different types that actually occur in the execution of the program.

d) Receiver Call Polymorphic Bin (RCPB) metric shows the

percentage of all calls that occur from a call site with one, two

and more than two different receiver types. This metric

measures the importance of polymorphic calls.

e) Receiver Cache Miss Rate polymorphic Value (RCMRV)

metric shows as a percentage how often a call site switches

between receiver types. This is the most dynamic

measurement of receiver polymorphism, and it represents the miss rate of a true inline cache.

f) Target Polymorphic Bin (TPB) metric shows the percentage

of all call sites that have one, two and more than two different

target methods. This metric is dynamic, but does not reflect

the run time importance of call sites.

g) Target Call Polymorphic Bin (TCPB) metric shows the

percentage of all calls that occur from a call site with one, two

and more than two different target methods.

h) Target Cache Miss Rate polymorphic Value (TCMRV) metric shows as a percentage how often a call site switches

between target methods. It represents the miss rate of an

idealized branch target buffer. It is always lower than the

corresponding inline cache miss rate since targets can be equal

for different receiver types. Accordingly, this metric can also

be heavily influenced by the order in which target methods

occur.

i) Dynamic polymorphism in ancestors (DPA) is the sum of

number of dynamic polymorphism function members in

ancestor that appears in the different classes.

j) Dynamic Polymorphism in Descendants (DPD) is the sum

of number of dynamic polymorphism function members in

descendant that appears in the different classes.

k) Average Changing Rate of Virtual methods (ACRV) are

used to check the efficiency by using run time method

resolution.

IV. MODELING POLYMORPHISM: THE YOYO GRAPH

One of the major difficulties with using polymorphism

and dynamic binding is that of modeling and visualizing the complicated interactions. The essential problems are that of

understanding which version of a method will be executed and

which versions can be executed. The fact that execution can

sometimes “bounce” up and down among levels of inheritance

has been called the yo-yo effect by Binder and he introduced a

preliminary graph [8]. We have extended this notion as a basis

for a graphical representation that we call the “yo-yo graph” to

show all possible actual executions in the presence of dynamic

binding.

A

t u v w

D() G() H() I() J() L()

B

x

H() I()

K()

C

I() J() L()

Fig1. Data Flow anomalies with polymorphism

Consider the inheritance hierarchy shown in Figure 1.

Assume that in A’s implementation, d() calls g(), g() calls h(),

h() calls i(), and i() calls j(). Further, assume that in B’s

implementation, h() calls i(), i() calls its parent’s (that is, A’s)

Methods Defines Users

A..h A..u, A..w

A..i A..u

A..j A..v A..w

A..l A..v

B..h A..x

B..i B..x

C..i A..y

C..j C..v

C..l A..v

302

version of i(), and k() calls l(). Finally, assume that in C’s

implementation, i() calls its parent’s (this time B’s) version of

i(), and j() calls k().

A

d() g() h() i() j() l()

Fig 2. Calls to d() when object has various actual types.

Figure 2 is a yo-yo graph of this situation and expresses the

actual sequence of calls if a call is made to d() through an

instance of actual type A, B, and C. At the top level of the

graph, it is assumed that a call is made to method d() through

an object of actual type A. In this case, the sequence of calls is

simple and straightforward. In the second level, where the

object is of actual type B, the situation starts to get more

complex. When g() calls h(), the version of h() defined in B is

executed (the light dashed line from A::g() to A::h()

emphasizes that A::h() is not executed). Then control

continues to B::i(), A::i(), and then to A::j(). When the object

is of actual type C, it becomes clear why the term “yo-yo” is

used. Control proceeds from A::g() to B::h() to C::i(), then

back up through B::i() to A::i(), back to C::j(), back up to B::k(), and finally down to C::l(). Along with this induced

complexity comes more difficulty and required effort in

testing.

V. CATEGORIES OF ANOMALIES

Benefits of using inheritance include more creativity,

efficiency, and reuse. Unfortunately, it also allows a number

of anomalies and potential faults that anecdotal evidence has

shown to be some of the most difficult problems to detect,

diagnose and correct. This section presents a list of fault types

that can be manifested by polymorphism. Table 1 summarizes

the fault types that result from inheritance and polymorphism.

The goal is a complete list of faults, though we do not make

this claim. Most of the types are programming language-

independent, although the language that is used will affect how the faults manifest. In all cases, we are concerned with

how each anomaly or fault is manifested through

polymorphism in a context that uses an instance of the

ancestor [5][7][9]. Thus, we assume that instances of

descendant classes can be substituted for instances of the

ancestor.

TABLE I

ANOMALIES DUE TO INHERITANCE AND POLYMORPHISM.

VI. STANDARDS FOR APPROPRIATE USE OF

POLYPORPHISM

Software is increasingly being built by combining and

extending pre-existing software components. In particular, we often create new classes through inheritance by extending

from pre-existing classes. Moreover, we often do not have

access to the source for these library classes. Although this is

a very powerful abstraction mechanism, the implementation

can be somewhat problematic. In particular, careless

inheritance and overriding can create problems in the state

space interactions of the resulting objects. Unfortunately,

providers of class libraries often want to keep the

implementation proprietary, and thus do not provide the

source. Since the developer may not know the internals of the

parent class, there is no way to know what type of inheritance

and polymorphisms. In cooperation with dynamic binding, polymorphism becomes a very useful feature, as it allows a

program to have dynamic behaviour according to the actual

A d() g() h() i() j() l()

B h() i() k()

C i() j() l()

B

h() i() k()

A d() g() h() i() j() l()

B h() i() k()

C i() j() l()

C

i()

j()

l()

A d() g() h() i() j() l()

B h() i() k()

C i() j() l()

ACRONYM ANOMALY

ITU Inconsistent Type Use (context swapping)

SDA State Definition Anomaly (possible post-condition violation)

SDIH State Definition Inconsistency (due to state variable hiding)

SDI State Defined Incorrectly (possible post condition violation)

IISD Indirect Inconsistent State Definition

ACB1 Anomalies Construction Behavior 1

ACB2 Anomalies Construction Behavior 2

IC Incomplete construction

SVA State visibility Anomaly

303

type of object under execution. A number of design patterns

are formalized based on this concept. Although it is a very

useful feature many researchers address that polymorphism

does not prevent software from defects [2][3][8][9]. Barbey

and Strohmeier state that to gain confidence in an operation a

statement of which is a call to a dynamically invoked operation, it becomes necessary to make assumptions

regarding the nature of the polymorphic parameters, i.e. a

“hierarchy specification” must be provided for each of their

operations, which specifies the expected minimal behaviour of

their operation and all their possible redefinitions. Therefore

whenever methods are inherited by a derived class some

further testing is necessary which must be in context to the

derived class. This is true both for those methods which have

been inherited unchanged and those which have been

overridden.

VII. EXPERIMENTAL RESULTS

A polymorphic call in object oriented system takes the form of invoke virtual or invoke interface.

TABLE II

ANALYSIS OF RESULTS

METRIC I II III IV V

CSPV 1 2 3 4 5

IDVP 6 19 13 11 23

RPB 22% 7.14% 8.09% 9.09% 4.35%

RCPB 20% 57.14% 54.55% 63.64% 60.67%

RCMRV 0% 28.57% 27.27% 36.36% 17.36%

TPB 20% 85.71% 75% 99% 76.92%

TCPB

20% 80% 75% 100% 78.78%

TCMRV 0% 16.67% 16.67% 0.14% 0.1%

DPA 2 1 2 1 2

DPD 1 1 1 1 1

ACRV 0.38% 0.43% 0.5% 0.39% 0.53%

This paper is presenting eleven dynamic metrics as a means of

assessing the actual run time behavior of a program by providing a high-level overview of several of its key aspects.

This dynamic information can be more relevant than the more

common static measures. These metrics are designed with the

goals of being unambiguous, dynamic, robust, discriminating,

and platform-independent. These metrics are also classified in

four categories value, percentile, bin and continuous. The

utility of the metrics is evaluating by applying them to five

specific dynamic polymorphic problems and determining to

which extent they provided useful information for each task.

We furthermore want to establish the preciseness of using

dynamic metrics as helping tools for exploratory program understanding in object oriented system. Implementation of

dynamic metrics is given in this section on the set of five

different polymorphic programs. The results are mentioned in

Table II.

VIII.CONCLUSIONS AND FUTHERWORK

We have shown the types of faults are likely to be

encountered in object-oriented programs. The results of our

empirical study suggest that traditional testing techniques,

such as functional testing and white-box approaches that

utilize the statement and branch coverage criteria, are not

adequate for detecting object-oriented faults. The object flow

based testing strategy, however, takes object-oriented features

into account and is reliable with respect to most OO faults.

Therefore, to have confidence in testing object oriented

programs, it is important to incorporate traditional testing

techniques with the object-flow based testing strategy. Nevertheless, with a few feasible efforts the reliability of the

program can be improved effectively and the maintenance

cost can be reduced significantly. Much work remains to be

done in this research area. We are currently investigating the

correlation of Object Oriented design metrics with fault

population and distribution. It is expected that such a

relationship can be used to select effective testing techniques

and to prioritize these selected techniques, so that most faults

can be detected with an affordable cost. Meanwhile, we are

studying the granularity of the object flow based testing

strategy to compare the strengths and the cost of this strategy with the data flow testing strategy which is a further

refinement of the object-flow strategy.

REFERENCES

[1] M.D. Smith and D.J. Robson, “Object-Oriented Programming: The

Problems of Validation”, Proc. IEEE Conf. Software Maintenance, IEEE

Computer Society Press, Los Alamitos, Calif., pp. 272-281, 1990.

[2] M.H. Ghai and M. H.Kao,”Testing Object-Oriented Programs an

Integrated Approach. In Proceedings of the 10th international

Symposium on Software Reliability Engineering”, ISSRE.IEEE

Computer Society, Washington, DC, pp.73-83, November 01-04, 1999.

[3] R. Alexander and J. Offutt.,“Criteria for Testing Polymorphic

Relationships”, In Proceedings of the 11th international Symposium on

Software Reliability Engineering (Issre'00). ISSRE. IEEE Computer

Society, Washington, DC, pp.15-23, October 08 - 11, 2000.

[4] Atanas Rountev, Ana Milanova and Barbara G. Ryder, “Fragment Class

Analysis for Testing of Polymorphism in Java Software”, IEEE Trans.

Software Engineering, vol. 30, no.6, pp.372-387, Jun. 2004.

[5] S. Barbey and A. Strohmeier, “The Problematic of Testing Object-

Oriented Software”, Proceedings of SQM '94 Second Conference on

Software Quality Management, Edinburgh, Scotland, UK, vol.2, pp.

411-426, 1994.

[6] Stephane Barbey and Alfred Strohmeier,”The problematic of testing

object-oriented software”, In SQM'94 Second Conference on Software

Quality Management, Edinburgh, Scotland, UK, July 26-28, vol. 2,

pp.411-426, 1994.

[7] Parvinder Singh Sandhu, and Gurdev Singh,”Dynamic Metrics for

Polymorphism in Object Oriented Systems”, In World Academy of

Science, Engineering and Technology , vol. 39, no.3, pp.412-121, 2008.

[8] Jeff Offutt and Roger Alexander, “A Fault Model for Subtype

Inheritance and Polymorphism”, The Twelfth IEEE International

Symposium on Software Reliability Engineering (ISSRE ’01), Hong

Kong, PRC, November 2001.

304

[9] Mei-Hwa Chen and Howard M. Kao,”Testing Object-Oriented

Programs - An Integrated Approach”, ISSRE. IEEE Computer Society,

Los Alamitos, 2007.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

305

A paper on Efficient Reverse Turing Test in

Artificial Intelligence

G. Geetha

PG Computer Science,

Sree Saraswathy Thyagaraja College, Pollachi

1 [email protected]

Abstract— This paper presents an introduction to

Artificial intelligence (AI). Artificial intelligence is

exhibited by artificial entity, a system is generally

assumed to be a human intelligence of computer.

This paper takes into the consideration of the

relationship between the two independent lines of

research into user modelling activities in human-

computer interaction (HCI) and artificial intelligence

(AI). The user modelling research conducted in HCI

and AI respectively is then discussed along with the

goals of HCI and AI. In addition a view of how the

various Turing Testing Techniques are employing AI

user models including the Reverse Turing Tests is

evaluated. The paper begins by considering the AI

techniques and the field of HCI with the later

sections discussing the historical evolution of AI,

various Turing Test Techniques involved. The

remaining paper focuses on presenting the Reverse

Turing Test involved with AI its applications.

Finally, we bring up the Enhanced methodology

using CAPTCHA as a contributed work which is

more efficient Reverse Turing Test than the existing

method. This Enhanced methodology, EPIX is

discussed with its technical work clearly projected

out along with its advantages and future work.

Keywords— Artificial Intelligence (AI), CAPTCHA

(Completely Automated Public Turing test to tell Computers and Humans Apart), Human Computer

Interaction (HCI)

I. INTRODUCTION

The scientific and engineering discipline that makes up or is closely allied with AI is relatively recent origin.

They date back to the 1950s at the earliest, with the first

encyclopaedia article on AI appearing in 1967. AI have been rapidly changing and expanding both in intellectual

content and in application, especially in the last decade.

The recent accelerated pace of change has been due in no small part to the almost breathtaking innovations and

product developments in the supporting microelectronics,

electro optics, and display technologies—all hardware-intensive areas. Growth in the amount of new

terminology and technical jargon always accompanies

such changes and advances in technology. Recognizing

the futility of attempts to craft comprehensive definitions in light of these rapid changes, the panel has

opted to provide brief descriptions of four frequently

used terms: artificial intelligence, human-computer interface, virtual worlds, and synthetic environments.

A. Artificial intelligence (AI)

Artificial intelligence (AI) is defined as intelligence

exhibited by an artificial entity. Such a system is

generally assumed to be a computer [3]. Although AI

has a strong science fiction connotation, it forms a vital branch of computer science, dealing with intelligent

behaviour, learning and adaptation in machines.

Research in AI is concerned with producing machines to automate tasks requiring intelligent behaviour. Examples

include control, planning and scheduling, the ability to

answer diagnostic and consumer questions, handwriting,

speech, and facial recognition. As such, it has become a scientific discipline, focused on providing solutions to

real life problems. AI systems are now in routine use in

economics, medicine, engineering and the military, as well as being built into many common home computer

software applications, traditional strategy games like

computer chess and other video games. Artificial intelligence is the collection of

computations that at any time make it possible to assist

users to perceive, reason, and act. Since it is

computations that make up AI, the functions of perceiving, reasoning, and acting can be accomplished

under the control of the computational device (e.g.,

computers or robotics) in question. AI at a minimum includes representations of "reality,"

cognition, and information, along with associated

methods of representation, Machine learning, representations of vision and language, Robotics and

Virtual reality.

306

B. Human-computer interface (HCI)

Human-computer interface (HCI) consists of the machine integration and interpretation of data and their

presentation in a form convenient to the human operator

or user (i.e., displays, human intelligence emulated in computational devices, and simulation and synthetic

environments) [1]. The bidirectional communication of

information between two powerful

Humans and computers are destined to interact on a more personal level someday in the future. As

computers become smarter, they will no doubt require a

more personal interaction to achieve greater results. No longer will we use a computer through interface devices

such as a mouse and keyboard to enter data, but instead

we will ask it to do required tasks for us, and it will do them with ease, in the same fashion that we ourselves

would [2]. In fact, it will become feasible that computers

will act and emulate humans −− even to the fact that

they will assist us in our daily tasks. In short, these computers will end up masquerading as people.

II. EVOLUTION OF AI

The intellectual roots of AI, and the concept of intelligent machines, may be found in Greek mythology.

Intelligent artifacts appear in literature since then, with

real mechanical devices actually demonstrating behaviour with some degree of intelligence. After

modern computers became available following World

War-II, it has become possible to create programs that

perform difficult intellectual tasks. In 1950 – 1960 the first working AI programs were

written in 1951 to run on the Ferranti Mark I machine of

the University of Manchester (UK): a draughts-playing program written by Christopher Strachey and a chess-

playing program written by Dietrich Prinz.

In 1960 – 1970, during the 1960s and 1970s Marvin

Minsky and Seymour Papert publish Perceptrons, demonstrating limits of simple neural nets and Alain

Colmerauer developed the Prolog computer language.

Ted Shortliffe demonstrated the power of rule-based systems for knowledge representation and inference in

medical diagnosis and therapy in what is sometimes

called the first expert system. Hans Moravec developed the first computer-controlled vehicle to autonomously

negotiate cluttered obstacle courses.

From 1980’s onwards, neural networks became

widely used with the back propagation algorithm, first described by Paul John Werbos in 1974. The 1990s

marked major achievements in many areas of AI and

demonstrations of various applications. Most notably Deep Blue, a chess-playing computer, beat Garry

Kasparov in a famous six-game match in 1997.

III. RELATED WORK - TESTING TECHNIQUES OF AI

A. Turing Test

The approach is best embodied by the concept of the Turing Test. Turing held that in future computers can be

programmed to acquire abilities rivalling human

intelligence. As part of his argument Turing put forward the idea of an 'imitation game', in which a human being

and a computer would be interrogated under conditions

where the interrogator would not know which was which, the communication being entirely by textual messages.

Turing argued that if the interrogator could not

distinguish them by questioning, then it would be

unreasonable not to call the computer intelligent. Turing's 'imitation game' is now usually called 'the

Turing test' for intelligence. Consider the following

setting. There are two rooms, A and B. One of the rooms contains a computer. The other contains a human. The

interrogator is outside and does not know which one is a

computer. He can ask questions through a teletype and receives answers from both A and B. The interrogator

needs to identify whether A or B are humans. To pass

the Turing test, the machine has to fool the interrogator

into believing that it is human. Logic and laws of thought deals with studies of ideal

or rational thought process and inference, the emphasis

in this case is on the inference mechanism, and its properties. That is how the system arrives at a

conclusion, or the reasoning behind its selection of

actions is very important in this point of view. The

soundness and completeness of the inference mechanisms are important here.

The view of AI is that it is the study of rational agents.

This view deals with building machines that act rationally. The focus is on how the system acts and

performs, and not so much on the reasoning process

Fig. 1. Illustrating Turning Test in AI

B. Reverse Turing Test

The term reverse Turing test has no single clear

definition, but has been used to describe various

situations based on the Turing test in which the objective

307

and/or one or more of the roles have been reversed

between computers and humans.

Conventionally, the Turing test is conceived as having a human judge and a computer subject which attempts to

appear human. Critical to the concept is the parallel

situation of a human judge and a human subject, who

also attempts to appear human. The intent of the test is for the judge to attempt to distinguish which of these

two situations is actually occurring. It is presumed that a

human subject will always be judged human, and a computer is then said to "pass the Turing test" if it too is

judged human. Any of these roles may be changed to

form a "Reverse Turing test".

Arguably the standard form of reverse Turing test is one in which the subjects attempt to appear to be a

computer rather than a human.

A formal reverse Turing test follows the same format as a Turing test. Human subjects attempt to imitate the

conversational style of a conversation program such as

ELIZA. Doing this well involves deliberately ignoring, to some degree, the meaning of the conversation that is

immediately apparent to a human, and the simulation of

the kinds of errors that conversational programs

typically make. Arguably unlike the conventional Turing test, this is most interesting when the judges are very

familiar with the art of conversation programs, meaning

that in the regular Turing test they can very rapidly tell the difference between a computer program and a human

acting normally.

The humans that perform best (some would say worst) in the reverse Turing test are those that know computers

best, and so know the types of errors that computers can

be expected to make in conversation. There is much

shared ground between the skill of the reverse Turing test and the skill of mentally simulating a program's

operation in the course of programming and especially

debugging. As a result, programmers (especially hackers) will sometimes indulge in an informal reverse Turing

test for recreation.

An informal reverse Turing test involves an attempt to

simulate a computer without the formal structure of the Turing test. The judges of the test are typically not

aware in advance that a reverse Turing test is occurring

and the test subject attempts to elicit from the 'judges' (who, correctly, think they are speaking to a human) a

response along the lines of "is this really a human?‖

Describing such a situation as a "reverse Turing test" typically occurs retroactively.

C. Design of a Reverse Turing Test

We propose what we call a ―reverse Turing test‖ of the following kind. When a user — human or machine —

chooses to take the test (e.g. in order to enter a protected

Web site), a program challenges the user with one

synthetically generated image of text; the user must type back the text correctly in order to enter. This differs

from Turing’s proposal in at least four ways:

The judge is a machine, rather than human;

There is only one user, rather than two;

The design goal is to distinguish, rather than to

fail to

Distinguish, between human and machine; and

The test poses only one challenge – or very

few—rather than an indefinitely long sequence

of challenges.

The challenges must be substantially different almost every time, else they might be recorded exhaustively,

answered off-line by humans, and then used to answer

future challenges. Thus we propose that they be generated pseudo randomly from a potentially very large

space of distinct challenges.

"CAPTCHA" is an acronym for "Completely

Automated Public Turing test to tell Computers and Humans Apart" so that the original designers of the test

regard the test as a Turing test.

The term "reverse Turing test" has also been applied to a Turing test (test of humanity) that is administered by

a computer. In other words, a computer administers a

test to determine if the subject is or is not human.

CAPTCHA are used in some anti-spam systems to prevent automated bulk use of communications systems.

CAPTCHA is cropping up everywhere as it is used to

defend against:

Skewing search-engine rankings (Altavista,

1999)

Infesting chat rooms, etc (Yahoo!, 2000)

Gaming financial accounts (PayPal, 2001)

Robot spamming (MailBlocks, SpamArrest

2002)

In the last two years: Overture, Chinese

website, HotMail,

CD-rebate, TicketMaster, MailFrontier, Qurb,

Madonnarama, Gay.com, …

On the horizon it is being covered out at all ballot

stuffing, password guessing, denial-of-service attacks

`blunt force’ attacks (e.g. UT Austin break-in, Mar ’03). The use of CAPTCHA is controversial.

Circumvention methods exist that reduce their

effectiveness. Also, many implementations of CAPTCHA (particularly ones desired to counter

circumvention) are inaccessible to humans with

308

disabilities, and/or are difficult for humans to pass.

Types of CAPTCHA are

Text based i. Gimpy, ez-gimpy

ii. Gimpy-r, Google CAPTCHA

iii. Simard’s HIP (MSN)

Graphic based i. Bongo

ii. Pix

Audio based

D. Text Based CAPTCHA

In Text Based CAPTCHA the methodologies like

Gimpy, ez-gimpy a word or words from a small dictionary is picked up and they are distort them and add

noise and background

In Gimpy-r, Google’s CAPTCHA methodology we pick random letters and distort them, then add noise and

background

In Simard’s HIP methodology we pick random letters and numbers and distort them and then finally add arcs

Fig. 2. Illustrating Text Based CAPTCHA

E. Graphic Based CAPTCHA

In graphics based CAPTCHA the Bongo methodology is used which displays two series of blocks and the user

must find the characteristic that sets the two series apart

user is asked to determine which series each of four single blocks belongs to.

Fig. 3. Illustrating Graphics Based CAPTCHA

F. Audio Based CAPTCHA

In Audio Based CAPTCHA, the general

methodologies is to pick a word or a sequence of numbers at random then render them into an audio clip

using TTS software and distort the audio clip then ask

the user to identify and type the word or numbers

G. Related work in Reverse Turing Test

• Gimpy

Gimpy picks seven random words out of a dictionary,

distorts them, and renders them to users. The user needs

to recognize three out of the seven words to prove that he or she is a human user. Because words in Gimpy

overlap and undergo nonlinear transformations, they

pose serious challenges to existing OCR systems.

However, they also place a burden on human users. The burden is so great that Yahoo pulled Gimpy off its Web

site. The CMU team later developed an easier version,

EZ Gimpy. It shows a single word over a cluttered background and is currently used on Yahoo’s Web site.

• Bongo

Bongo explores human ability in visual pattern

recognition. It presents to a user two groups of visual patterns (e.g., lines, circles, and squares), named LEFT

and RIGHT. It then shows new visual patterns and asks

the user to decide if the new patterns belong to LEFT or RIGHT.

• Pix

Pix rely on a large database of labelled images. It first randomly picks an object label (e.g., flower, baby, lion,

etc.) from the label list and then randomly selects six

images containing that object from the database and

shows the images to a user. The user needs to enter the correct object label to prove he or she is a human user.

Fig. 4. Illustrating the Image for the object Dog

IV. THE EPIX -CONTRIBUTION WORK

EPIX is the Enhanced methodology of the PIX which the reverse Turing test techniques. It is very similar to

PIX in which we pick a concrete object then get 6

images at random from images.google.com that match

the object and then distort the images. Then a list of 100 words is built in which 90 is taken from a full dictionary,

10 from the objects dictionary. Then the user is

prompted to pick the object from the list of words

A. EPIX – Technical Issues

Make an HTTP call to images.google.com and

search for the object

309

Screen scrape the result of 2-3 pages to get the

list of images

Pick 6 images at random

Randomly distort both the images and their

URLs before displaying them

Expire the CAPTCHA in 30-45 seconds

B. EPIX – Advantages

The database already exists and is public

The database is constantly being updated and

maintained

Adding ―concrete objects‖ to the dictionary is

virtually instantaneous

Distortion prevents caching hacks

Quick expiration limits streaming hacks

V. CONCLUSIONS

We tried to explain with brief ideas the fundamentals

of AI. The paper then discusses the human computer

interaction and they both are significantly linked to each other. Then the historical evolutions of over years of AI

are projected. The Turing test Techniques are discussed

in detail along with Reverse Turing Test and its

categories are brought out. The related works of Reverse Turing test are studied and a proficient methodology of

Enhanced PIX, EPIX has been contributed through this

paper which can perform more efficient then the existing methodologies.

We conclude this paper by contributing a new

methodology by making enhancements in the existing system which could bring out a tremendous and

effective performance ratio when compared to the

existing methodologies

REFERENCES

[1] M. Schmitz, J. Baus, and S. Schmidt.

Anthropomorphized objects: A novel interaction

metaphor for instrumented spaces. In A. Q. Thomas Strang, Vinny Cahill, editor, Pervasive 2006

Workshop Proceedings. Deutsches Zentrum f¨ur

Luftund Raumfahrt. ISBN 978-3-00-018411-6, 2006.

[2] Schmidhuber, J. 2003. Goedel machines: self-

referential universal problem solvers making

provably optimal self-improvements. In Artificial General Intelligence, eds. B. Goertzel and C.

Pennachin. Forthcoming. New York: Springer-

Verlag. [3] Advances in Artificial Intelligence subject of

conference 23rd AAAI Conference on AI (AAAI-

08) July 13-17, 2008 Hyatt Regency McCormick

Place, Chicago, Illinois.

[4] F. O. Karray, C. de Silva. Soft Computing and

Intelligent Systems Design. Addison Wesley, 2004.

Concepts of Artificial Intelligence. [5] Artificial Intelligence – An Introduction to

Robotics Tim Niemueller and Sumedha

Widyadharma July 8, 2003.

[6] J. Hawkins, S. Blakeslee. On Intelligence. Times Books, 2004.An interesting book on a

comprehensive theory of intelligence based on the

human brain and neuro-science. [7] GregWelch and Gary Bishop. An Introduction to

the Kalman Filter. Technical report, Department of

Computer Science – University of North Carolina

at Chapel Hill, 2003. [8] RWTH Aachen Robocup Team. AllemaniACs.

http://robocup.rwth-aachen.de, 2003.

[9] Sebastian Thrun. Robotic Mapping: A Survey. Technical report, School of Computer Science –

Carnegie Mellon University, 2003.

[10] Stuart Russel and Peter Norvig. Artificial Intelligence. A Modern Approach, 2nd Edition.

Prentice Hall, 2003

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

310

BOMB DETECTION USING WIRELESS SENSOR AND

NEURAL NETWORK

Sharon Rosy S, M.E- Power System (PT) Doing, Anna University, Trichy

[email protected]

Shakena Grace S, M.E- Computer Science.

[email protected]

ABSTRACT : Today, if we want to

know where a terrorist attack will

happen, then we need to think like a

terrorist. Terrorists want to kill as many

people as possible.

At present bombs are detected

by using different machines like x-ray

machines, handheld detectors etc..In case

of x-ray machine the pictures are

captured and transmitted to experts.

They will analyze the presence and type

of bomb. Other detectors also used only

with the presences of experts. But this

way of analyzing takes more time and

make risk to life of experts

Solution for this problem is

provided in this paper using neural

network technology and GPS tracking

system. Kalman‟s filter algorithm is

used to train the network for automatic

recognition of bomb.

The wireless sensor network is

used to connect all the sensors which are

present at different locations and it is

used to produce the result to the

controller. This the static part of finding

the bomb

GPS tracking system is used to

position the exact place of bomb .This is

the dynamic part of the operation.

Index terms—Sensor networks, bomb,

neural network, GPS, KF

I. INTRODUCTION

Terrorists have a huge economic

advantage over law enforcement because it

is, many times, more expensive to detect

terrorist threats than it is to deploy terrorist

threats. The fundamental concept behind

explosives is very simple. At the most basic

level, an explosive is just something that

burns or decomposes very quickly,

producing a lot of heat and gas in a short

amount of time.

A bomb is a weapon designed to

explode when thrown, dropped or detonated

by some form of signal. A bomb is a hollow

case filled with explosives or other

destructive chemicals and exploded by

percussion or a timing device. The outer

case may be metal, plastic, concrete or glass

and the shape, size and contents of a bomb

311

depends on its application. Explosives used

for peaceful purposes such as mining and

heavy construction, are referred to as

explosive charges, but when used to maim

or destroy persons or property, they are

called "bombs".

Problem observed in the existing

work:

At present the bomb are detected

using different instruments. In all the cases

the human expert presences is necessary to

analysis the presences of bomb. In this

situation, we may lose very good analyst

when bomb gets explode. In case of x-ray

machine the pictures are captured and

transmitted to experts. They will analyze the

presence and type of bomb but they take

more time to analyze and 24 hours manual

work is necessary to get a proper result.

Training should be provided to x-ray

operators as well as security personnel and

supervisors. It takes long time to make them

expert.

Some of the drawbacks in existing system

are

Fault in suspension

Impossible to continue work

Takes long time to analyze

May miss the sensed object before

analyzing

Dog is also exiting method but there

concentration can be distracted

The main aim to propose this

concept.

To produce immediate result

To reduce manual power

To avoid manual error

To collect information from different

node at a time

To locate a exact position

II SENSING

Sensors are used here to sense the

explosive materials which are also called as

bomb detecting devices or sensors. Wireless

Sensor Network (WSN) is a network

designed with bomb sensor nodes and

deployed in different locations to sense the

bombs. WSN is scattered in a region where

it is meant to collect data through bomb

sensor nodes. Sensor networks involve three

areas:

Sensing (hardware)

Communications (software)

Computation (Algorithm).

The main architectural components

of a sensor network: Data acquisition

(Achievement) by sensors, Data transport to

system, Data analysis by the trained

network.

Bomb detecting device is taken as

sensor node. This sensor can be placed in

one static place and information can be

continuously gathered .Based on coverage

capacity of sensors, the number of sensors is

used in particular region. We can use any

kind of trustworthy bomb sensors and

portioning of area depends on these sensors.

For understanding, we can name the sensors

as S1, S2, S3…..These sensors will sense

the bomb when it enters their sensing area.

In case of joining different nodes in

different place to the controller through

wired mean is expensive, at the same time it

alerts the terrorist to damage the sensor. So

sensors are connected to the controller room

through wireless means. These sensors sense

and transmit the information trough

transmitter to the receiver at the controller

room. Here the wireless sensor is chosen to

312

change the position or to replace sensors

when their life span ends.

III TRAINED NETWORK

Neural networks are based on the

parallel architecture Neural networks are a

form of multiprocessor computer system,

with

simple processing elements

a high degree of interconnection

Interacting with noisy data or data

from the environment

Massive parallelism

Fault tolerance

Adapting to circumstances

Based on various approaches, several

different learning algorithms have been

given in the literature for neural networks.

Almost all algorithms have constant learning

rates or constant accelerative parameters,

though they have been shown to be effective

for some practical applications. The learning

procedure of neural networks can be

regarded as a problem of estimating (or

identifying) constant parameters (i.e.

connection weights of network) with a

nonlinear or linear observation equation.

The Kalman filter purpose is to use

measurements that are observed over time

that contain noise (random variations) and

other inaccuracies, and produce values that

tend to be closer to the true values of the

measurements and their associated

calculated values. The Kalman filter has

many applications in technology, and is an

essential part of the development of space

and military technology.

The Kalman filter produces estimates

of the true values of measurements and their

associated calculated values by predicting a

value, estimating the uncertainty of the

predicted value, and computing a weighted

average of the predicted value and the

measured value. The most weight is given to

the value with the least uncertainty.

The Kalman filter uses a system's

dynamics model (i.e. physical laws of

motion), known control inputs to that

system, and measurements (such as from

sensors) to form an estimate of the system's

varying quantities (its state) that is better

than the estimate obtained by using any one

measurement alone. As such, it is a common

sensor fusion algorithm.

All measurements and calculations

based on models are estimates to some

degree. Noisy sensor data, approximations

in the equations that describe how a system

changes, and external factors that are not

accounted for introduce some uncertainty

about the inferred values for a system's state.

Here, consider the problem of

determining the precise location of a moving

vehicle, say helicopter. The vehicle can be

equipped with a GPS unit that provides an

estimate of the position within a few meters.

The GPS estimate is likely to be very noisy

and jump around at a high frequency, though

always remaining relatively close to the real

position. The vehicle's position can also be

estimated by integrating its velocity and

direction over time, determined by keeping

track of the amount the accelerator is

depressed and how much the steering wheel

is turned. This is a technique known as dead

reckoning. Typically, dead reckoning will

provide a very smooth estimate of the

helicopter's position,

313

The Kalman filter is an efficient

recursive filter that estimates the internal

state of a linear dynamic system from a

series of noisy measurements. However, by

combining a series of measurements, the

Kalman filter can estimate the entire internal

state.

The sensor which detects the

radiation on basis of Vapour Pressures

(Normalized to atmospheric pressure) of

some common explosives (Parts-Per-

Billion) is taken as example to train the

network.

Temp.(in

Celsius)

DNT TNT PETN RDX

0 8

10 36 1 0.001

20 160 5 0.007 0.003

30 490 19 0.046 0.014

40 72 0.270 0.070

50 250 1.40 0.294

Vapour Pressures (Normalized to atmospheric pressure) of some common explosives (Parts-Per-Billion)

Most explosives have low-pressures

at ambient temperatures. The following table

shows vapor pressures of trinitrotoluene

(TNT), 2,4 dinitrotoluene (DNT),

pentaerythritol tetra nitrate (PETN),and

hexahydro-1,3,5-trizine (RDX) at a range of

temperatures.TNT is one if the more

commonly used explosives. DNT is a

synthetic byproduct of TNT. The saturation

concentrations of DNT in air are

approximately 25 times larger than TNT.

High explosives, such as PETN and RDX,

are the most serious threats.

Network is trained by using an

annealing schedule to determine how many

training passes to complete at every

radiation and temperature interval

Once the network is trained, it will

produce an output identifying the probable

cause indicated by the input symptom map.

The network will do this when the input is

equivalent to one of the training inputs, as

expected, and it will produce an output

indicating the likely cause of the problem

when the input is similar to, but different

from, any training input.

Once the network is trained with the

values of radiation and temperature. It will

produce an output after identifying the

probable cause indicated by the input

symptom map. The network will do this

when the input is equivalent to one of the

training inputs from the sensor, as expected,

and it will produce an output indicating the

likely cause of the problem when the input is

similar to, but different from, any training

input.

Similarly, the network is trained

according to the sensor input. If sensor sense

the radiation it transmit the information

through wireless means to the receiver

which is connected to trained system and the

system take the information as „1‟in case of

radiation, if no radiation it will be stetted as

„0‟.after,reciving „1‟ the alarmed signal is

produced to the concern authority. At a time

we can receive information from multiple

sensors.

IV. GPS TRACKING

Now here, we use Global Positioning

System (GPS) tracking technique. Place a

good bomb detecting sensor and GPS device

inside a helicopter which is controlled by a

remote. Remote can operated from anyplace

but the distance of operation will be based

on transmitting signal‟s type.

314

The principle behind GPS is that

receivers are able to use the technique of

“trilateration” to calculate their coordinates

on Earth by measuring the time taken for

signals from various satellites to reach them.

The GPS software will account for any

irregularities in the signal strength and clock

differences between itself and the GPS

satellite network by using signals from four

separate satellites to improve accuracy.

Usually the coordinates are then used

to locate the GPS device on a map, which is

either displayed to the user or used as a basis

for calculating routes, navigation, or as input

into mapping programs.

In fact, it is this use which

represents the simplest form of GPS

tracking. The user is able, using a portable

GPS device, to keep a track of where they

have been, in order to be able to either

retrace their steps, or follow the same path

again in the future. For example, Say,

Helicopter is designed as Light weight

By visualizing through GPS, we can

monitor position, and control the movement

of helicopter carrying GPS device and

sensor.GPS device is the transmitter here to

show the map of that particular region,

receiver receives the signals and visualize it

to the computer. From control room, the

authorized person can control the sensor. If,

the bomb is detected by the sensor, the

intensity of the LED at sensor gets increase,

when, it goes nearby to the bomb placed

from the respective sensor .Now, process

gets easy. Sensor senses the bomb and GPS

shows the path to search. Applying any

process to remove the bomb, the process of

removing bombs may be done faster and

easier.

V.CONCLUSION

There may be many varieties of

sensors and so many types of bombs. But

selecting precise sensor and removing bomb

earlier by analyzing easier is important.

By using technologies which is in

our hand, trust we can make developments

in technologies and do everything effective

and save our precious life from dangers. Let

the small thing may make great change with

many ideas in future.

REFERENCES

1. James A. Freeman and David M.

Skapura (2002) “Neural network

2. Algorithms, Application and

Programming Techniques”, Low

price edition. Page 189-211.

3. Simon Haykins (2001) “Neural

network, A Comprehensive

foundation”, Low price edition,2nd

ed. Page 1-5.

4. Sean M. Brennan, Angela M.

Mielke, and David C. Torney (2004)

“Radiation Detection with

Distributed Sensor Networks”

5. Edward J. Seuter, “Understanding

the strengths and limitations of each

bomb detection technology is the key

to a good protection strategy”

6. James Dunnigan (2004) “The Truth

about Dirty Bombs

7. Dr. Frank Barna (2005) “Dirty

Bombs and Primitive Nuclear

Weapons”

8. M. W. Siegel and A. M. Guzman,

315

“Robotic Systems for Deployment of

Explosive Detection Sensors and

Instruments”

9. “FAS Public Interest Report: The

Journal of the Federation of

American Scientists”

10. Lisa Thiesan, David Hannum, Dale

W. Murray, John E. Parmeter (2004)

“Survey of Commercially Available

Explosives, Detection techn. And

Equipment

11. Journal of lntelligent and Robotic

Systems 3: 305-319, 1990. Kluwer

Academic Publishers. Printed in

the Netherlands

12. Dan Simon (2001) “Kalman

Filtering”

13. Jianguo Jack Wang a*,Jinling Wang,

David Sinclair, Leo Watts “Neural

Network Aided Kalman Filtering

For Integrated Gps/Ins Geo-

Referencing Platform”

14. Greg Welch and Gary Bishop “An

Introduction to the Kalman Filter”

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

316

WARRIOR ROBOTS

A.Sandeep Kumar & S.M.Siddharth

ABSTRACT:

Robotics is the science that makes the

Robot as agent to move around, react to

their environment in pursuit of our goals.

Any electronic system that uses a computer

chip, but that is not a general-purpose

workstation, desktop or laptop computer is

called an embedded system. Such systems

use microcontrollers or microprocessors.

In this paper we discussed about the

implementation of embedded system based

robot in security field. The robots are used

in many real world applications such as in

military sensing, rescuing operations during

disaster management etc. We have analyzed

and worked on this topic and came out with

a project of MILITARY SECURITY

APPLICATION. It holds the capability of

detecting and conveying the information of

presence of human being (including the

image of the person) to the controller. The

controller can decide whether the person is

allowable inside our border or not. The

detection is performed by Pyroelectric Infra

red sensor and it was conveyed using wired

network. The strange person can be allowed

or shot by passing the respective command

to the Robot which is placed along with the

sensor in the border end.

BENEFICIAL IMPROVEMENTS:

The same project can be

implemented with the help of wireless

communication technologies so that it may

enable the control from distant place. More

over, when the project’s size is reduced and

with more improved features such as sensing

and rescuing facility etc., it may be also

employed in disaster management

applications. The size can be reduced with

the help of advance technology such as

NANOTECHNOLOGY.

INTRODUCTION:

EMBEDDED SYSTEM:

An embedded system is some combination

of computer hardware and software that is

specifically designed for a particular kind of

application device. Industrial machines,

317

automobiles, medical equipment, cameras,

household appliances, airplanes, and toys

are among the possible hosts of an

embedded system.

DEFINITION OF ROBOTICS:

The science and technology that deals with

robots are called robotics (or) Robotics is

the science or study of the technology

associated with the design, fabrication,

theory, and application of robots.

Robots and robotic technologies have an

intellectual and emotional appeal that

surpasses any other type of engineered

product. This appeal has attracted the

inquisitive minds of children and young

adults.

ROBOTS:

Robots are mechanical devices that operate

automatically. They can perform a wide

variety of tasks; jobs that are repetitious and

boring, difficult, or too dangerous for people

to perform. Industrial robots efficiently

complete routine tasks such as drilling,

welding, assembly, painting, and packaging.

They are used in manufacturing, electronics

assembly, and food processing.

A typical robot completes its task by

following a set of specific instructions that

tell it what and how the job is to be

completed. These instructions are

programmed and stored in the robot's

control center, a computer or partial

computer.

These new generation robots are controlled

by both their stored instructions (software

programs) and by feedback that they receive

from the sensors. Such robots might be used

on the ocean floor at depths man is unable to

reach and in planetary exploration and other

scientific research.

TYPES OF ROBOTICS:

Analog Robotics:

Analog robotics indulges the analog

robots.

Analog robot is a type of robot which uses

analog circuitry to go toward a simple goal

such as finding more light or responding to

sound. The first real analog robot was

invented in 1940.

Autonomous robotics:

A fully autonomous robot has the ability to

Gain information about the

environment.

Work for an extended period without

human intervention.

318

Move either all or part of itself

throughout its operating environment

without human assistance.

Avoid situations that are harmful to

people, property, or itself unless

those are part of its design

specifications.

NANOROBOTICS:

Nanorobotics is the technology of creating

machines or robots at or close to the

microscopic scale of nanometers (10-9

meters). More specifically, nanorobotics

refers to the still largely hypothetical

nanotechnology engineering discipline of

designing and building nanorobots.

Nanorobots (nanobots, nanoids or nanites)

would be typically devices ranging in size

from 0.1-10 micrometers and constructed of

nanoscale or molecular components. As no

artificial non-biological nanorobots have yet

been created, they remain a hypothetical

concept.

OUR PROJECT:

Our project titled “warrior Robot”

is equipped with PIR sensor, which is

capable of sensing live human being alone.

Once live person is detected nearby, robot

sends a logic 1 signal to Microcontroller

which in turns activates the weapon.

Parallely the video camera fixed at

head of robot captures the video information

and transmits to remote controller.

EMBEDDED SYSTEM:

Embedded systems are the computing

systems embedded within electronic

devices.

AIM OF THIS PROJECT:

o To replace the armed men in the

country border.

o So that the precious human life may

be prevented from death.

o To withstand extreme climatic

conditions or so intensive

environment.

o Implementing the advanced

technology for proper security

purpose.

319

BLOCK DIAGRAM OF WARRIOR

ROBOT:

320

PIR SENSOR:

Pyroelectric infrared sensor to detect

infrared energy radiation from human body.

It can detect the human presence (like

security alarm) in the range up to 500cm.

THEORY OF OPERATION:

Infrared radiation exists in the

electromagnetic spectrum at the wave length

that is longer than visible light .objects that

generate infrared radiation includes animals

and the human body, the infrared radiation

generated by human is strongest

This PIR sensor is designed specifically to

sense the infrared radiation emitted from

human body.

FEATURES OF PIR SENSOR:

Human infrared radiation detection

Human presence detection up to 5 meters

Human motion direction up to 1.5 meters

APPLICATIONS OF PIR SENSOR:

Human moving direction measurement

Human-following device

Human avoidance and security robot

MICROCONTROLLER

CONTROLLER FEATURES:

4K Bytes of In-System Reprogrammable

Flash Memory-

Endurance: 1,000 Write/Erase Cycles

Fully Static Operation: 0 Hz to 24 MHz

Three-Level Program Memory Lock

128 x 8-Bit Internal RAM

32 Programmable I/O Lines

Two 16-Bit Timer/Counters

321

Six Interrupt Sources

Programmable Serial Channel

MOTOR DRIVERS:

The dc motors are used for the wheel

movement

Four relay circuits drive the motors

The relay circuits are activated with

respect to the command from the

controller.

CONTROL FOR WHEELS

ROBOT MOVEMENT:

To move the robot left, the left motor is

not powered and the right motor is

made to rotate

To move the robot right the right motor

is not powered and the left motor is

made to rotate.

The motor rotation is governed by the

control from the processor (computer)

through the controller and the relay

circuit.

WEAPON DRIVER:

The firing gun is fixed at the robot

terminal

The control to the weapon is given from

the processor through the separate relay

circuit

RELAY CIRCUIT

CONTROL TO WEAPON DRIVER:

At the robot terminal a camera is fixed,

which sends the image of the

surrounding environment

The visuals from the camera is directly

sent to the processor(computer)

If any person is identified, the weapon

is fired depending on the respective

command from the processor through

the controller and the relay circuit.

HARDWARE REQUIREMENT:

1. Robot Chassis

2. AT89s8252 – 8 bit Microcontroller

3. Stepper motor driver

4. Video Transmitter

5. Computer with windows 9(x) OS

SOFTWARE REQUIREMENT:

322

1. Visual basic 6.0

WORKING:

1. The PIR sensors detect the human

being and send the signal to the

controller.

2. The controller transmits the signal to

the computer.

3. The computer also receives the video

signal from camera.

4. By using the computer we can make

the firing equipment to fire or to

move front, back, to the left or right

through relay circuit.

ADVANTAGES OF THIS ROBOT:

It fulfills the necessity of

presence of human warrior

(the human can’t be alert for

all 24 hours)

It can be implemented at low

cost

It is smaller in size.

PHOTO FLIPS OF OUR PROJECT EMBEDDED ROBO:

VISUAL EQUIPMENT (CAMERA)

323

OVERVIEW OF THE ROBOT:

324

CONCLUSION:

Using robots in defence prevents

precious Human life from death

By using Artificial Intelligence ,we can

develop robots which can be used for

different purposes in defence

“NECESSITY IS THE MOTHER OF

INNOVATION!!!”

We need more. So innovate more

REFERENCES:

1. WWW.ATMAL.COM

2. WWW.KEIL.COM

SSOOFFTTWWAARREE--FFRROONNTT EENNDD

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

325

PERFORMANCE ANALYSIS OF ATM NETWORKS USING

MODIFIED PARTICLE APPROACH

M.Sundarambal

#1, M.Dhivya

#2, L.Nitisshish Anand

#3,

#1 Department of Electrical and Electronics Engineering,

Coimbatore Institute of Technology, Coimbarore-641014.INDIA.

#2

[email protected],

Department of Electrical and Electronics Engineering,

#3

Department of Mechanical Engineering

Anna University Coimbatore, Coimbatore-641047.INDIA.

Abstract:

This paper presents generalized particle approach

to optimize the bandwidth allocation and QoS parameters

in ATM networks. The generalized particle approach

algorithm (GPAA) transforms the network bandwidth

allocation problem into dynamics and kinematics of

numerous particles in a force-field. In ATM networks

optimization, the GPA deals with a variety of random and

emergent phenomena, like congestion, failure, and

interaction. The bandwidth allocated, success rate and the

shortest path determined by GPAA are compared with the

same parameters determined using Max-Min fairness

(MMF) algorithm. The comparison among the algorithms gives the effective performance analysis of ATM networks.

1. INTRODUCTION

ATM network is to support multiple classes of traffic

(e.g. video, audio, and data) with widely different

characteristics and Quality of Service (QoS) requirements. The

major challenges are guaranteeing the QoS required for all the

admitted users and dynamically allocating appropriate

resources to maximize the resource utilization. Many

algorithms and evaluation criteria for the bandwidth allocation

have been proposed are based on maximizing an aggregate

utility function of users, an average priority or average

reliability of traffics; minimizing the overhead of data transmission and the highest call blocking Probability of all

source-destination pairs. In some duality models the network's

optimization problem can be solved by a primal Parallel

algorithm for allotted rates and a dual parallel algorithm for

shadow prices or congestion signals [1].

In this paper GPAA is used to determine bandwidth to

be allocated based on pricing and shortest path. MMF is used

to determine bandwidth to be allocated based on fairness and

pricing. The success rate, shortest path are also calculated

using both methods separately and compared.

2. GENERALIZED PARTICLE APPROACH

The GP model consists of numerous particles and forces,

with each particle having its own dynamic equations to

represent the network entities and force having its own time-

varying properties to represent various social interactions

among the network entities. Each particle in GP has four main

characteristics [2]:

1. Each particle in GP has an autonomous self-

driving force, to embody its own autonomy and the personality of network entity.

2. The dynamic state of every particle in GP is a

piecewise linear function of its stimulus, to guarantee a stable

equilibrium state.

3. The stimulus of a particle in GP is related to its own

objective, utility and intention, and to realize the multiple

objective optimizations.

4. There is variety of interactive forces among particles,

including unilateral forces, to embody various

social interactions in networks.

The bandwidth allocation policy of a network link depends on the payment policy of a network path. It changes

dynamically according to the situation of the demands and

supplies of bandwidths and influences the kinematics and

dynamics of the link particles and path particles in their force-

fields. The link particles move in resource force-field FR and

the path particles that require bandwidth assignment move in

demand force-field FD as shown in figure1. Normally a

network path pays less amount under the condition that its

bandwidth requirement is satisfied. It is embodied in the

proposed model by the corresponding path particle. Similarly

to maximize the price benefit a link tries to assign its own

bandwidth to the path that pays maximal price.

3. MODEL FOR BANDWIDTH ALLOCATION

3.1 EVOLUTION OF GP:

The mathematical model based on GP [1] for dynamic

bandwidth allocation that involves m links, n paths and p

channels is defined with following notations:

ri : The maximum bandwidth of link Ai.

d j : The bandwidth that channel T j requires.

q j : The maximum payoff that channel T j can afford for its d j. Ai : The ith physical link in the network.

Tk : The kth path of the jth channel T j.

a ik : The bandwidth that link Ai allots to path Tk.

P ik : The price per unit bandwidth that path Tk tries to

Pay link Ai.

326

3.1.1Total Utility: u k(t) is the distance from the current position of the

path particle T k to the upper boundary of the demand force-

field FD at time t, and let JD(t) be the total utility of all the path

particles in FD. uk(t) and JD(t) are defined respectively as stated

below;

m

u k(t) =δ 1 exp[- ∑ a ik(t) / p ik(t) ] (1) i=1

p n

J D (t) = α1 ∑ ∑ u k(t) (2) j=1 k=1

u i (t) is the distance from the current position of the link particle Ai to the upper boundary of the resource Force-field

FR at time t, and let JR(t) be the total utility of all the link

particles in FR. ui(t) and JR(t)are defined respectively as stated

below;

p n

u i(t) =δ 2 exp [- ∑ ∑ p ik(t) / a ik(t) ] (3) j=1 k=1

m

J R (t) = α2 ∑ u i(t) (4)

i=1 δ 1, δ 2 >1,and 0< α1, α2 <1.

3.1.2 Potential Energy: At time t, the potential energy functions, PD(t) and

PR(t), that are caused by the upward gravitational forces of the

force-fields, FD and FR, are defined respectively as stated

below;

p n

PD(t) = ε 2 ln[- ∑ ∑ exp [( u k(t))2 / 2 ε 2] - ε 2 ln(n) (5)

j=1 k=1

m

PR(t) = ε 2 ln ∑ exp [( ui (t))2 / 2 ε 2] - ε 2 ln(m) (6)

i=1

At time t, the potential energy functions, QD(t) and QR(t),

that are caused by the interactive forces among the particles in

FD and FR are defined respectively as stated below;

p n

Q D (t)= β1 ∑ [∑ a ik(t) – d j(t)]2 + E D(t) j=1 k=1

p n m

+ ρ∑ [ ∑ ∑ a ik(t) p ik(t) – q j(t)]2 (7) j=1 k=1 i=1

m n p

Q R (t) = β2 ∑ [∑ ∑ a ik(t) – r j(t)]2 + ER(t) (8) i=1 k=1 j=1

0< ε<1, 0< β1 β2, ρ<1;

ED(t) and ER(t) are the potential energy functions that involve

other kinds of the interactive forces among the particles in

FD and FR, respectively.

3.1.3 Stability:

Dynamic equations for path particle T k and

link particle Ai are defined respectively as stated below;

du k (t)/dt =φ1(t)+φ2(t) (9)

φ1(t) = - u k(t) + γ1 v1 (t) (9a)

φ2(t) =-[ε1 +ε2 dJ D(t)/d u k(t) + ε3dPD(t)/du k(t)

+ ε4 dQD(t)/d u k(t)] *

m

∑ [d u k(t) / dp ik(t)]2 (9b)

i=1

(OR)

du i (t)/dt = ψ1(t)+ ψ 2(t) (10)

ψ 1(t) = - u i(t) + γ2 v2(t) (10 a)

ψ 2(t) = -[ λ 1 + λ 2 dJ R(t)/d u i(t)+ λ 3 dPR(t)/d ui (t)

+ λ 4 dQR(t)/d u k(t) ] *

n p ∑ ∑ [d u i (t) / da ik(t)]

2 (10 b)

k=1 j=1

v1(t) and v2(t) is a piecewise linear function. γ1, γ2 >1; 0< λ 1, λ 2, λ 3 ,λ 4, ε1, ε2, ε3 ,ε4,<1

dp ik(t)/dt = - ε1[ d u k(t) / dp ik(t)] - ε2[dJ D(t)/dp ik(t)]

-ε3[dPD(t)/dp ik(t)] - ε4[dQD (t)/dp ik(t)]

(11)

da ik(t)/dt = - λ 1[ d u i(t) / da ik(t) - λ 2[dJ R(t)/da ik(t)]

-λ 3[dPR(t)/ da ik(t)] - λ 4[dQR (t)/da ik(t)]

Figure1: GPAA to optimize bandwidth allocation- a) Demand force- field FD for path particles.

b) Resource force-field FR for link particles.

327

(12)

3.1 GP ALGORITHM:

Based on the above mathematical model the algorithm is

written as given below.

Input:

Maximum bandwidth ri , required bandwidth d j ,

Maximum payoff q j

Output:

1. Initialization:

At time t=0;

Initiate bandwidth a ik(t) and price p ik(t)

2. calculate utility, potential energy functions as per the

Equations (1) to (8).

3. If

du k (t)/dt =0; (Or) du i (t)/dt =0;

hold for every particle ,finish with success;

Else goto step 4.

4. Calculate du k(t)/dt and update u k(t)

calculate dui (t)/dt and update ui(t)

calculate da ik(t)/dt ,

a ik(t) = a ik(t -1) + da ik(t)/dt

calculate dp ik(t)/dt,

p ik(t) = p ik(t -1)+ dp ik(t)/dt

goto step 2.

A network with 7 nodes and 11 edges as exhibited in

Figure 2: is considered with the set of node pairs:

V =(v1,v7),(v2,v6),(v4,v6),(v1,v3),(v1,v5)

Figure 2: A network with 7 nodes and 11 edges

The communication bandwidth bij requested between given

set of (v=vi,vj) network node pairs are:

B =15,25,21,20,17

The virtual path sets are:

T18 =(e8,e9),(e1,e2,e3),(e1,e11,e9),(e1,e2,e12,e9)

T27 = (e2, e3,e4),(e11,e10,e5),(e11,e9,e4),

(e2, e12,e10,e5), (e2,e12,e9,e4)

T47 =(e10,e5),(e9,e4)

T16 =(e7,e6),(e8,e10),(e1,e11,e10)

T26 =(e11,e10),(e2,e12,e10)

And the set of links with maximum allowable bandwidths are

e1=10 ;e2=8; e3=3; e4=18; e5=19; e6=8; e7=12; e8=15;

e9=4; e10=6;e11=15;e12=20

The minimal price, bandwidth requested, maximal price,

maximum bandwidths of link are given as input, to the GPA

algorithm. The QOS index required by every virtual path

connection (VPC) is generated randomly within 0.5 to 1.5.

4. MAX-MIN FAIRNESS ALGORITHM:

The Max-Min fairness is computed via an iterative

procedure. The algorithm classifies every virtual session as

either completely utilized or not[3].

The allocated rate for session P is denoted by rp and

the allocated flow on network is[4];

Fa= ∑ δp(a) r p

pεP

δp(a)=1 if a is on path P and 0 otherwise.

r p>=0, for all p ε Pk

Fa<=Ca, for all a ε A.

Ca is capacity of arc a and r is vector of allocated rates.

4.1 MMF ALGORITHM :

Initial Conditions:

K=1,where k is the iterative index.

Capacity Fa =0; rate r p=0; P1=P; A1=A;

1. nk =number of paths ; p ε Pk with δp(a)=1

2. rate vector r- k = min(Ca-Fa(k-1))/nk

3. r p = r p(k-1)+ r- k for p ε Pk

Else

r p = r p(k-1)

4. Fa = ∑ δp(a) r p

pεP

5. A(k+1) = a|Ca-Fa > 0

6. P(k+1) = p| δp(a)=0,for all ¢ A(k+1)

7. k = k+1

8. If P k is empty, then stop; else goto 1.

The above algorithm terminates and finds the max-

min fairness vector r, if it exists, within k steps.

Initially all sessions are unsaturated, and their status

change from unsaturation to saturation. A session is allocated

a rate r p equal to minimum of the link bandwidth on its path.

Initially bandwidth allocated to the link is one third of the

minimum bandwidth of link on its session P. The algorithm

checks the saturation condition and updates the bandwidth till

the session is saturated. The algorithm terminates if all the

sessions are saturated.

For the network in figure 2: the bandwidth requested,

minimal price (rp) and the allowable bandwidth of link (Ca)

are given as input to MMF algorithm.

5. PERFORMANCE ANALYSIS:

The Generalized particle approach (GPAA) and Max-

Min fairness algorithm is implemented with MATLAB 7.0.1.

The solution obtained is given below.

TABLE 1: BANDWIDTH ALLOCATION

Path

path set

Bandwidth allocation

GPAA MMF

12

2

1 8 4

5

3 6 7

10

8

8 19

18 6

14 4

20

15 3

328

T18

e8,e9 8.2500 4

e1,e2,e3

3.7125 3

e1,e11,e9 1.6706 4

e1,e2,e12,e9 1.3669 4

T27

e2,e3,e4

13.7500

3

e11,e9,e4

6.1875

4

e11,e10,e5 2.7844 6

e2,e12,e9,e4 1.2530 4

e2,e12,e10,e5 1.0252 6

T47

e9,e4 11.5500 4

e10,e5 9.4500 6

T26 e11,e10 9.3500 6

e2,e12,e10 7.6500 6

T16

e7,e6 11.0000 8

e8,e10 4.9500 6

e1,e11,e10 4.0500 6

In Table1: the performance comparisons of

bandwidth allocation for paths are given. Every path set in a path is analyzed separately using both algorithms.. The

amount of bandwidth allocated is tabulated respectively. The

requested bandwidth is distributed among the path sets

according to the respective algorithms.

In figure 3: and 4: the bandwidth allocation for path 1

and path2 are given. For path1 the bandwidth allocation using

Max-Min fairness algorithm gives fair allocation. For path 2

the bandwidth allocation using Max-Min fairness algorithm

doesn’t meet the requirement. The GPAA exhibits much better

performance than the MMF in terms of bandwidth allocation

for path2.

In figure 5: the success rates of both the algorithms are compared. The success rate is computed as the difference

between requested bandwidth and allocated bandwidth. The

success rate of GPAA is better compared to Max-Min fairness

algorithm. For path 2 and 3 the success rate is very less in

Max-Min fairness algorithm.

Figure 3: Illustrates bandwidth allocation for path1, where fairness is

obtained in MMF and priority analysis is seen in GPAA.

1 2 3 4 50

2

4

6

8

10

12

14

path set no

bandw

idth

allocation (

Mbps)

bandwidth allocation for path 2

MMF

GPAA

Figure 4: Illustrates bandwidth allocation for path 2; where in GPAA the need

is met and in MMA the allocation is not satisfied.

1 2 3 4 50.4

0.5

0.6

0.7

0.8

0.9

1

path no

success r

ate

(%)

path no VS success rate

MMF

GPAA

Figure5: the comparison of GPAA and MMF algorithm with success rate.

In figure 6: the price allocation for every path set of path 3 is

analyzed. For shortest path the price allocated is low.

1 2 3 4 50.5

0.55

0.6

0.65

0.7

0.75

0.8

0.85

0.9

path set no

price b

oundary

PRICE ALLOCATION FOR PATH 2

Figure 6: The satisfactory degree of path 2 through GPAA.

Taking 100 no of particles of size 10*10; price distribution

and utility distribution are computed as shown in figure 7: and

8: respectively.

329

0

5

10

0

5

100.5

1

1.5

resource no

Price distribution of 10 * 10 particles

user no

pric

e di

strib

utio

n

Figure 7: The particles distribution within the price boundary.

0

5

10

0

5

101

2

3

utilit

y d

istr

ibution

Utility distribution of 10 *10 particles

resource nouser no

Figure 8: The total utility value for 10*10 particles

In figure 8: the utility distribution of all the particles at

the final stage of GPAA execution is illustrated.

The routing optimality can be determined with the shortest

path and the feasible bandwidth allocated for that.

In Table 2: The performance comparisons of shortest path, success rate, net bandwidth allotted are given. It

illustrates that the algorithm GPAA exhibit better performance

than the MMF algorithm in terms of bandwidth allocation and

success rate, whereas they have approximately same shortest

paths.

TABLE 2: OPTIMAL ANALYSIS OF PARAMETERS

Path

Requested

Bandwidth

Network

Bandwidth

Utilization

Success rate Shortest path

GPAA MMF GPAA MMF GPAA MMF

T18 15 15 15 100% 100% e8,

e9

e8,

e9

T27 25 25 23 100% 92%

e2,

e3,

e4

e11,

e9,

e4

T47

21

21

10

100%

47%%

e9,

e4

e10,

e5

T2 6

17

17

12

100%

70%

e11,

e10

e11,

e10

T16

20

20

20

100%

100%

e7,

e6

e7,

e6

6.CONCLUSION: In this paper a generalized particle approach

algorithm is used to estimate the bandwidth allocation in ATM

networks. The performance evaluation using GPAA is

compared with MMF algorithm and found better in terms of

success rate, network bandwidth utilization, price allocation

and quality of service. In future, congestion factor, breakdown

factor, fairness factor can be incorporated in GPA model to

optimize the bandwidth allocation.

REFERENCES:

[1] Dianxun Shuai, Xiang Feng, Francis C.M.Lau, A new generalized

Particle approach to parallel bandwidth allocation. Computer

Communications July 2006 3933-3945.

[2] Dianxun Shuai, Yuming Dong, Qing Shuai, Optimal control of

Network services based on Generalised particle model.

[3] Saswati Sarkar, Leandros Tassiulas,Fair Distributed congestion control

in multirate multicast networks IEEE/ACM Transactions on

Networking vol.13.no1, February 2005

[4] Dimitri Bertsekas, Robert Gallager, Data Networks. Englewood cliffs

NJ: Prentice Hall,1987

[5] Ammar W.Mohammed and Nirod Chandra sahoo,

Efficient computation of shortest paths in networks using Particle

Swarm optimization and Noising metaheuristics ,Hindawi publishing

Corporation ,volume 2007,Article Id 27383,

[6] Bozidar Radunovic,A unified framework for Max-Min and Min-max

Fairness with applications,july2002.

[7] Chang Wook Ahn,R.S.Ramakrishna, A Genetic Algorithm for Shortest

Path routing problem and the sizing of populations.

IEEE/ Transactions on Evolutionary Computation, vol, 6, N0:6,

December

2002.

[8] Particle Swarm optimization Tutorial

www.swarmintelligence.org/tutorials

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology,Sathyamangalam- 638401

9-10 April 2010

331

CURTAIL THE COST AND IMPACT IN WIRELESS SENSOR

NETWORK

T.MUNIRAJ

M.E, Communication Systems

Anna University Coimbatore.

[email protected]

K.N.VIJEYAKUMAR

Lecturer, Dept of ECE,

Anna University Coimbatore.

[email protected]

ABSTRACT The paper presents an original integrated MAC and

routing scheme for wireless sensor networks The design objective is to elect the next hop for data

forwarding by jointly minimizing the amount of

signaling to complete a contention and maximizing

the probability of electing the best candidate node.

Toward this aim, to represent the suitability of a

node to be the relay by means of locally calculated

and generic cost metrics. Based on these costs, and

analytically model the access selection problem

through dynamic programming techniques, which

we use to find the optimal access policy. Hence, the

paper proposes a contention-based MAC and

forwarding technique, called Cost- and Collision-Minimizing Routing (CCMR). This scheme is then

thoroughly validated and characterized through

analysis, simulation, and experimental results.

Index Terms—Routing protocols, distributed

applications, algorithm/protocol design and

analysis, dynamic programming, wireless sensor

networks

1. INTRODUCTION

FORWARDING algorithms for Wireless Sensor

Networks (WSNs) should be simple, as sensor nodes

are inherently resource constrained. Moreover, they

should also be efficient in terms of energy

consumption and quality of the paths that are used to

route packets toward the data gathering point

(referred to here as sink). A trend in recent research

is to select the next hop for data forwarding locally

and without using routing tables. Such a localized

neighbor election is aimed at minimizing the

overhead incurred in creating and maintaining the

routing paths. Often, nodes are assumed to know

their geographical location. Such a knowledge can be

exploited to implement online routing solutions

where the next hop is chosen depending on the

advancement toward the sink. However, in addition

to the maximization of the advancement, other

objectives such as the maximization of residual

energies should be taken into account. The schemes

in are localized MAC/routing algorithms (LRAs),

where nodes only exchange information with their

one-hop neighbors (i.e., the nodes within

transmission range). This local information exchange

is essential to achieve scalability, while avoiding the

substantial communication costs incurred in

propagating path discovery/update messages.

GeRaF is an example of a geographical

integrated MAC and routing scheme where the

forwarding area (whose nodes offer geographical

advancement toward the sink) is subdivided into a

number of priority regions. The next hop is elected

by means of a channel contention mechanism, where

the nodes with the highest priority (i.e., closest to the

sink) contend first. This has the effect of reducing

the number of nodes that simultaneously transmit

within a single contention, while increasing the

probability of electing a relay node with a good

geographical advancement. The authors in propose

Contention-Based Forwarding (CBF). In their

scheme, the next hop is elected by means of a

distributed contention. CBF makes use of biased

timers, i.e., nodes with higher advancements respond

first to contention requests. The value of the timers is

determined based on heuristics. A similar approach is

exploited in [3], where the authors propose Implicit

Geographic Forwarding (IGF). This technique

accounts for biased timers as well. Response times

are calculated by also considering the node’s residual

332

energy and a further random term. Advancements,

energies, and random components are encoded into

cost metrics. The random term improves the

performance when multiple nodes have similar costs.

To improve the performance of LRAs by presenting

the concept of partial topology knowledge

forwarding. Sensors are assumed to know the state of

the nodes within their communication range (called

knowledge range in [4]) only. Their goal is to

optimally tune, based on the local topology, the

communication range (local view) at each sensor in

order to approach globally optimal routing. Reference

[5] proposes the MACRO integrated MAC/ routing

protocol. This is a localized approach relying on

priority regions as [1] by, in addition, exploiting

power control features for improved energy

efficiency. A common denominator among these

forwarding schemes is that they are all based on some

sort of cost metrics, which are locally computed, and

take into consideration the goodness of a node to be

elected as the relay

To analytically characterize the

joint routing and relay election problem,

finding optimal online

policies.

To use these results for the design

of a practical solution for WSNs, which

we call CCMR.

CCMR is compared against state-

of-the-art solutions belonging to the

same class of protocols, showing its

effectiveness.

To describe the software

implementation of our algorithm and

present experimental results to

demonstrate the feasibility of our

approach.

The algorithm can be seen as a generalization of

[1], [5], and [6], as contentions are carried out by

considering cost-dependent access probabilities

instead of geographical [1] or transmission power-

aware [5] priority regions. Moreover, the

optimization performed in the present work is a

nontrivial extension of the approaches in [6] and in

[12]. In particular, the channel contention follows an

optimization process over multiple access slots and,

for each slot, over a two- dimensional cost-token

space (justified and formally presented in Section 2.1

this considerably improves the performance of the

forwarding scheme. In addition, the contention

strategy we devise here is optimal rather than

heuristic, and we add a new dimension to carry out

the optimization (i.e., the node “cost,” to be defined

shortly). Also, as our solution provides a method to

locally and optimally elect the next hop for a given

knowledge range (transmission power), we note that

it can be readily coupled with previous work [4].

Finally, our technique can be used in conjunction

with advanced sleeping behavior algorithms [10],

[11]. This is possible due to the stateless nature of

our scheme, which makes it well adaptable to system

dynamics. More specifically, we present an original

forwarding technique coupling the desirable features

of previous work, such as the local nature of the next-

hop election and the definition of suitable cost

metrics, with optimal access policies

2. ANALYTICAL FRAMEWORK

2.1 Cost Model

In this section, we introduce a simple analytical cost

model that we adopt to design our scheme. In doing

so, we explicitly account for the correlation among

costs, as this parameter affects the optimal channel

access behavior the most. In the next sections, such a

cost model is used to derive the optimal access policy

and to design an integrated channel access/routing

protocol. In Section 4, simulation results are given to

demonstrate the validity of the approach in the

presence of realistic costs, depending on

geographical advancements and energy levels.

Further, in Section 5, we show experimental results

where the cost is associated with geographical

advancements and is used to implement a greedy

geographical routing scheme. Let us consider a

generic set SN of N nodes, where we refer to ck as

the cost associated with node k 2 SN. In order to

model the cost correlation, we assume that ck is

given by ck ¼ c þ _k, where c is a cost component

common to all nodes, whereas _k is an additive

random displacement uniformly distributed in ½__c;

_ð1 _ cÞ_, _ 2 ½0; 1_, and independently picked for

every node k. c is drawn from a random variable with

domain in [0, 1] and whose exact statistics depends

on the specific environment where the sensors

operate. With the above model, _ ¼ 0 corresponds to

the fully correlated case as all node costs collapse to

c. Conversely, _ ¼ 1 gives the i.i.d. case ð_ ¼ 0Þ as

333

the costs of every pair of nodes in SN are

independent. Intermediate values of lead to a

correlation _ 2 ð0; 1Þ. The (linear) correlation

coefficient between the costs of any two nodes r; s 2

SN is defined as _r;s ¼ ðE½crcs_ _

E½cr_E½cs_Þ=ð_r_sÞ, where _2s ¼ E½ðcs _

E½cs_Þ2_, cr ¼ c þ _r, and cs ¼ c þ _s. Hence, the

correlation coefficient is given by

As an example, if c is drawn from a uniform distribution in [0, 1] (E½c2_ ¼ 1=3 and E½c_2 ¼ 1=4), (1) becomes

Note that as long as E½c2_ > E½c_2, the correlation can be tuned by varying the parameter _, and as anticipated

above, _ ¼ 0 and _ ¼ 1 lead to the fully correlated and to the i.i.d. case, respectively. Also, there is a one-to-one

mapping between _ and _ as (1) is invertible.

2.2 State Space Representation and Problem

Formulation

Let us consider the next-hop election

problem for a given node in the network. Such an

election is performed by means of MAC contentions,

which usually consume resources in terms of both time and energy. Broadly speaking, our goal is to let

the relay node by maximizing the joint probability

that a node wins the contention and it has the smallest

cost (or a sufficiently small cost) among all active

neighbors. The formal problem statement is given at

the end of this section. Here, we refer to this election

strategy as optimal. According to our scheme, the

node sends a request (REQ) addressed to all nodes in

its active (or forwarding) set, which is composed of

all active neighbors providing a positive

advancement toward the sink. Upon receiving the

REQ, the active nodes in this set transmit a reply (REP) by considering a slotted time frame of W slots.

Specifically, each node picks one of these slots

according to its cost and uses it to transmit a REP to

the inquiring node. The first node to send a REP

captures the channel so that the nodes choosing a

later slot refrain from transmitting their REPs

.

To model the above scheme and find the

optimal slot election strategy under any cost

correlation value, we proceed as follows: For each

node, we account for a cost and a token, the latter being a random number that is uniformly picked in

[0, 1] at every contention round. Tokens are used to

model cost-unaware access probabilities [13]. In

more detail, when costs are fully correlated ð_ ¼ 1Þ,

the nodes should pick the access slots by only

considering their tokens, as their costs are all

equivalent by definition. In this case, the aim of the

algorithm is to select any node in the forwarding set

by maximizing he probability of having a successful

contention, and the solution reduces to the one in

[12]. On the other hand, when costs are completely uncorrelated ð_ ¼ 0Þ, tokens must be disregarded,

and the slot selection should be made on the basis of

the node costs only. Finally, if the cost correlation is

in (0, 1), both costs and tokens should be taken into

account in the selection of the access slot. In addition,

in order to simplify the problem formulation, access

probabilities can be expressed in terms of access

intervals as we explain next. For illustration, consider

the case where _ ¼ 1, i.e., only tokens are accounted

for in making access decisions. In this case, at any

given node and for a given slot, accessing the channel

with a given probability p is equivalent to accessing the channel if the token is within the interval ½0; p_.

When _ ¼ 0, the same rationale can be used for the

costs, by defining intervals in the cost space. In the

most general case ð_ 2 ð0; 1ÞÞ, we can define

rectangular access regions spanning over both costs

and tokens. For the sake of explanation, we illustrate

the concept by means of Fig. 1, where we plot an

access slot selection example for W ¼ 4 slots.

Fig. 1. Example of access region and node

representation for W ¼ 4.

A formal treatment is given in Section 2.3 The active

set is composed of the three nodes n1, n2, and n3,

which are plotted in the token-cost space by means of

white-filled circles. We associate the access regions

R1, R2, R3, and R4 with the access slots 1, 2, 3, and

4, respectively Note that R1 _ R2 _ R3 _ R4; this

334

Fig. 2. cifor N ¼ 10, W ¼ 10, and c ¼ 0:5.

property holds in the general case; see

Section 2.3. For the slot selection, each node picks

the access slot corresponding to the smallest region

containing its (cost, token) pair. Specifically, node n1

cannot transmit in any of the slots as it is not within a valid access region. Moreover, in the first slot, none

of the remaining nodes can access the channel. In

fact, R3 is the first region containing a node, n2 in

our example, which therefore sends its REP in the

third slot. Note that according to our slot selection

strategy, n3 would be allowed to transmit its REP in

slot 4. However, it refrains from transmitting the REP

In this slot as it senses the ongoing communication

of node n2. In this example, a single node (n2)

accesses the channel, and this is the node with the minimum cost in the active set. We observe that

collisions (multiple nodes select the same slot) are

possible. Moreover, although it could also be possible

that the winner of the contention is not the node with

the minimum cost, our solution is aimed at

minimizing the probability of occurrence of this

event.

Optimal Access Schedules: Discussion of Results

As a first result, in Fig. 2, we show the probability

’ð1;R0Þ of having a successful contention round

using the optimal policy, by averaging over c (uniformly distributed in [0, 1]). The parameters for

this figure are " ¼ 0 and _ ¼ 0:5. Perfect knowledge

is assumed at the transmitter for the number of

contenders N, the cost correlation _, and c. As

expected, ’ð1;R0Þ increases with an increasing

number of access slots W: increasing W from 2 to 10

almost doubles the performance, whereas further

increasing it (10 ! 20) only provides marginal

improvements. Also, for a given W, the success

Probability quickly stabilizes ðN _ 10Þ to its

asymptotic value. obtain a reasonable trade-off between complexity and effectiveness

Fig. 3. cifor N ¼ W ¼ 10, c ¼ 0:5, and " ¼ 0:1

. In Figs. 3 and 4, we plot the optimal access regions

for costs and tokens, respectively. Notably, the value

of _ does have an impact on the shape of the regions.

In practice, the case _ ¼ 0 is the most selective in the

sense that high costs, for any given slot, are penalized

the most. Also, we observe that for _ ¼ 1, all costs are equal by construction, and hence, they should not

affect the slot selection process. This is in fact

verified in Fig. 3, where cost regions are all equal to

one for _ ¼ 1. This concept can be remarked by

looking at Fig. 4, where we plot the token regions t i

for the same system parameters. In this case, t i are

equal to one for _ 2 ½0; 1Þ. This means that for these

values of _, the tokens do not influence the slot

selection, which is only driven by the costs. On the

other hand, for _ ¼ 1, costs are no longer relevant to

3. COST-AND COLLISION-

MINIMIZING ROUTING

In this section, based on the previously

discussed results, we present an integrated channel

access and routing scheme that we name as CCMR.

Our cross-layer design relies on the definition of the

costs, which are used in the channel access to

discriminate among nodes. This is achieved by

accounting for routing metrics such as the

geographical advancement, right in the cost

calculation. Realistic cost models are presented in

Section 4, where we report extensive simulation results to validate our approach. Next, we outline our

integrated scheme by considering the costs as give

Consider a generic node n. When the node has a

packet to send, it first senses the channel according to

a CSMA policy, as done in, e.g., IEEE 802.15.4. If

the channel is sensed idle for a predetermined

interval, the contention starts. The contention to elect

the next hop works in rounds and ends as soon as a

round is successful. At the generic round r _ 1, node

n sends a request (REQ) including its identifier and

an estimate for the number of contenders N and specifies a cost interval ½cmin;r; cmax;r_, where

cmin;1 ¼ 0, and cmax;1 ¼ 1. We detail how this

interval is modified for r > 1 in point 3 belo All

active devices providing a positive advancement

335

toward the sink contend for the channel. Upon

receiving the REQ, at round r _ 1, every node

considers W access slots and calculates cost and

token regions as follows: The node first computes a

decay function dðrÞ ¼ r_=ðr_ þ 1Þ depending on the

round number r and on a constant _ > 0. If ðcmax;r _ cmin;rÞ > dðrÞ, the c i s are calculated by means of

(15) and t i ¼ 1 8 i; otherwise, c i ¼ 1 8 I and t i ¼ t

i_1 þ p i ðt 0 ¼ 0Þ, where p i are as in (14) and i 2 f1;

2; . . .;Wg. In other words, dðrÞ is used to estimate

when costs can be assumed to be equal, and

therefore, the corresponding theory for _ ¼ 1 should

be used. We refer to the cost region associated with

the last slot W as c W;r. Subsequently, using these

access regions and its own cost, the node picks a slot

in f1; 2; . . .;Wg according to the scheme in Section 2

and schedules a reply (REP) in this slot.

4. RESULTS

All sensors generate traffic according to a

Poisson arrival process. The network generation rate

net is varied from 0.2 to 1.5 packets per second to

respect the capacity limits we found in the previous

section. The generation rate for a single node is ¼

net=Nu. We repeated at least 10 experiments for each

value of net, where each node generates 100 packets

for each experiment. For the transmission of RTS

messages, we adopted an IEEE 802.11-like CSMA

technique. In particular, in case of a busy medium, we set the bakeoff timer so that new RTSs cannot

interfere with ongoing contentions. Costs are

Fig 4 Nodes Creation.

Packets are transferred by constructing only

three nodes which is shown in NAM output. Packet

receiving output shows the data transferring from

source node to sink node (Destination node) only in a

single path direction. Cost function denotes the

distance between the two nodes. Time function denotes the time taken by packets to reach from one

node to another node.

Packet Receiving

During each experiment, the packets

received at the sink are collected, storing the source

and the number of hops traversed. Also, each node

saves statistics about all its (one-hop) contentions.

The performance metrics we show here are the

delivery rate, defined as the average number of

packets delivered to the sink over the total number of

packets generated, the fraction of duplicate packets reaching the sink, and the distribution of the

contention outcome. This last metric represents the

statistics of the number of contention rounds for the

election of the next hop. Fig.5 shows the delivery

rate, averaged over all nodes, as a function of net.

Both simulation and experimental points are shown

in the plot. To this end, we used the simulator

presented in Section 4, which was configured to

reproduce, as closely as possible, our experimental

setting. Two types of simulations were run. In the

former, nodes are in a real multi hop scenario (referred to as MH in the figure)

Fig 5 Packet Received

336

5.CONCLUSIONS

In this paper, we presented an original

integrated channel access and routing technique for

WSNs. Our design objectives were to minimize the

energy consumption (the number of packets sent for each channel contention), while maximizing the

probability of picking the best (lowest cost) node in

the forwarding set. By analysis, we found an online

algorithm that we called CCMR. We tested this

scheme via simulation, by comparing its performance

against that of state-of-the-art solutions. Besides

proving the effectiveness of the scheme, our results

show its robustness against critical network

parameters. Finally, we demonstrated the feasibility

of implementing CCMR as a lightweight software

module for TinyOS and validated the protocol

experimentally.

6. FUTURE WORK:

Further optimizations of the software implementing

CCMR are possible, including the possibility of using

the protocol as a basis for data aggregation and

distributed network coding techniques.

REFERENCE

[1] M. Zorzi and R.R. Rao, “Geographic Random

Forwarding (GeRaF)

for Ad Hoc and Sensor Networks: Multihop Performance,” IEEE

Trans. Mobile Computing, vol. 2, no. 4, pp. 337-348,

Oct.-Dec. 2003.

[2] H. Fu¨ ßler, J. Widmer, M. Ka¨semann, M.

Mauve, and

H. Hartenstein, “Contention-Based Forwarding for

Mobile

Ad-Hoc Networks,” Elsevier’s Ad Hoc Networks,

vol. 1, no. 4,

pp. 351-569, Nov. 2003.

[3] T. He, B.M. Blum, Q. Cao, J.A. Stankovic, S.H.

Son, and T.F. Abdelzaher, “Robust and Timely

Communication over

Highly Dynamic Sensor Networks,” Real-Time

Systems, vol. 37,

no. 3, pp. 261-289, Dec. 2007.

[4] T. Melodia, D. Pompili, and I.F. Akyildiz,

“Optimal Local

Topology Knowledge for Energy Efficient

Geographical Routing

in Sensor Networks,” Proc. IEEE INFOCOM ’04,

Mar. 2004. [5] D. Ferrara, L. Galluccio, A. Leonardi, G.

Morabito, and S. Palazzo,

“MACRO: An Integrated MAC/ROuting Protocol for

Geographical

Forwarding in Wireless Sensor Networks,” Proc.

IEEE

INFOCOM ’05, Mar. 2005.

[6] M. Rossi and M. Zorzi, “Integrated Cost-Based

MAC and

Routing Techniques for Hop Count Forwarding in Wireless

Sensor Networks,” IEEE Trans. Mobile Computing,

vol. 6, no. 4,

pp. 434-448, Apr. 2007.

[7] M.C. Vuran and I.F. Akyildiz, “Spatial

Correlation-Based Collaborative

Medium Access Control in Wireless Sensor

Networks,”

IEEE/ACM Trans. Networking, vol. 14, no. 2, pp.

316-329, Apr. 2006.

[8] S. Lee, B. Bhattacharjee, and S. Banerjee,

“Efficient Geographic Routing in Multihop Wireless Networks,” Proc.

ACM MobiHoc ’05,

May 2005.

[9] W. Ye, J. Heidemann, and D. Estrin, “An Energy-

Efficient

MAC Protocol for Wireless Sensor Networks,” Proc.

IEEE

INFOCOM ’02, June 2002.

[10] A. Keshavarzian, H. Lee, L. Venkatraman, D.

Lal, K. Chintalapudi,

and B. Srinivasan, “Wakeup Scheduling in Wireless Sensor

Networks,” Proc. ACM MobiHoc ’06, May 2006.

[11] S. Du, A.K. Saha, and D.B. Johnson, “RMAC: A

Routing-Enhanced

Duty-Cycle MAC Protocol for Wireless Sensor

Networks,” Proc.

IEEE INFOCOM ’07, May 2007.

[12] Y.C. Tay, K. Jamieson, and H. Balakrishnan,

“Collision-

Minimizing CSMA and Its Applications to Wireless

Sensor

Networks,” IEEE J. Selected Areas in Comm., vol. 22, no. 6,

pp. 1048-1057, Aug. 2004.

[13] P. Popovski, F.H. Fitzek, and R. Prasad, “Batch

Conflict

Resolution Algorithm with Progressively Accurate

Multiplicity

Estimation,” Proc. Joint Workshop Foundations of

Mobile Computing

(DIAL M-POMC ’04), Oct. 2004.

[14] E.V. Denardo, Dynamic Programming: Models

and Applications, second ed. Dover Publications, 2003.

[15] S. Ghosh and S.G. Henderson, “Behavior of the

NORTA Method.

Proceedings of the Third National Conference on RTICT 2010 Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

337

DATA ALLOCATION OPTIMIZATION ALGORITHM IN

MOBILE COMPUTING

AISWARYA.P.V, SABIREEN.H

PANIMALAR ENGINEERING COLLEGE

MAIL-ID:[email protected],[email protected]

Abstract-In this paper, we devise data allocation

algorithms that can utilize the knowledge of user

moving patterns for proper allocation of shared

data in a mobile computing system. By employing

the data allocation algorithms devised, the

occurrences of costly remote accesses can be

minimized and the performance of a mobile

computing system is thus improved. The data

allocation algorithms for shared data, which are

able to achieve local optimization and global

optimization, are developed. Local optimization

refers to the optimization that the likelihood of local

data access by an individual mobile user is

maximized whereas global optimization refers to

the optimization that the likelihood of local data

access by all mobile users is maximized.

Specifically, by exploring the features of local

optimization and global optimization, we devise

algorithm SD-local and algorithm SD-global to

achieve local optimization and global optimization,

respectively. In general, the mobile users are

divided into two types, namely, frequently moving

users and infrequently moving users. A

measurement, called closeness measure which

corresponds to the amount of the intersection

between the set of frequently moving user patterns

and that of infrequently moving user patterns, is

derived to assess the quality of solutions provided

by SD-local and SD-global. Performance of these

data allocation algorithms is comparatively

analyzed. From the analysis of SD-local and SD-

global, it is shown that SD-local favors infrequently

moving users whereas SD-global is good for

frequently moving users. The simulation results

show that the knowledge obtained from the user

moving patterns is very important in devising

effective data allocation algorithms which can lead

to prominent performance improvement in a mobile

computing system.

1. INTRODUCTION

Due to recent technology advances, an

increasing number of users are accessing various

information systems via wireless communication.

Such information systems as stock trading, banking,

and wireless conferencing, are being provided by

information services and application providers and

mobile users are able to access such information via

wireless communication from anywhere at any

time. For cost-performance reasons, a mobile

338

computing system is usually of a distributed server

architecture. In general, mobile users tend to submit

transactions to servers nearby for execution so as to

minimize the communication overhead incurred [1],

[2]. Data objects are assumed to be stored at servers

to facilitate coherency control and also for memory

saving at mobile units. Data replication is helpful

because it is able to improve the execution

performance of servers and facilitate the location

lookup of mobile users [3], [4]. The replication

scheme of a data object involves how many replicas

of that object to be created and to which servers

those replicas are allocated. Clearly, though

avoiding many costly remote accesses, the

approach of data replication increases the cost of

data storage and update. Thus, it has been

recognized as an important issue to strike a

compromise between access efficiency and storage

cost when a data allocation scheme is devised. The

data allocation schemes for traditional distributed

databases are mostly designed in static manners,

and the user moving patterns, which are particularly

relevant to a mobile computing system where users

travel between service areas frequently, were not

fully explored. The server is expected to take over

the transactions submitted by mobile users and

static data allocation schemes may suffer severe

performance problems in a mobile computing

system.

For example consider the network topology given

in fig,1 . There are 12 servers and 4 replicated

servers which are shown by slant lines. Suppose

that the shared data are replicated statically at sites

A, F, J, and L under the data allocation schemes for

traditional distributed databases. Assume that the

mobile user U1 is found to frequently travel in

service areas of A and E (i.e., A,E is called the

moving pattern of mobile user U1) and the mobile

user U2 frequently moves in the service areas of A,

B, C, and D. It can be seen that the advantage of

having replicas on F, J and L cannot be fully taken

by mobile users U1 and U2, and the extra cost of

maintaining those replicas is not justified by the

moving patterns of users U1 and U2.

A B C D

E F G H

I J K L

Fig 1. an example network scenario

The extra cost of maintaining those replicas is not

justified by the moving patterns of users U1 and U2.

Similarly to the calling patterns of users, it is

envisioned that most users tend to have their own

user moving patterns since the behaviors of users

are likely to be regular [5]. Since each mobile user

has his / her own moving pattern, in this paper, we

devise data allocation algorithms that can utilize the

knowledge of user moving patterns for proper

allocation of shared data. Explicitly, the data

allocation schemes for shared data, which are able

to achieve local optimization and global

optimization, are developed. Local optimization

refers to the likelihood of local data access by an

individual mobile user is maximized whereas global

optimization refers to the likelihood of local data

access by all mobile users is maximized based on

339

which we devise algorithm SD-local and algorithm

SD-global to achieve local optimization and global

optimization, respectively.

It is assumed that the mobile users who have

regular moving behaviors can be further divided

into two types, namely, frequently moving users

and infrequently moving users.

A measurement, called closeness measure which

corresponds to the amount of the intersection

between the set of frequently moving user patterns

and that of infrequently moving user patterns, is

derived to assess the quality of solutions resulted by

SD-local and SD-global. From the analysis of SD-

local and SD-global, it is shown that SD-local

favors infrequently moving users and SD-global is

good for frequently moving users.

There is a proposed data distribution scheme that is

based on the read/write patterns of the data objects.

Given some user calling patterns, the proposed

algorithm in [6] employed the concept of

minimum-cost maximum-flow to compute the set

of sites where user profiles should be replicated.

The attention of the study in [6] was mainly paid to

the distribution of location data for mobile users.

Schemes for personal data allocation were explored

in [7], whose attention was mainly paid to

developing mining procedure to cope with personal

data allocation, but not for the shared data

allocation explored in this paper.The contributions

of this paper are twofold. We not only devise data

allocation for shared data in a mobile computing

system, but also in light of user moving patterns

obtained, optimize the data allocation schemes

devised.

This paper is organized as follows: Preliminaries

are given in Section 2. Shared data allocation

algorithms based on user moving patterns are

developed in Section 3. Analysis results are

presented in Section 4. This paper concludes with

Section 5.

2. PROBLEM FORMULATION

In this paper, we devise data allocation algorithms

that can utilize user moving patterns to determine

the set of replicated servers for proper shared data

allocation. Since each mobile user has his / her own

moving pattern, how to select proper sites for

shared data allocation is the very problem we shall

deal with in this paper. The problem that we study

in this paper can be formally defined as follows:

Problem of shared data allocation based on user

moving patterns: Given the number of mobile

users with their moving patterns, the number of

servers and the number of replicated servers, we

shall determine the proper set of sites to which

shared data are allocated with the purpose of

maximizing the number of local access of shared

data. In this paper, with the proper allocation of

shared data, the number of local access of shared

data is improved and the properties of data objects

are read-only in order to fully focus our problem on

design the shared data allocation based on user

moving patterns. Table 1 shows the description of

symbols used in modeling the problem.

340

Table 1: Descriptions of symbols used

fig2. Sharing data where number of users are 3

Fig. 2 shows the problem formulation of allocating

shared data where the number of mobile users is 3.

The set of total servers is expressed by S, where |S|

is the total number of servers. Denote the set of

replicated sites for shared data as R. The union set

of moving patterns for mobile user Ui is expressed

by FSi (standing for frequent set), where |FSi| is the

number of distinct sites within the set FSi. The

number of moving paths for mobile user Ui is

denoted by ni, where a moving path is a sequence

of servers accessed by a mobile user.

Clearly, the probability of local access of mobile

user Ui, denoted by

L(Ui) is proportional to | R FSi |

|FSi|

which is formulated as follows:

L(Ui)=f* | R FSi | (1)

|FSi|

where f is a hit coefficient and 0<f<1

Consider the mobile user U1 in Table 2, where the

network topology is shown in Fig. 1. Assume that

without exploring user moving patterns, the set of

replicated sites R = A, F, J, L and the value

f=0.8. From Table 2, the set of FS1 can be obtained

by unifying two moving patterns of mobile user U1

into one set, i.e., FS1= AEABC=ABCE. It

can be verified that the set of R FS1 is A. Then,

we have the estimated probability of local access of

mobile user U1 is 0.8*1/4-=0.2. Since each mobile

user has his/her own moving patterns, how to select

proper sites for shared data allocation, i.e., R, is a

very important issue which will be dealt with in this

paper. To facilitate the presentation, we denote PT

as the threshold to determine whether the mobile

user belongs to the group of frequently moving

users or not.

Table 2: example for data allocation patterns for

users Ui

Frequent set of mobile users Ui FSi

Probability of local access hit for mobile

user Ui

L(Ui)

Set of total servers S

Set of replicated servers R

No. of moving paths for mobile user Ui ni

No. of mobile users N

Threshold value to identify frequent

moving mobile users

PT

Ui Moving

patterns

Frequent

set FSi

No. of

moving

paths

1 AE,ABC ABCE 1500

2 BC,GJ BCGJ 350

3 BCD BCD 300

4 CGF CGF 200

341

The union set of frequent sets of frequently moving

users is defined as

FFS= i,1 i N and ni PT FSi.

The union set of frequent sets of infrequently

moving users is defined as

UFS = i,1 i N and ni < PT FSi.

To quantify how closely FFS approximates UFS, we

use a closeness measure, denoted by C(FFS, UFS

that returns normalized value in [0, 1] to indicate

the closeness between FFS and UFS. The larger the

value of C(FFS, UFS) is, the more closely FFS

approximates to UFS. C(FFS, UFS) is formulated as

follows:

C(FFS, UFS)= |FFS UFS|

|FFS UFS|

For the example profile in Table 2, assuming the

value of PT is 500. U1 is the frequently moving user

(with n1 = 1, 500 movements), and U2, U3, and U4

are infrequently moving users (with n2 = 350, n3 =

300, and n4 = 200 movements, respectively). Also,

the set of FFS is ABCE (i.e., FS1) and the set of

UFS is BCDFGJ (i.e., FS2 FS3 FS4). It can be

verified that the value of C(FFS, UFS) is 0.25 (i.e.,

2/8).As can be seen later, the closeness measure

between FFS and UFS influences the solution quality

resulted by shared data allocation algorithms.

3.SHARED DATA ALLOCATION

ALGORITHMS BASED ON MOVING

PATTERNS

In Section 3.1, we develop two shared data

allocation algorithms, SD-local and SD- global, to

improve the performance of a mobile computing

system. An analysis of Algorithms SD-local and

SD-global is given in Section 3.2.

L2 User

occurrence

count for

sd-local

Movement

Occurrence

count

for sd-global

AB 1 n AB (U1)=800

BC 3 nBC (U1)+ nBC (U2)+ and nBC

(U3)=600

CD 1 nCD (U3)=200

CG 2 nCG (U2)+ nCG (U4)=400

GJ 2 nGJ (U2)+ nGJ (U4)=350

AE 1 nAE (U1)=500

CF 1 nCF (U3)=500

Table 3: Occurrence rating of the paths

3.1.1 Data Allocation Scheme in a Fixed Pattern

In the scheme which allocates data in a fixed

pattern (referred to as DF), the replication sites are

determined when the database is created. Explicitly,

the number of replicated sites and the sites at which

the shared data can be replicated are predetermined.

Though being adopted in some traditional

distributed database systems due to its ease of

implementation, DF is not suitable for mobile

computing environments where mobile users move

frequently. DF suffers from poor performance since

it does not take user moving patterns into

consideration.

342

3.1.2 Data Allocation Scheme in Moving Pattern

As described before, shared data refers to those data

that are used by many mobile users. By properly

determining the set of replicated servers used by a

group of mobile users, data allocation for shared

data is able to increase the local data access ratio in

the sense of both local and global optimization.

Local optimization refers to the optimization that

the likelihood of local data access by an individual

mobile user is maximized, meaning that the

probability of average local access is maximized.

Accordingly, we have the following objective

function for local optimization,

OPTlocal(N) = 1 Ni=1 L(Ui) = 1

Ni=1 f*| R FSi |

N N |FSi|

= f Ni=1 | R FSi |

N |FSi|

where N is the number of mobile users and f is the

hit coefficient. In contrast, global optimization

refers to the optimization that the likelihood of local

data access by all mobile users is maximized,

meaning that the number of total local accesses is

maximized. Hence, the objective function for global

optimization can be formulated as follows:

OPTlocal(N)=1 Ni=1 L(Ui) * ni

N

=1 Ni=1 f*| R FSi | * ni

N |FSi|

= f Ni=1 | R FSi | * ni

|FSi|

where N is the number of mobile users, f is the hit

coefficient, and ni is the number of moving paths

for mobile user Ui. With the user moving patterns

obtained, we can develop shared data allocation

algorithms to determine the set of replicated

servers. The moving patterns of mobile users may

contain different large k-moving sequences,

denoted as Lk; where a k moving sequence is called

a large k-moving sequence if there are a sufficient

number of moving paths containing this k-moving

sequence [8]. A large moving sequence can be

determined from all moving paths of each

individual user based on its occurrences in those

moving paths .We first convert these Lk„s into L2‟s

and the allocation of shared data will be made in

accordance with the occurrences of these L2‟s.

Thus, in algorithm SD-local, we use the user

occurrence count of L2, where the user occurrence

count of L2 is the number of mobile users whose

moving patterns contain that L2.

An example profile for the counting in algorithm

SD-local is given in Table 3. For example, since

AB can only be found in the moving patterns of

U1 the user occurrence count of AB is one. Also,

sinceU1,U2, and U3 contain BC in their moving

patterns, the user occurrence count of BC is 3.

Hence, as mentioned above, those L2 pairs with

larger values of user occurrences should be

included in the set of R so as to maximize the

objective function of local optimization.

3.2.1 SD-LOCAL ALGORITHM:

Algorithm SD-Local

Input: All user moving patterns of mobile users

Output: Set of replicated servers ,R

begin

1. Determine user occurrence counts for all

frequent L2„s from all user‟s moving

patterns

2. Repeat until |P|0;/* |P| is the number of

replicated servers yet to determine*/

343

3. begin

4. Include those L2„s that have maximal

occurrence count into the set C-max /*C

denotes the L2 pair in C-max*/

5. if |R|=0 /*R is set of replicated servers */

begin

Choose an L2 pair from c-max;

Include this pair into R;

|V| = |V| - 2;

end

6. else if (c C-max and R c 0)

begin

In c-max, choose an L2 pair that has an

intersection with pairs in R;

V| = |V| -1;

end

7. else/* there are no intersection pairs in R */

begin

Choose an L2 pair from C-max;

|V| = |V| - 2;

end

8. R=Rc;

9. end

end

Fig. 3, where the network topology is four by three

mesh network. Once the user occurrence counts of

all L2 pairs are obtained, in Fig.3, the number next

to each edge represents the user occurrence count of

the corresponding L2. Then, we include the L2

which has maximal user occurrence count (i.e.,

BC according to the profile in Table 3) resulting

in the configuration with BC added. In general, if

the number of replicated server, |R|, is not equal to

the number of replicated servers required, we

select, from existing L2 pairs that have maximal

user occurrence count (i.e., c-max), the one that has

an intersection with pairs in R (from line 12 to line

15 of algorithm SD-local). The pair CG is hence

selected. After the inclusion of CG, R becomes

BCG.

Following this procedure, we shall identify and

include more proper L2 pairs until |R| reaches the

number of replicated servers required (i.e., |V| =

0).We add GJ and R is composed of the most

frequent moving sites for all mobile users in the

sense of local optimization.

fig.3 selection of replicated server pairs by SD-

local algorithm

On the other hand, based on the objective function

of global optimization, we develop algorithm SD-

global. Since the objective function of global

optimization takes the number of moving paths into

account, the movement occurrence count should be

used for counting, where the movement occurrence

count is the sum of all the movement occurrence

counts of that L2 from all mobile users. An

illustrative example profile is given in Table 3. Let

nBC(Ui) denote the occurrence count of BC in

moving paths of mobile user Ui. The movement

occurrence count of BC is thus the sum of nBC

(U1), nBC (U2), and nBC (U3).

344

3.2.2 SD-GLOBAL ALGORITHM:

Algorithm SD-global

Input: All users moving patterns of mobile users

Output: Set of replicated servers, R

begin

1. Determine , from the counting statistics of

mining user moving patterns [20],movement

occurrence counts of all frequent L2„s

/* line 2 to line 9 are same */

An illustrative example profile is given in Table 3.

Let nBC(Ui) denote the occurrence count of BC in

moving paths of mobile user Ui. L2 with larger

values of movement occurrences will be selected so

as to maximize the value of the objective function

of global optimization. With the same profile in

Table 3, Fig. 4 shows the execution scenario of SD-

global, where the number next to each edge

represents the movement occurrence count of the

corresponding L2. The set of replicated servers by

SD-global can be obtained similarly.

fig.4 Selection of replicated server pairs by SD-

global algorithm

Using both SD-local and SD-global

algorithms improves the local hit ratios of both

frequently and infrequently moving users. Thus the

reason of using replicated servers is satisfied.

4. ANALYSIS

By analysis it is clearly shown that by the usage of

SD-global for frequently moving users and SD-

local for infrequently moving users, the selection of

the replicated server positions become more

accurate and thus increasing the usage of replicated

servers and serving its purpose.

The graph in the figure below with number of users

in the X-axis and the usage of the replicated servers

in Y-axis is clearly given, which shows the increase

of efficiency at least by 52%.

0

20

40

60

80

100

120

250 500 750 1000 1250 1500

SD-local and SD-global allocation

MCTS allocation

Fig.5 comparison between MCTS and SD-local

and SD-global algorithms

345

5. CONCLUSION

In this paper, we devised data allocation schemes

that utilize the knowledge of user moving patterns

for proper allocation of shared data in a mobile

computing system. Specifically, by exploring the

features of local optimization and global

optimization, we derived the objective functions of

local optimization and global optimization. With

the objective functions, we devised algorithm SD-

local and algorithm SD-global to achieve local

optimization and global optimization, respectively.

A measurement, called closeness measure which

corresponds to the amount of the intersection

between the set of frequently moving user patterns

and that of infrequently moving user patterns, was

derived to assess the quality of solutions resulted by

SD-local and SD-global. It was shown by our

analysis that the knowledge obtained from the user

moving patterns is very important in devising

effective shared data allocation algorithms which

can lead to prominent performance improvement in

a mobile computing system.

6. REFERENCES:

[1] J. Jing, A. Helal, and A. Elmagarmid, “Client-

Server Computing in Mobile Environments,” ACM

Computing Surveys, vol. 31, no. 2, pp. 117-157,

June 1999.

[2]N. Krishnakumar and R. Jain, “Escrow

Techniques for Mobile Sales and Inventory

Applications,” ACM J. Wireless Network, vol.

3,no. 3, pp. 235-246, July 1997.

[3]J. Jannink, D. Lam, N. Shivakumar, J. Widom,

and D. Cox,“Efficient and Flexible Location

Management Techniques for

Wireless Communication Systems,” ACM J.

Wireless Networks,vol. 3, no. 5, pp. 361-374, 1997.

[4]O. Wolfson, S. Jajodia, and Y. Huang, “An

Adaptive DataReplication Algorithm,” ACM Trans.

Database Systems, vol. 22,no. 4, pp. 255-314, June

1997.

[5]H.-K. Wu, M.-H. Jin, J.-T. Horng, and C.-Y. Ke,

“Personal Paging Area Design Based on Mobile‟s

Moving Behaviors,” Proc. IEEE Infocom 2001, pp.

21-30, Apr. 2001.

[6]N. Shivakumar, J. Jannink, and J. Widom, “Per-

User Profile Replication in Mobile Environments:

Algorithms, Analysis and Simulation Results,”

ACM J. Mobile Networks and Applications, vol. 2,

pp. 129-140, 1997.

[7]E. Pitoura and G. Samaras, “Locating Objects in

Mobile Computing,”IEEE Trans. Knowledge and

Data Eng., vol. 13, no. 4, pp. 571- 592, July/Aug.

2001.

[8]W.-C. Peng and M.-S. Chen, “Mining User

Moving Patterns for Personal Data Allocation in a

Mobile Computing System,” Proc. 29th Int‟l Conf.

Parallel Processing (ICPP 2000), Aug. 2000.

[9]M.H. Dunham, A. Helal, and S. Balakrishnan,

“A MobileTransaction Model That Captures Both

the Data and MovementBehavior,” ACM J. Mobile

Networks and Applications, vol. 2, pp. 149-

162,1997.

[10] EIA/TIA, Cellular Radio Telecomm.

Intersystem Operations,1991.

[11] T. Imielinski and B.R. Badrinath, “Mobile

Wireless Computing,”Comm. ACM, vol. 37, no.

10, pp. 18-28, Oct. 1994.

[12] Intelligent Transportation Systems,

http://www.artimis.org/,2004.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

346

Greedy Resource Allocation Algorithm for OFDMA

Based Wireless Network Dhananjay Kumar P. Subramanian

Sr. Lecturer M. Tech. Student

Department of Information Technology, Anna University, Madras Institute of Technology Campus,

Chennai-600 044

[email protected], [email protected]

Abstract — A dynamic channel allocation algorithm for an

Orthogonal Frequency Division Multiple Access (OFDMA)

network is presented in this paper. The main objective is to

perform variable resource allocation based on channel gain and

channel state information. Base stations will know the channel

gain based on the feedback given by the mobile station. An

algorithm based on the greedy approach is used to allocate the

resources such as power, bit rate and sub channels. The sub-

carrier allocation algorithm (SAA) consists of two stages. The first

stage determines the number of subcarriers that a user needs and

the second stage deals with the assignment of the subcarriers to

satisfy their rate requirements. The SAA helps to utilize the

available spectrum efficiently. The snap shot of average

transmitted power and observed average Signal to Noise Ratio

(SNR) is presented to predict the system behaviour. Bit Error

Rate (BER) and the transmitting power is measured. The

simulation result shows that with an SNR of 12dB, the average

un-coded BER falls below 10-3

.

Keywords— Adaptive modulation, Resource allocation,

Multiuser channel, OFDM, SNR values

I. INTRODUCTION

The demands for high speed wireless networks are

increasing day by day. The higher throughput per bandwidth

to accommodate more users with higher data rates while

retaining a guaranteed quality of service becomes the main goal of system design. A dynamic multi user resource

allocation strategy is needed achieve this. This requirement is

explored here in the context of an OFDM based multiple

access technique called OFDMA. An OFDMA system is

defined as one in which each user is assigned a subset of the

subcarriers for use, and each carrier is assigned exclusively to

one user [1]. In OFDMA systems the power control and

bandwidth allocation for mobile station is still largely

unexplored that maximizes the system performance. The

OFDMA system performs well because instead of

transmitting over a single channel, users are allowed to choose best sub-channels as per the prevailing conditions [5]. An

adaptation in the frequency domain requires a large amount

of feedback, which makes the channel variation tracking

difficult in mobile applications. To reduce the feedback

overhead, it is often preferred to allow frequency diversity in

each subchannel (SCH) and then SCH wise link adaptation

is performed. In Code Division Multiplexing (CDM), data

symbols of same user are multiplexed and hence the

interference is reduced [2]. Higher data rates can be achieved

by higher levels of modulation techniques.But this may lead to

inter symbol interference(ISI) among the subchannels.This ISI can be mitigated easily by using multi user OFDM.In general

the subcarriers with high channel gain usually employ higher

levels of modulation to carry more bits per OFDM

symbol[3].While those sub carriers that are having low

channel gain are in deep fades and they may or may not carry

bits.The transmission power is distributed across all sub

channels to maximize the throughput..

In typical mobile communication scenario, different sub-carriers experience different channel gain, various data rates

and fade with respect to time and distance. At any instance all

the users may not see the channels that are in deep fade. The

static allocation for the OFDMA system is the simplest one in

which once the sub carriers are assigned to users, they remain

the same through out that session irrespective of their fading

characteristics. This motivates us to consider an adaptive

multiuser subcarrier allocation scheme where the subcarriers are assigned to the users based on instantaneous channel

information. This approach will allow all the subcarriers to be

used more effectively because a subcarrier will be left unused

only if it appears to be in deep fade to all users.

The objective of the proposed algorithm is to maximize the

sum data rate of the system by selecting the channel that

requires minimum transmitting power.Bit loading is done

based on the instantaneous fading characteristics of each sub

channel. An iterative algorithm based on the greedy approach

is used to allocate the subcarriers to multi users in the system. A base station will have some overhead while transmitting the

information about the allocated sub carriers and the number of

bits assigned to each carrier to the users via a dedicated

control channel in the downlink. But this overhead can be

slightly reduced if channel fading is slow and this information

is sent once every OFDM symbols.

The organisation of paper is as follows. In Section II, the

description of the system model and the problem formulation

of the minimum overall transmit power is presented. In

Section III, the bit and power allocation algorithm based on

the greedy approach is studied. The simulation result has been provided in Section IV, followed by conclusions in Section V.

II. SYSTEM MODEL

347

Assume that we have k users in the system and the data rate for the kth user is Rk bits per OFDM symbol. At the transmitter

Fig.1 Block diagram of OFDMA system

side we have the serial data from k number of users. These

serial data are fed to the sub carrier and bit allocation blocks

(Fig.1). It is assumed that the channel bandwidth is much

smaller than the coherence bandwidth of the channel. The

instantaneous channel gain of all sub channels are known at

the transmitter side. Using this information, the block assigns

the bits from different users to the different carriers with the

help of sub carrier bit and power allocation. Based on the

number of bits assigned to each sub carrier, adaptive modulation schemes such as QPSK, 4-QAM, 16-QAM, 64-

QAM and transmitting power is adjusted accordingly. This

varies from subcarrier to subcarrier. Let ck, n be the number of

bits assigned to user k on sub-carrier n. It is assumed that each

user is assigned aspecific sub-carrier which cannot be shared.

Consider a downlink of the cellular frame. A frame consists of

number of OFDM symbols each with number of sub-carriers.

The user information is modulated and assigned to the sub-

carriers. These sub carriers are then passed through the IFFT

block i.e. Inverse Fast Fourier Transform. The function of this

block is to combine all those individual sub-carriers in to an OFDM symbol in time domain.Then the cyclic prefix is added

to the OFDM symbol to avoid the multi path effect. Cyclic

extension of the time domain samples are known to be guard

interval. This is mainly done to ensure the orthogonality

among the sub-carriers. It is then transmitted to different

frequency selective fading channels to various users [4].

At the receiver side, the Fast Fourier Transform is

performed to remove the ISI effects. Now time domain

samples become frequency domain samples.With

demodulators knowing the bit allocation information for each

channel, the information intended for corresponding users is retrieved.

Let αk, n denote the magnitude of channel gain of nth sub

carrier for kth user and Rk be the data rate of user k. W(k) is the

number of subcarriers allocated to user k based on his data

rate requirements.The minimum data rate of the all users are

denoted as rmin and rmax is the maximum data rate of the

subcarrier. It is assumed that the power spectral density No of

all sub carriers are assumed to be unity.This value is same for

all sub-carriers. The power required for receiving c

information bits/symbol when the channel gain is unity is

denoted as fk(c).This function fk(c) is convex i.e. montonically

increasing function.The required transmit power allocated to

the nth subcarrier of kth user is denoted as pk,n.

Pk,n = fk(ck,n)/α 2k,n (1)

This convex function depends on user k.This helps in allowing

different users to have differentiated quality of services

requirements by using an adaptive modulation and coding

scheme (AMC). If fk (0) = 0, then it represents that no power

is needed and no data is transmitted on a particular sub-carrier.

The additional power required to transmit an additional bit

increases with c. This is formulated to find the best

assignment of c information bits over the available subcarriers so that the total transmission power decreases.

Mathematically, it can be formulated as

N K

Pt = min ∑ ∑ fk (ck, n)/α2

k, n (2)

Ck, n ЄM n=1 k=1

The minimization is subjected to the constraints

N

C1: For all users k Є 1, 2, K, Rk = ∑ ck,n (3) n=1

C2: M is the set of all possible values for ck, n

M= 0, 1, 2, D and D is the maximum number of

information bits per OFDM symbol.

Adaptive demodulator N

User 1, rate 1

User 2, rate 2

User K, rate K

User K, rate k

Subcarrier and bit

allocation

Extract bits

for user k

Subcarrier bit and power

allocation algorithm

Adaptive modulator 1

Adaptive modulator N

Adaptive demodulator 1 IFFT

FFT and

Add guard

interval

Frequency

selective

channel

Remove

guard

interval

Subcarrier and bit allocation

information

Channel conditions from

k users

348

C3: W (k) = rmin/rmax, for all users. (4)

C4: Rk ≥ rmin, for all users.

C5: ∑w (k) ≤ N, for all users

III GREEDY ALGORITHM

Initially the problem of resource allocation is divided in to

two stages.

A) Subcarrier Estimation: In this stage we find out the

number of subcarrier that each user needs based on their data rate requirements.

B) Bit Allocation: Based on the channel state

information of all subcarriers, the subcarriers are

allocated to the users.

A) Subcarrier Estimation

The number of subcarrier that each user needs is

determined by using (4).

W (k) = rmin / rmax

The SNR of all the sub-carriers at any instance will not be

same. Hence in a wireless environment some users will have very less SNR values compared to other users. As per LR

algorithm once users have sufficient subcarriers to satisfy their

minimum rate requirements, giving more subcarriers to users

with lower average SNR helps to reduce the total transmission

power.In this way the fairness among the users can be

maintained.

B) Bit Allocation Algorithm

The SAA as shown in Fig.2 assigns bits to the subcarriers

one bit at a time and in each assignment the subcarrier that

requires least transmission is selected. The selected greedy approach is optimal because the power needed to transmit

certain number of bits in a subcarrier is independent of the

number of bits allocated to other subcarriers.

Initialization:

Cn = 0 for all n and ∆pn= [f (1)-f (0)]/α2n;

For each user k:

Repeat the following until all the bits are assigned.

Step1: Select the subchannel that requires the least

transmission power.

n = arg Minn ∆pn;

Cn, k = Cn, k +1;

Step2: Increase the data rate of the selected subchannel and power proportionally based on the values in table 1 given

below.

Step3: After assignment of the bits, calculate the power

needed for assigniing each additional bit.

∆pn= [f (Cn, k +1)-f (Cn, k)]/ α2n

Step4: Repeat the same procedure for selecting the other

subcarriers until data rate requirements of the user is satisfied.

The Table I provide information about data rates of the

various modulation schemes being used. If the channel gain

of any particular subcarrier is very less then a lower order

modulation will be employed. For those that are having good

channel gain, data rates are increased.

TABLE I

ADAPTIVE MODULATION SCHEMES

S. No Data rate in bps

Modulation Schemes

1 1 BPSK

2 2 QPSK

3 4 4-QAM

4 6 16-QAM

Fig.2 Flow chart of the subcarrier allocation algorithm

IV.SIMULATIONS

The main system parameter under consideration is listed in

Table II. The channel model is based on the Rayleigh fading

model with Rician absolute value. The system is simulated

using Matlab version 7.8.0. The single-sided power spectral

density level is taken to unity. The average required transmit

349

power (Fig.3) in defined here is the ratio of the overall

transmit energy per OFDM symbol including all subcarriers

and all users to the total number of bits transmitted per OFDM

symbol. The average SNR is computed as the ratio of average

transmit power to the noise power spectral density level No.

The overall transmit power is proportional to the average SNR

(Fig.4). TABLE III

SYSTEM PARAMETERS

No. of users

1 - 5

Bandwidth 4 MHz

No. of sub-carriers

128

Modulation BPSK, QPSK, 16,64-QAM

Convolution codes Rate

½, 2/3, 3/4

1 2 3 4 50

5

10

15

20

25

Number of users

Tra

nsm

itte

d P

ow

er

in d

b

Fig.3 Transmitted power in db versus Number of users

1 2 3 4 50

5

10

15

20

25

Number of users

SN

R in d

b

Fig.4 SNR in db versus Number of users

It is also observed during simulation that the transmitting

power requirement for each user depends on their data rate

requirements. In other words the SNR requirement is dictated

by the band width requirement of the mobile user. The BER

depends on the SNR and the order of modulation (M) used in

M-QAM. The average BER plot in Fig.5 corresponds to an

un-coded channel.

10 12 14 16 18 20 22 2410

-6

10-5

10-4

10-3

10-2

10-1

100

Average signal to noise ratio

Avera

ge B

it E

rror

Rate

Fig.5 Average BER versus Average SNR

V CONCLUSIONS

The dynamic resource allocation in OFDMA allows

efficient use of system resources in terms of power and

bandwidth. The feasibility of the algorithm is influenced by

the factors such as the channel variation, and the accuracy and

overhead of the channel estimation.The transmission power and bit error rate observed here represents a practical OFDMA

systems.

ACKNOWLEDGMENT

We wish to express our sincere thanks to the Department of

Information Technology, Anna University, MIT Campus,

Chennai for providing the required hardware and software

tools to carry out simulation.

REFERENCES

[1] Kwang soon kimVariable, Yuu hee Kim, ―Power and Rate Allocation

using simple CQI for Multiuser OFDMA-CDM systems,‖IEEE

Trans.Wireless Comm, Vol.8, no.6, june2009.

[2] E. S. Lo, P. W. C. Chan, V. K. N. Lau, R. S. Cheng, K. B. Letaief, R. D.

Murch, and W. H. Mow, “Adaptive resource allocation and capacity com-

parison of downlink multiuser MIMO-MC-CDMA and MIMO-OFDMA,"

IEEE Trans. Wireless Common., vol. 6, no. 3, pp. 1083-1093, Mar. 2007.

[3] K. S. Kim, Y. H. Kim, and J. Y. Ann, “An efficient adaptive transmission technique for LDPC coded OFDM cellular systems using multiple antennas," IEE Electron. Let. vol. 40, no. 6, pp. 396-397, Mar. 2004.

[4] D. Ivan, G. Li, and H. Liu, “Computationally efficient bandwidth

allocation and power control for OFDMA," IEEE Trans. Wireless Common., vol. 2, no. 6, pp. 1150-1158, Oct. 2003.

[5] D. Test and P. Viswanath, Fundamentals of Wireless Communication.

New York: Cambridge University Press, 2005.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

351

Optimal power path routing in wireless sensor

networks Sathiyapriyen.V.B

#1, Shrilekha.M

# 2

#Electronics and Communication Engineering Department, Easwari Engineering College

Ramapuram, Chennai- 600089, India

[email protected]

[email protected]

Abstract—this paper aims to design an optimal power path routing

protocol for IEEE802.15.4 based LRWPAN. IEEE802.15.4 LR-

WPAN has a challenge that power consumption has to be reduced to

maximize the lifetime. The current strategy being used is variable

range transmission power control. Here the coordinator is made to

radiate power depending on the distance and other several

parameters between itself and the end devices and the concepts have

been applied to only data packets. This idea is further stretched to

include the low power route paths also. We will initially simulate the

scenario by incorporating the changes specified, using Network

Simulator 2 (NS2) and then implement it using real time wireless

sensors and compare the results of the original with the modified

analysis

Keywords— wireless sensor networks, AODV, Ns2

1. INTRODUCTION

Wireless Sensor Network (WSN) is an interconnection

of spatially distributed autonomous devices that are used to

monitor a particular physical or environmental condition. It is a

well known fact that in WSN the major constraint is energy

conservation and battery-power consumption, since the sensor is

required to operate under remote conditions without fresh supply

of power to replenish themselves. This paper aims to design a routing protocol at network layer with optimal power as a

primary routing metric.

WSN falls under the category of Low Range Wireless

Personal Area Network (LR-WPAN). The nodes consist of

sensors, a radio transceiver, battery and an embedded processor.

The main applications of sensor networks include industrial

control, asset tracking and supply chain management,

environmental sensing, health monitoring and traffic control.

WSN is different from traditional ad hoc networks in certain

aspects such as being highly energy efficient, having low traffic

rate, low mobility and predefined traffic patterns [1]. Energy conservation is a critical factor for the design of these networks.

Energy conservation involves optimizing protocols at various

layers of the protocol stack. Optimal placement of nodes in

sensor field also contributes to energy conservation but typically

the philosophy in WSN is to have a large number of cheap nodes

which could fail and thus resilience in the distributed protocols is

preferred over strategies such as an optimal placement of nodes.

The major constraint in WSN is energy conservation since the

sensor is required to operate under remote conditions without

fresh supply of power to replenish itself.

2. WSN ARCHITECHTURE

A WSN device can be configured as one of the three types of

nodes namely, a coordinator (PANCOORD), a router (RTR) or

an end device (ED). The end device will sense the environmental

conditions and will communicate the data to the co-ordinator.

The router in addition to its normal routing operations can also

act as an end device. The co-ordinator will collect the data from

end devices and router and process them accordingly. A LR-

WPAN normally consists of one coordinator, several routers and

many end devices. All the nodes operate at a given transmit power irrespective of their location.[2] Some of the

characteristics are Sensors are of :

• Low cost

• Low processing capability System strength based on sensor collaboration

• Large scale networks Multihop communication

• Sensors are battery operated for long unattended period: Saving energy is a primary objective

Some of the applications of WSN are

Remote area monitoring

Object location

Industry machinery monitoring

Disaster prevention

Wireless medical systems

2.1 ENERGY CONSERVATION IN WSN

According to IEEE standard 802.15.4 MAC protocol

[3], energy conservation is accomplished through idle modes and

sleep modes in a sensor node. The node switches between active

and sleep states based on the operations. During sleep states, all

the blocks of the sensor mode are switched off, saving

considerable power. The node wakes up at the time when it wants to transmit or when it expects information from the coordinator.

352

The extent of sleep periods and active periods depends on the

configuration of the network, namely, beacon-enabled mode and

non-beacon-enabled mode.

Although this saves power considerably, there is a

possibility of control over transmit power which can further bring

down the energy spent on transmissions.

3. AODV ROUTING PROTOCOL

AODV [4] is an on-demand routing protocol, only when

the source has data to send, a short route request (RREQ)

message will be initiated and broadcast by the source, with an estimated and pre-defined lifetime (TTL). This RREQ is

rebroadcast until the TTL reaches zero or a valid route is

detected. The nodes receiving the RREQ will add a valid route

entry to its routing table to reach the RREQ source, called reverse

route formation. When the RREQ reaches the destination or a

node that has a valid route to the destination, a route reply

(RREP) message is uni-cast by this node to the source. If one

round of route discovery fails (the RREQ TTL decreases to zero),

the source will re-initiate a new RREQ with a larger initial TTL

after time-out. If several rounds of route request all fail, it means

no valid route can be found. The RREP message will go to the source, following the reverse route formed by the RREQ. For

every hop the RREP is forwarded, intermediate nodes will add a

valid route pointing to the destination, called forward route

formation, until the RREP reaches the source. If the RREP is

generated by an intermediate node that already has a valid route

to the destination, a special message called gratuitous route reply

(G-RREP) is uni-cast to the destination, notifying it that the

source has route request and then a bi-directional route is formed.

By then, both the data source and destination have routes to each

other, and all the intermediate nodes have routes to the source

and destination. In the original AODV, source node will choose

the shortest path if there are multi-routes discovered (with several route replies).

3.1 ROUTE MAINTANANCE IN AODV

Hello messages in AODV are used in the route

maintenance part of the AODV protocol. Hello messages are enabled so that every node knows entire details about its

neighboring nodes. When a node receives a hello message from

its neighbor, it creates or refreshes the neighbor table entry to the

neighbor. Hello message is not enabled by default. You need to

comment the following 2 lines in aodv.h file

#defineAODV_LINK_LAYER_DETECTION

#define AODV_USE_LL_METRIC

hello hello hello

hello hello hello hello hello

Fig 1 Hello messages

3.2 POWER CALCULATION AND UPDATING NEIGHBOUR TABLE

Power calculation is done using Distance and LQI values.

Distance and LQI are split up into five regions: Low, Medium

and High. Transmitting power values are decided using the nine

combinations. LQI is inversely proportional to power. Distance is

directly proportional to power

Only neighbor address and expiry time are available in

the neighbor table initially. Program is altered to include power

values in the neighbor table. After this alteration, when a node

refreshes the neighbor table while receiving a hello message,

power values also get refreshed along with it.

4. ANALYSIS OF WSN UNDER CONTROLLED CONDITIONS

The required transmit power levels are obtained by

observing the parameters under consideration for an ideal

network scenario. The simulation environment is created with the

parameters as given in Table 1.

Table 1: Parameters used to create Simulation Environment

Sl. No. Parameters Values

Fixed Parameters

1 Frequency of

operation

914 MHz

2 Receiver

Threshold

-82 dBm

3 Carrier Sense

Threshold

-92 dBm

4 MAC Protocol IEEE 802.15.4

5 Network

Routing Protocol

AODV

6 Traffic CBR 512

bytes/pkt

7 Simulation

Time

200 seconds

Variable Parameters

1 No. of nodes 2 / 4 / 6 / 25

(Four scenarios)

2 Mobility Induced Mobility in

selected nodes

3 Transmit Power Varied from 0.2

W to 0.2818W

4 Propagation

Model

Two Ray

Ground

Reflection /

Free space

In order to minimise transmit power, we propose a

protocol to be implemented in devices that follow the IEEE

802.15.4 standard such that the transmit power varies with

respect to LQI, distance in between the communicating nodes and

353

collisions in the channel in between the pair of communicating

nodes. The distance between the mobile nodes has a direct

relation to LQI [5].

The factors LQI and distance are divided into three

logic levels and the transmit power is varied in five levels based

on those values. Based on the fuzzy logic table several decisions can be taken such that the transmit power of nodes can be varied

automatically based on the table.

DIST: HI > 175

MED [ 75 – 175]

LOW < 75

LQI: HI >200

MED [100 – 200]

LOW <100

Pt: HI 0.2818W

HIMED 0.26W

MED 0.24W LOMED 0.22W

LOW 0.20W

5. POWER CONTROL ALGORITHM

The algorithmic steps are as follows:

Step 1: Find DIST from the physical layer. Store it in 'DIST'

variable.

Step 2: Packetize and send it to MAC layer

Step 3: In the MAC layer, find LQI, Source MAC address and the

status of the frame (normal, corrupted or collided). Store them in

the variables LQI, MACSRC and ERR respectively.

Step 4: Calculate the average values of DIST and LQI over a time

period, i.e., 5 seconds and store it in variables ADIST and ALQI

respectively. Also store the total no. of error frames in TOTERR.

Step 5: Check if the packet type is DATA or ACK. Step 6: Based

on the accumulated values of TOTDIST and TOTLQI, calculate Average DIST and Average LQI.

Decide the transmit power using the following conditions:

Case 1: if ((DIST==HI) && (LQI==HI or MED)) Pt =LOW

Case2: if ((DIST==HI) && (LQI== LOW))

Pt =LOWMED

Case3: if ((DIST==MED) && (LQI==HI or MED))

Pt = MED

Case4: if ((DIST==MED) && (LQI==LOW))

Pt = HIMED

Case5: if ((DIST==LOW) && (LQI==LOW or MED))

Pt =HI Case6: if ((DIST==LOW) && (LQI==HI))

Pt =HIMED

Step 7: In case of Frame Errors >2, increase the Power level one

step higher

Step 8: Repeat the above steps for every 5 seconds

6.PERFORMANCE ANALYSIS OF IEEE 802.15.4 WITH OPTIMAL POWER

CONTROL

The IEEE 802.15.4 is analysed based on their Distance

LQI values. A look up table is formed is based on that. Then the

power control algorithm is implemented. To visualize the effects

of the power control algorithm, the performance is analysed

based on the energy consumption. The graphs are plotted for

different environment conditions (say number of nodes) to provide an optimal level of solutions.

Fig 2 LQI vs Distance for WSN with 4 nodes using Two Ray Ground Reflection model

Fig. 3 Transmit Power vs Distance for WSN with 2 nodes with

and without power control

Fig. 4 Transmit Power vs Time for WSN with 4 nodes with and

without power control

354

Fig. 5 NAM screenshot of WSN with 25 nodes

The analysis done are collected as a group and then formatted as

a tabular column containing power conserved by implementing

the algorithm and the energy conservation done. The tabular column is given for 2nodes, 4nodes, 6nodes and 25nodes with

two ray model as propagation model. The tabular column is given

as follows:

Table 2: Energy consumption for different number of nodes

Condition Energy consumed(mJ)

2nodes two ray model 36.89(without power

control)

34.036(with power control)

4nodes two ray model 35.23(without power

control)

34.23(with power

control)

6nodes two ray model 37.23(without power

control)

34.394(with power

control)

25 nodes two ray

model

43.23(without power

control)

42.62(with power

control)

Using the table 2 a graph is plotted for number of nodes vs

energy consumed. The line in red color indicates the plot without

power control and the line in green indicates the plot with power

control.

Thus from the graph it can be clearly viewed that the energy has

been successfully conserved thereby increasing the life time of

the sensors used.

7. CONCLUSION

Energy efficient protocol design is the need of the hour

because of increasing dependence on independent systems that

could operate without utilizing much power. Through this paper

we plan to upgrade the existing technique with optimal usage of

power. This technique is being worked out currently in Network

Simulator 2 (NS 2).

8.REFERENCES

[1] Holger Karl and Andreas Willig, ―Protocols and

Architectures for Wireless Sensor Networks‖, First Edition,

Wiley and sons, 2007 .

[2] Energy Efficient Medium Access Control for Wireless Sensor

Networks, Sabitha Ramakrishnan,

T.Thyagarajan, IJCSNS International Journal of Computer

Science and Network Security, VOL.9 No.6, June 2009.

[3] IEEE 802.15.4a: Wireless Medium Access Control (MAC)

and Physical Layer (PHY) Specifications for Low-Rate Wireless

Personal Area Networks (WPANs), Internet draft.

[4] Energy-Aware Algorithms for AODV in Ad Hoc Networks, Xiangpeng Jing and Myung J. Lee

[5] Variable-Range Transmission power control in Wireless Ad

Hoc Networks, IEEE transactions on mobile computing, vol 6,

no.1, January 2007

[6]Energy efficient traffic scheduling in IEEE 802.15.4 for home

automation networks, IEEE Transactions on consumer

electronics, Volume 53, No.2, May 2007.

[7] Variable-Range Transmission Power Control in Wireless Ad

Hoc Networks, IEEE transactions on mobile computing, vol. 6,

no. 1, January 2007

[8] Power Saving Algorithms for Wireless Sensor Networks on IEEE 802.15.4, IEEE Communications Magazine, June 2008.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

355

Analysis of Propagation Models used in

Wireless Communication System Design K.PHANI SRINIVAS Asst Professor KL University

Email: [email protected] Cell: 9849462633

ABSTRACT---The communication path between the transmitter

and the receiver in wireless communication can vary from simple

line of sight (LOS) to one that is rigorously obstructed by buildings,

mountains and shrubbery. Unlike wired channel that are

motionless and predictable, radio channels are extremely random

and do not offer simple analysis. Even the speed of motion impacts

how rapidly the signal level fades as a mobile terminal moving in

space. The radio propagation model has previously been one of the

most complex parts of mobile system design, and is usually done in

a statistical fashion, based on measurements made specifically for

an intended communication system or spectrum allocation. Various

propagation models exist in the open literature. These models are

basically divided into two distinct classes namely physical and

analytical models. Physical models focus on the location of

transmitter and receiver without taking the antenna design into

account. Analytical models capture physical wave propagation and

antenna configuration simultaneously by describing the transfer

function between the antenna arrays at both link ends. This paper

deals with various propagation models and studies their application

to wireless communications system design. The advantages and

disadvantages of each model are given where applicable. Keywords— Quality of service (QoS),Radio frequency(RF), .

Multiple Input Multiple Output (MIMO), Angle of Arrival (AOA),

Angle of Departure (AOD),Line of sight(LOS),Multiple path

components(MPCs)

I. INTRODUCTION Wireless communications is, by any measure, the fastest

growing segment of the communications Industry. As such, it

has captured the attention of the media and the imagination of

the public. Cellular phones have experienced exponential

growth over the last decade, and this growth continues unabated

worldwide, with more than a billion worldwide cell phone users

projected in the near future. Indeed, cellular phones have

become a critical business tool and part of everyday life in most

developed countries, and are rapidly supplanting antiquated wire

line systems in many developing countries. This development is

being driven primarily by the transformation of what has been

largely a medium for supporting voice telephony into a medium for supporting other services, such as the transmission of video,

images, text, and data. Many new applications, including

wireless sensor networks, automated highways and factories,

smart homes and appliances, and remote telemedicine, are

emerging from research ideas to concrete systems. The

explosive growth of wireless systems coupled with the

proliferation of laptop and palmtop computers indicate a bright future for wireless networks, both as stand-alone systems and as

part of the larger networking infrastructure. Thus, similar to the

developments in wire line capacity in the 1990s, the demand for

new wireless capacity is growing at a very rapid pace. Although

there are, of course, still a great many technical problems to be

solved in wire line communications, demands for additional

wire line capacity can be fulfilled largely with the addition of

new private infrastructure, such as additional optical fiber,

routers, switches, and so on. On the other hand, the traditional

resources that have been used to add capacity to wireless

systems are radio bandwidth and transmitter power.

Unfortunately, these two resources are among the most severely limited in the deployment of modern wireless networks: radio

bandwidth because of the very tight situation with regard to

useful radio spectrum, and transmitter power because mobile

and other portable services require the use of battery power,

which is limited. These two resources are simply not growing or

improving at rates that can support anticipated demands for

wireless capacity. many technical challenges remain in

designing robust wireless networks that deliver the performance

necessary to support emerging applications. On the other hand,

one resource that is growing at a very rapid rate is that of

processing power. The term "wireless" has become a generic and all-encompassing word used to describe communications in

which electromagnetic waves or RF (rather than some form of

wire) carries a signal over part or the entire communication path

Like in Fig 1. Wireless operations permits services, such as long

range communications, that are impossible or impractical to

implement with the use of wires. The term is commonly used in

the telecommunications industry to refer to telecommunications

systems (e.g. radio transmitters and receivers, remote controls,

computer networks, network terminals, etc.) which use some

form of energy to transfer information without the use of wires.

Information is transferred in this manner over both short and

long distances. Electromagnetic waves propagate through

356

environments where they are reflected, scattered, and diffracted

by walls, terrain, buildings, and other objects. The ultimate

details of this propagation can be obtained by solving Maxwell’s

equations with boundary conditions that express the physical

Figure 1.Basic wireless communication system

characteristics of these obstructing objects. The two most

important requirements for a modern day wireless

communication system are to support high data rates within the

limited available bandwidth and to offer the maximum

reliability. Multiple Input Multiple Output (MIMO) technology,

which exploits the spatial dimension, has shown potential in

providing enormous capacity gains and improvements in the

quality of service (QoS) [1-5]. in any communication system,

the capacity is dependent on the propagation channel conditions, which in turn are dependent on the environment. Propagation

Modeling is an area of active research and several models have

been developed to predict, simulate and design a high

performance communication system. Propagation models can be

classified into two broad categories, namely physical models

and analytical models as shown in Table 1. Site specific physical

models help in network deployment and planning, while site

independent models are mostly used for system design and

testing. The physical models may be further classified into

deterministic and stochastic models. Deterministic models

characterize the physical propagation parameters in a completely deterministic manner (examples are ray tracing and

stored measurement data). With geometry-based stochastic

propagation models (GSPM), the impulse response is

characterized by the laws of wave propagation applied to

specific Tx, Rx, and scattered geometries, which are chosen in a

stochastic (random) manner. A deterministic model tries to

reproduce / repeat the actual physical radio propagation process

for a given environment along with the reflection, diffraction,

shadowing by discrete obstacles, and the wave guiding in street

canyons. Recorded impulse response and ray tracing

techniques are some of the examples of deterministic

channel modeling techniques. The stochastic models are

based on the fact that the wireless propagation channels are

unpredictable and time varying in nature but its parameters, like

the Angle of Arrival (AOA), Angle of Departure (AOD), time

delay profiles etc., follow a defined stochastic/ statistical

behavior, depending on the environment. The stochastic channel

models are generally computationally efficient. Most stochastic

models have a geometrical basis; In the realm of geometrically

based stochastic models, large variants of the model have been

proposed, but the basic philosophy remains the same. Different

geometrically based stochastic models reproduce different sets of environments like indoor or outdoor scenarios, and narrow

band or wide band environments. Usually, the models are

validated by comparing the values or distributions of certain

physical parameters like AOA, AOD, Time of arrival (TOA),

and power spectrum etc., obtained through the model with those

acquired through measurements under specific conditions. In

contrast, non-geometric stochastic models describe and

determine physical parameters (the direction of departure

(DoD), and the direction of arrival (DoA) delay, etc.) in a

completely stochastic way by prescribing underlying probability

distribution functions without assuming an underlying geometry

(examples are the extensions of the Saleh-Valenzuela model, Zwick model [14], [15]).

Table 1.

Physical models: Analtical models 1. Deterministic : -Ray tracing -Stored measurements

2. Geometry –based Stochastic: -GSPM 3 .Non-Geometrical Stochastic: -Saleh-Valenzula type -Zwick model

1.Correlation-based: - canonical model

-Kronecker model -weichselberger Model 2. Propagation -motivated: -Finite scatterer

model -Maximum entropy model -Virtual channel representation

II. PHYSICAL MODELS

A. Deterministic Physical Models

Physical propagation models are termed “deterministic” if they

plan to reproduce the actual physical radio propagation process

for a given atmosphere. The deterministic channel modeling

techniques try to repeat the actual physical scenario between the

transmit and the receive arrays. In urban environments, the

geometric and electromagnetic characteristics of the

environment and of the radio link can be easily stored in files (environment databases) and the corresponding propagation

process can be simulated through computer programs. Most

often, the antenna parameters like the antenna patterns, array

size and geometry, the effects of mutual coupling between the

array elements, polarization etc. are not accounted for [6].

Although electromagnetic models such as the method of

moments (MoM) or the finite-difference in time domain

357

(FDTD) model may be useful to study near field problems in the

vicinity of the Transmitter or Receiver antennas.

Ray tracing softwares and techniques are one of the most

accepted ways for modeling the channel deterministically. In ray

tracing software’s, the geometry and the electromagnetic

characteristics of any particular situation/ environment is stored in files. The ray tracing Software’s are basically based on the

phenomenon of geometrical optics like reflection, refraction,

diffraction etc. These files are later used for simulating the

electromagnetic propagation process between the transmitter

and the receiver. These models are fairly accurate and may be

used in place of measurement campaigns, when time is at

premium. In ray tracing techniques, flat top polygons of

different sizes and shapes are generally used to represent

buildings as shown in fig.2 Geometrical optics is based on the

so-called ray approximation, which assumes that the wavelength

is sufficiently small compared to the dimensions of the obstacles

in the environment. This assumption is usually valid in urban radio propagation and allows to express the electromagnetic

field in terms of a set of rays, each one of them corresponding to

a piece-wise linear path connecting two terminals. Each

“corner” in a path corresponds to an “interaction” with an

obstacle (e.g. wall reflection, edge diffraction). Rays have a null

transverse dimension and therefore can in principle describe the

field with infinite resolution. If beams with a finite transverse

dimension are used instead of rays, then the resulting model is

called beam launching, or ray splitting. Beam launching models

allow faster field strength prediction but are less accurate in

characterizing the radio channel between two SISO or MIMO terminals.

Fig. 2 propagation scenario (Rectangular blocks shows buildings) B. Geometry-Based Stochastic Physical Models

Any geometry-based model is resolute by the scatterer locations.

In deterministic geometrical approaches (like RT discussed in

the previous subsection), the scatterer locations are prescribed in

a database. In contrast, geometry-based stochastic propagation

models (GSPM) choose the scatterer locations in a random style

according to a certain probability distribution. The actual

channel impulse response is then found by a simplified RT

procedure. 1).Single-Bounce Scattering The antecedent of the GSPM in [7] placed scatterers in a

deterministic mode on a circle around the mobile station, and

assumed that only single scattering occurs. approximately

twenty years later, several groups simultaneously suggested to

augment this single-scattering model by using randomly placed

scatterers [10], [11], [12], This random assignment reflects

physical reality much better. The single-scattering assumption

makes RT extremely simple: apart of the LoS, all paths consist

of two sub paths connecting the scatterer to the Tx and Rx,

respectively. [13], [14], [15]. These sub paths characterize the

DoD, DoA, and propagation time (which in turn determines the overall attenuation, usually according to a power law). The

scatterer interaction itself can be taken into account via an

additional random phase shift. Different types of the GSPM differ mainly in the proposed

scatterer distributions. The simplest GSPM is obtained by

assuming that the scatterers are spatially uniformly distributed.

Contributions from far scatterers carry less power since they

propagate over longer distances and are thus attenuated more

strongly; this model is also often called single-bounce

geometrical model. An alternative approach suggests to place

the scatterers randomly around the MS [12], [14]. Various other scatterer distributions around the MS were analyzed; a one-sided

Gaussian distribution w.r.t. distance from the MS resulted in an

approximately exponential PDP, which is in good agreement

with many measurement results[16]. To make the density or

strength of the scatterers depend on distance, two

implementations are possible. In the ”classical” approach, the

probability density function of the scatterers is adjusted such

that scatterers occur less likely at large distances from the MS.

Alternatively, the“non-uniform scattering cross section” method

places scatterers with uniform density in the considered area, but

down-weights their contributions with increasing distance from

the MS [17]. For very high scatterer density, the two approaches are equivalent. However, non-uniform scattering cross section

can have numerical advantages, in particular less statistical

fluctuations of the power-delay profile when the number of

scatterers is finite. Another important propagation effect arises

from the existence of clusters of far scatterers (e.g. large

buildings, mountains . . .). Far scatterers lead to increased

temporal and angular dispersion and can thus significantly

influence the performance of MIMO systems. In a GSPM, they

can be accounted for by placing clusters of far scatterers at

random locations in the cell.

2) Multiple-Bounce Scattering

358

The single bounce scattering based on the theory that only

single-bounce scattering is present. This is limiting insofar as

the position of a scatterer completely determines DoD, DoA,

and delay, i.e., only two of these parameters can be chosen

independently. However, many environments (e.g., micro- and

Pico cells) feature .multiple-bounce scattering for which DoD, DoA, and delay are completely decoupled. In microcells, the BS

is below rooftop height, so that propagation mostly consists of

wave guiding through street canyons, which involves multiple

reflections and diffractions (this effect can be significant even in

macro cells[18,19]For Pico cells, propagation within a single

large room is mainly determined by LoS propagation and single-

bounce reflections. However, if the Tx and Rx are in different

rooms, then the radio waves either propagate through the walls

or they leave the Tx room e.g. through a window or door, are

waveguide through a corridor, and be diffracted into the room

with the Rx [20]. If the directional channel properties need to be

reproduced only for one link end (i.e., multiple antennas only at the Tx or Rx), multiple-bounce scattering can be incorporated

into a GSPM via the concept of equivalent scatterers. These are

virtual single-bounce scatterers whose position is chosen such

that they mimic multiple bounce contributions in terms of their

delay and DoA. This is always possible since the delay,

azimuth, and elevation of a single-bounce scatterer are in one-

to-one correspondence with its Cartesian coordinates.

C. Non-geometrical Stochastic Physical Models Non-geometrical stochastic(random) models tell paths from Tx

to Rx by statistical parameters only, without indication to the geometry of a physical surroundings. There are two classes of

stochastic non-geometrical models reported in the literature. The

first one uses clusters of MPCs and is generally called the

extended Saleh-Valenzuela model since it generalizes the

temporal cluster model developed in [21]. The second one

Zwick model, treats MPCs individually.[22]

1) Saleh-Valenzuela Model:

Saleh and Valenzuela projected to model clusters of MPCs in

the delay domain via a doubly exponential decay process . The

Saleh-Valenzuela model uses one exponentially decaying

profile to control the power of a multipath cluster.[21] The MPCs within the individual clusters are then characterized by a

second exponential profile with a steeper slope.The Saleh-

Valenzuela model has been extended to the spatial domain in

[23]. In particular, the extended Saleh-Valenzuela MIMO

model in is based on the assumptions that the DoD and DoA

statistics are self-sufficient and identical. Usually, the mean

cluster angle is assume to be uniformly distributed within

0,2 and the angle ' of the MPCs in the cluster are

Laplacian distributed, i.e., their probability density function

equals

(1)

where characterizes the cluster’s angular spread. The mean

delay for each cluster is characterized by a Poisson process, and

the individual delays of the MPCs within the cluster are

characterized by a second Poisson process relative to the mean

delay.[24]

2) Zwick Model: It is argued that for indoor channels clustering and multipath

fading do not occur when the sampling rate is sufficiently large

. Thus, in the Zwick model, MPCs are generated independently

(no clustering) and without amplitude fading. However, phase

changes of MPCs are incorporated into the model via geometric

considerations describing Tx, Rx, and scatterer motion[24]. The

geometry of the scenario of course also determines the existence

of a specific MPC, which thus appear and disappear as the

channel impulse response evolves with time. For non-line of

sight (NLoS) MPCs, this effect is modeled using a marked

Poisson process. If a line-of-sight (LoS) component shall be included, it is simply added in a separate step. This allows to use

the same basic procedure for both LoS and NLoS

environments.[22]

III. ANALYTICAL MODELS

A. Correlation-based Analytical Models

Various narrowband analytical models are based on a

multivariate complex Gaussian distribution of the MIMO

channel coefficients. The channel matrix can be split into a zero-mean stochastic part Hs and a purely deterministic part Hd

[9,26]

(2)

Where K 0 denotes the Rice factor. The matrix Hd refers for LoS components and other non-fading contributions. In the

following, we focus on the NLoS components characterized by

the Gaussian matrix Hs.

For simplicity, we thus assume K = 0, i.e., H = Hs. In its most

general form, the zero-mean multivariate complex Gaussian

distribution of h = vecH is given by

(3)

The nm × nm matrix

(4)

359

is known as full correlation matrix and describes the spatial

MIMO channel statistics. It contains the correlations of all

channel matrix elements. Realizations of MIMO channels with

distribution (3) can be obtained by

(5)

Here, denotes an arbitrary matrix square root and g is an

nm × 1 vector with i.i.d. Gaussian elements with zero mean and

unit variance.[30]

1) canonical model

canonical model is also called the i.i.d. model. Here

i.e., all elements of the MIMO channel matrix H are

uncorrelated (and hence statistically independent) and have

equal variance . Physically, this corresponds to a spatially

white MIMO channel which occurs only in rich scattering

environments characterized by independent MPCs uniformly

distributed in all directions. The i.i.d. model consists just of a

single parameter (the channel power ) and is often used for

theoretical considerations like the information theoretic analysis

of MIMO systems [27].

2) The Kronecker Model: The Kronecker model became popular because of its simple

analytic treatment. However, the main drawback of this model

is that it forces both link ends to be separable , irrespective of

whether the channel supports this or not. Kronecker model was

used in for capacity analysis before being proposed by [30] in

the framework of the European Union SATURN project. It

assumes that spatial Tx and Rx correlation are separable, which

is equivalent to restricting to correlation matrices that can be

written as Kronecker product[28,29,31]

(6)

With the Tx and Rx correlation matrices

(7) Respectively. It can be shown that under the above assumption,

(14) simplifies to the Kronecker model

(8)

With G = unvec(g) an i.i.d. unit-variance MIMO channel matrix. The model requires specification of the Tx and Rx correlation

matrices, which amounts to n2 + m2 real parameters. _ _ _

_ _ _ _ _ _ _

_ _ _ _ _ _ _

_ _ _ _ _ _ _

3) The Weichselberger Model:

The idea of Weichselberger was to relax the separability

Restriction of the Kronecker model and t o allow for any

arbitrary coupling between the transmit and receive Eigen base,

i.e. to model the correlation properties at the receiver and

transmitter jointly. Its definition is based on the eigen value

decomposition of the Tx and Rx correlation matrices,

x, (9)

(10)

Here, UTx and URx are unitary matrices whose columns are the

eigenvectors of RTx and RRx, respectively,[31] and and

are diagonal matrices with the corresponding eigen values. The model itself is given by

(11)

where G is again an n × m i.i.d. MIMO matrix, denotes the

Schur-Hadamard product (element-wise multiplication), and is an n×m coupling matrix whose (real-valued and nonnegative)

elements determine the average power coupling between the Tx

and Rx eigen modes. This coupling matrix allows for joint

modeling of the Tx and Rx channel correlations. We note that

the Kronecker model is a special case of the Weichselberger

model obtained with the rank-one coupling matrix

where and are vectors containing the eigen values of the

Tx and Rx correlation matrix, respectively. [22,24]

B. Propagation-motivated Analytical Models

1) Finite Scatterer Model:

The fundamental assumption of the finite scatterer model is that

propagation can be modeled in terms of a finite number P of

multipath components as shown in Fig. 3. For each of the

components (indexed by p), a DoD , DoA , complex

amplitude , and delay is specified.[24]

Fig. 3 finite scatterer model with single-bounce scattering (solid line), multiple-bounce scattering (dashed line), and a “split” component (dotted line)

The model allows for single-bounce and multiple-bounce

scattering, which is in contrast to GSPMs that usually only

incorporate single-bounce and double-bounce scattering. Given

the parameters of all MPCs, the MIMO channel matrix H for the

narrowband case (i.e., neglecting the delays ) is given by

(12) where

and are the Tx and Rx steering vectors Corresponding to the pth MPC, and is a diagonal matrix

360

consisting of the multipath amplitudes. For wideband systems,

also the delays must be taken into account. Including the band

limitation to the system bandwidth B = 1/Ts into the channel, the

resulting tapped delay line representation of the channel reads

(13) with

where and Tl is a diagonal matrix with diagonal elements

. For example, the

measurements in [71] suggest that in an urban environment all

multipath parameters are statistically independent and the

DoAs and DoDs are approximately uniformly distributed,the complex amplitudes »p have a log-normally

distributed magnitude and uniform phase, and the delays are

exponentially distributed.[24,31,32]

2) Maximum Entropy Model:

Analytically wireless propagation models are derived from the

maximum entropy principle,when only limited information about the environment is available. These models are useful in

situations where analytical models of the fading characteristics

of a multiple-antennas wireless channel are needed,and where

the classical Rayleigh fading model is too coarse. In scrupulous, the maximum entropy principle was proposed to

conclude the distribution of the MIMO channel matrix based on

a priori information that is existing. This a priori information

might include properties of the propagation environment and

system parameters (e.g., bandwidth, DoAs, etc.). The maximum

entropy principle was justified by the objective to avoid any

model assumptions not supported by the prior information. As

far as consistency is concerned, shows that the target application for which the model has to be consistent can influence the

proper choice of the model. Hence, one may obtain different

channels models for capacity calculations than for bit error rate

simulations. Since this is obviously undesirable, it was proposed

ignore information about any target application when

constructing practically useful models. Then, the maximum

entropy channel model was shown to equal [24,31]

(14)

where G is an sRx × sTx Gaussian matrix with i.i.d. elements. We note that this model is consistent in the sense that less detailed

models (for which parts of the prior information are not

available) can be obtained by “marginalizing” (14) with respect

to the unknown parameters9. Examples include the i.i.d.

Gaussian model where only the channel energy is known

(obtained with = Fm where Fm is the length-m DFT matrix,

= Fn, PTx = I, and PRx = I), the DoA model where steering

vectors and powers are known only for the Rx side (obtained

with = Fm, PTx = I), and the DoD model where steering vectors and powers are known only for the Tx side. (obtained

with = Fn, PRx = I). We conclude that a useful feature of the

maximum entropy approach is the simplicity of translating an

increase or decrease of (physical) prior information into the

channel distribution model in a consistent fashion.

3) Virtual Channel Representation: In contrast to the two prior models , the virtual channel

Representation (VCR) models the MIMO channel in the

beamspace instead of the eigenspace. In particular, the

eigenvectors are replaced by fixed and predefined steering

vectors. A MIMO model called virtual channel representation

was proposed as follows:

Here, the DFT matrices Fm and Fn contain the steering vectors

for m virtual Tx and n virtual Rx scatterers, G is an n × m i.i.d.

zero-mean Gaussian matrix, and is an n × m matrix whose elements characterize the coupling of each pair of virtual

scatterers, i.e. represents the “inner” propagation

environment between virtual Tx and Rx scatterers. We note that

the virtual channel model can be viewed as a special case of the

Weichselberger model with Tx and Rx eigen modes equal to the

columns of the DFT matrices.[33] In the case where = 1, the virtual channel model reduces to the i.i.d. channel model,

i.e., rich scattering with full connection of (virtual) Tx and Rx

scatterer clusters. Due to its simplicity, the virtual channel

model is mostly useful for theoretical considerations like the

analysing the capacity scaling behavior of MIMO channels. It

was also shown to be capacity complying in However, one has

to keep in mind that the virtual representation in terms of DFT

steering matrices is appropriate only for uniform linear arrays at

Tx and Rx.[34,35]

IV. CONCLUSIONS

The tremendous development in wireless communications Leads to the emergence of new ideas and techniques to increase

Capacity and improve the QoS. Smaller cell sizes higher

frequencies, and more complex environments need to be more

accurately modeled and site- specific propagation prediction

models need to be developed to achieve optimum design of

Next-generation wireless communication systems. This paper

provided a survey of the most important concepts in radio

propagation modeling for wireless systems. We advocated an

intuitive classification into physical models that focus on double

directional propagation and analytical models that concentrate

on the channel impulse response (including antenna properties). For both types, we take the examples that are widely used for

the design and evaluation of MIMO wireless systems.

361

REFERENCES

1. D.-S. Shiu, G. J. Foschini, M. J. Gans, and J. M. Kahn, “Fading

correlation and its effect on the capacity of multielement antenna

systems,” IEEE Transactions on Communication., vol. 48, no. 3,

pp. 502-513, March 2000.

2. I. E. Telatar, “Capacity of multi-antenna Gaussian channels,” European

Trans. Telecommun. Related Technol., vol. 10, pp. 585-595, 1999.

3. Volker Kuhn, Wireless Communications over MIMO Channels,

John Wiley and sons, 2006.

4. Claude Oesteges, and Bruno Clerckx, MIMO Wireless Communications:

Real-World Propagation to Space-TimeCode Design, Academic Press, 2007.

5. George Tsoulos, MIMO System Technology for Wireless

Communications, CRC Press, 2006.

6. A. F. Molisch, Wireless Communications. IEEE Press – Wiley, 2005. 7. W. Lee, “Effects on Correlations between Two Mobile Base-Station Antennas,” IEEE Trans. Comm., vol. 21, pp. 1214–1224, 1973.

8. J. Wallace and M. Jensen, “Statistical Characteristics of Measured MIMO Wireless Channel Data and Comparison to Conventional Models,” in Proc. IEEE Vehicular Technology Conference, VTC 2001

Fall, vol. 2, Sidney, Australia, Oct. 2001, pp. 1078–1082. 9. J. Wallace and M. Jensen “Modeling the Indoor MIMO Wireless Channel,” IEEE Trans. on Antennas and Propagation, vol. 50, no. 5

, pp. 591–599, May2002. 10. P. Petrus, J. Reed, and T. Rappaport, “Geometrical-based Statistical Macrocell Channel Model for Mobile Environments,” IEEE Trans.

Comm., vol. 50, no. 3, pp. 495–502, Mar. 2002. 11. J. C. Liberti and T. Rappaport, “A Geometrically Based Model for Line-Of-Sight Multipath Radio Channels,” in Proc. IEEE Vehicular

Technology Conf., Apr.May 1996, pp. 844–848. 12. J. Blanz and P. Jung, “A Flexibly Configurable Spatial Model for Mobile Radio Channels,” IEEE Trans. Comm., vol. 46, no. 3, pp.

367–371, Mar. 1998. 13. O. Norklit and J. Andersen, “Diffuse Channel Model and Experimental Results for Array Antennas in Mobile Environments,”

IEEETrans. on Antennas and Propagation, vol. 46, no. 6, pp. 834– 843, June 1998. 14. J. Fuhl, A. F. Molisch, and E. Bonek, “Unified Channel Model for

Mobile Radio Systems with Smart Antennas,” IEEE Proc. - Radar, Sonar and Navigation: Special Issue on Antenna Array Processing Techniques, vol. 145, no. 1, pp. 32–41, Feb. 1998.

15. C. Oestges, V. Erceg, and A. Paulraj, “A physical scattering model for MIMO macrocellular broadband wireless channels,” IEEE Journal on Selected Areas in Communications, vol. 21, no. 5, pp.

721–729, June 2003. 16. J. Laurila, A. F. Molisch, and E. Bonek, “Influence of the Scatter Distribution on Power Delay Profiles and Azimuthal Power Spectra

of Mobile Radio Channels,” in Proc. ISSSTA’98, Sept. 1998, pp. 267– 271. 17. [56] A. F. Molisch, A. Kuchar, J. Laurila, K. Hugl, and R.

Schmalenberger, “Geometry-based Directional Model for Mobile Radio Channels – Principles and Implementation,” European Trans. Telecomm., vol. 14, pp. 351–359, 2003.

18. M. Toeltsch, J. Laurila, A. F. Molisch, K. Kalliola, P. Vainikainen, and E. Bonek, “Statistical Characterization of Urban Mobile Radio Channels,” IEEE J. Sel. Areas Comm., vol. 20, no. 3, pp. 539–549,

Apr. 2002. 19.] A. Kuchar, J.-P. Rossi, and E. Bonek, “Directional Macro-Cell Channel Characterization from Urban Measurements,” IEEE Trans.

on Antennas and Propagation, vol. 48, no. 2, pp. 137–146, Feb.2000 20. C. Bergljung and P. Karlsson, “Propagation Characteristics for

Indoor Broadband Radio Access Networks in the 5 GHz Band,” in Proc. PIMRC’99, Sept. 1998, pp. 612–616. 21. A. Saleh and R. Valenzuela, “A Statistical Model for Indoor Multipath

Propagation,” IEEE J. Sel. Areas Comm., vol. 5, no. 2, pp.128–137,

Feb. 1987. 22. T. Zwick, C. Fischer, and W. Wiesbeck, “A Stochastic Multipath

Channel Model Including Path Directions for Indoor Environments,” IEEE J. Sel. Areas Comm., vol. 20, no. 6, pp. 1178–1192, Aug. 2002.

23. C. Chong, C. Tan, D. Laurenson, M. Beach, and A. Nix, “A New Statistical Wideband Spatio-temporal Channel Model for 5-GHz Band WLAN Systems,” IEEE J. Sel. Areas Comm., vol. 21, no. 2, pp.

139–150, Feb. 2003. 24. P. Almers, E. Bonek, A. Burr, N. Czink, M. Debbah, V. Degli-

Esposti, et al, “Survey of Channel and Radio Propagation Models

for Wireless MIMO Systems.” EURASIP Journal on Wireless

Communications and Networking. 2007.

26. P. Soma, D. Baum, V. Erceg, R. Krishnamoorthy, and A. Paulraj,

“Analysis and Modeling of Multiple-Input Multiple-Output (MIMO) Radio Channel Based on Outdoor Measurements Conducted at 2.5 GHz for Fixed BWA Applications,” in Proc. IEEE Intern. Conf. on

Comm., ICC 2002, vol. 1, Apr./May 2002, pp. 272–276. 27. I. Telatar, “Capacity of Multi-Antenna Gaussian Channels,” Technical Memorandum, Bell Laboratories, Lucent Technologies, Oct.

1998, published in European Transactions on Telecommunications, vol. 10, no. 6, pp. 585–595, Nov./Dec. 1999. 28. C.-N. Chuah, J. Kahn, and D. Tse, “Capacity of Multi-Antenna Array

Systems in Indoor Wireless Environment,” in Proc. IEEE Global Telecommunications Conf., vol. 4, Sidney, Australia, 1998, pp. 1894– 1899.

29. D. Chizhik, F. Rashid-Farrokhi, J. Ling, and A. Lozano, “Effect of Antenna Separation on the Capacity of BLAST in Correlated Channels,” IEEE Comm. Letters, vol. 4, no. 11, pp. 337–339, Nov.

2000. 30. J. Kermoal, L. Schumacher, K. Pedersen, P. Mogensen, and F. Frederiksen, “A Stochastic MIMO Radio Channel Model with

Experimental Validation,” IEEE J. Sel. Areas Comm., vol. 20, no. 6, pp. 1211–1226, Aug. 2002. 31. Babu Sena Paul and Ratnajit Bhattacharjee: MIMO Channel

Modeling: A Review; IETE TECHNICAL REVIEW , Vol

25 , ISSUE 6, NOV-DEC 2008

32. Li-Chun Wang, Senior Member, IEEE, Wei-Cheng Liu,

Student Member, IEEE, and Yun-Huai Statistical Analysis

of a Mobile-to-Mobile Rician Fading Channel Model; IEEE

TRANSACTIONS ON VEHICULAR TECHNOLOGY, VOL. 58, NO. 1,

JANUARY 2009.

33. A. Sayeed, “Deconstructing Multiantenna Fading Channels,” IEEE Trans. on Signal Proc., vol. 50, no. 10, pp. 2563–2579, Oct. 2002.

34. M. Debbah, R. M¨uller, H. Hofstetter, and P. Lehne, “Validation of Mutual Information Complying MIMO Models,” submitted to IEEE Transactions on Wireless Communications, can be downloaded at

http://www.eurecom.fr/»debbah, 2004. 35. M. Debbah and R. M¨uller, “Capacity Complying MIMO Channel Models,” in Proc. the 37th Annual Asilomar Conference on Signals,

Systems and Computers, Pacific Grove, California, USA, 2003, November 2003.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

362

GSM MOBILE TECHNOLOGY FOR RECHARGE SYSTEM

R.shanthi1

, G.Revathy2

, J.Anandhavalli3

M.phil(computer science) E-mail : [email protected], [email protected] , [email protected]

Periyar Maniammai University , Vallam, Thanjavur.

Mr.K.Sethurajan, Asst.Prof(MCA Dept)

E-mail : [email protected]

Periyar Maniammai University , Vallam, Thanjavur.

Abstract

Global system for mobile

communication system (GSM) provided

the efficient services and technologies

for mobile networks. Now a days

wireless communication technologies

providing a new innovation technologies

as well as challenges to tele

communications. The use of handheld

Devices such as mobile handsets that

charged by a Global system for mobile

communication network is increasing.

but Gsm services are faced the problem

like challenges of inefficient bandwidth

utilization and congestion in the

transmission control. One of the main

reason for this congestion that on the

network by need to reload/recharge

phone credit in order to enable calls.

quality of services(QoS) Demands that

wireless services be readily available to

end users without limitations occasioned

by congested highways, inefficient

bandwidth utilization and remote

location challenges in signal coverage

areas. Here we proposes a GSM

networks on mobile agent recharge

system.the main goal of this system is to

provide a bandwidth utilization and

remote challenges in signal coverage

areas. This paper explained a mobile

agent based recharging system for GSM

networks. And also improving a

bandwidth utilization by using a method

of query processing and report

feedbacks.

Keywords

GSM system, Mobile agents,

telecommunication networks, wireless

networks and software agents

1. Introduction

Wireless networks are complex

multidisciplinary systems, and a

description of their standards is often

very long and tedious. The Global

system for Mobile (GSM) is an ETSI

(European telecommunications standards

Institute) standard for 2G Pan-European

digital cellular with international

roaming. The original goal of the GSM

could be met only by defining a new air-

interface, the group went beyond just the

air-interface and defined a system that

compiled with emerging ISDN

(Integrated services digital networks)

like services and other emerging fixed

network features.

GSM is the most widely used

mobile network in the world. It

providing of good speech quality, low

terminal and service cost, support for

international roaming , support for

handheld terminals, spectral efficiency

and ISDN compatibility. Now a days

mobile phone networks are increasingly

used for much more than voice calls. As

a result of that mobile handsets can

363

offer access to email, sms , gps, mms ,

wap(wireless application protocol) based

on increasing technological

advancement in their operations.

GSM services is a multipurpose

system to identify the services that are

provided by that network because the

entire network is designed to support

these services. GSM is an integrated

voice data service that provides a

number of services beyond cellular

telephone. These services are divided

into three categories: Teleservices ,

Bearer services , Supplementary

Services.

2.Summary of the paper: In this paper

we defines the objectives and

architecture of the GSM. Then the

mechanisms that are designed to support

mobility to the GSM services. we give a

presentation of GSM networks model

with mobile agent areas of applications

and suitability for recharge systems and

describe the operational framework and

the proposed model and mobile agent

based recharging , retrieval ,and request

forwarding to their destinations.

2.1 Reference Architecture

Description of a wireless network

standard is a complex process that

involves detailed specification of the

terminal , fixed hardware backbone and

software data based that are need to

support the operation. To describe such a

complex a reference model or overall

architecture is divided into two

subsystems.

1.Mobile station(MS)

Base station subsystem(BSS)

2. Network and switching

subsystem(NSS)

2.2 Mechanisms to support Mobile

Environment

Four mechanisms are embedded

in all voice oriented wireless networks

that allow a mobile to establish and

maintain a connection with the network.

These mechanisms are

Registration

Call establishment

Handover(or handoff) and

Security

Registration take place as soon as on

turns the mobile unit on, Call

establishment occurs when the user

initiates or receives a call , handover

helps the MS to change its connection

point to the network and security

protects the user from fraud via

authentication , avoid revealing the

subscriber number over the air and

encrypt conversation where possible. All

these are achieved using proprietary

algorithms in GSM

3.0 Networks Models For GSM

In order to accommodate the

diversity of network components,

management applications incorporate

large numbers of interfaces and tools.

Mobile phone network management

systems are usually huge monoliths that

are difficult to maintain. In this section,

we review a number of application areas

aimed at illustrating the suitability of

mobile agents in the management of

mobile phone networks. The OSI

management model identifies distinct

functional areas based on the

requirement of each category.

These include:

Fault management

Configuration management

Performance management

364

Network monitoring and

accounting management

Security management

There exist several potential applications

in each of these areas. However, our

focus in this paper is on network

monitoring and accounting management.

3.1 Model Presentation The Agent model can be viewed as a

composition of the following

infrastructures:

(a) Agents

(b) Mobility

(c) Communication

(d) Servers and

(e) Security The simulated model consists of three

basic components.

• Mobile client/unit model

• Communication network and

• The server model

Agents

A mobile agent in this regard is a

composed object with ability to visit

agent-enabled environments(server(s)) in

a telecommunication network.

Generally, a mobile agent consists of its

code (completely or partially) and its

persistent state. The code is said to be

complete if all the instruction and

materials needed for its operation is

provided at the point of

composition/generation. However, due

to the task of some agents involving

several activities as it traverse a network,

large volume of codes usually slow

down the operation and migration

potentials of such agent. To this end,

such agents are provided with minimal

amount of code necessary for the

commencement of its operation while

other codes are downloaded in the

course of the operation. This special

compiled version of mobile agent is

necessitated by the fact that cellular

phones now operates by their own

operating system. The rapid growth of

Java-based mobile units/devices and

subsequent portability on mobile phones

provides execution environment with no

additional software except for the mobile

agent’s code that is expected to be

transferred once alongside the agent’s

state

Mobility

A clear distinguishing factor

proposed in this research is the migration

ability incorporated into its agents. Not

all agents are mobile, thus it behooves to

say that mobility is an important feature

in this research in other to allow

utilization of other attributes of an agent.

The ability of an agent to move from one

location to the other embedded with the

list of expected operation and places

(servers) to be visited cum flexible query

issuance and processing is a major

advantage to be employed to achieve

mobility. Java’s provision for the above

operation is known as Remote Method

Invocation (RMI) while OMG provides

mobile system agent interoperability

facility.

Communication

The notion of communication becomes

important as the need arise for an agent

to request (communicate) with other

agents in its environment or even the

database (server). Several prototypes

available for this operation.Standard

University providing support for

knowledge Query and Manipulation

Language (KQKL), language known for

Knowledge Representation (KR) or

Knowledge Interchange Format (KIF).

After their operation in the migrated

environment, it is also

365

expected that the same mode is

employed by the agent to communicate

with their owner and vice-versa by either

a status mail or SMS message to the

originator. Our approach employs

message passing through the Short Mail

Service (SMS), which is not costly to

process and compose.

Servers

Server in this regard refers to any

network environment supporting agent

operation for query issuance and

processing. The agent server is

responsible for service provision to the

mobile agents based on their requests.

Each server has a network

communicator employed for transferring

and receiving agents on one hand, and

processing of agents’ requests to/from

the owner or other collaborating agents.

The network communicator also

monitors agents’ attributes like resource

requirements, security issues e.t.c.

.

Security

The concepts of communication

and mobility would have been totally

rubbished if security were not involved.

Security is thus one of the most

important issues that can make or mar

the functionality and acceptability of a

mobile agent system. Imagine an agent

trying to interfere with the server host or

gaining unauthorized access to another

territory. There should be strict control

of sensitive information pullout. The

server is/and must be fully equipped to

authenticate the ownership of an agent,

its assigned level of access/permission

and also audit the agents’ execution.

This is achievable through the Java

Programming high security model.

3.2 Model Interactions and Notational

Representation

A large number of the mobile

agents will have their originating point

from the Mobile Client Model (MCM).

These not being unconnected with the

fact that most requests are queries

usually emanating from the mobile unit.

This is transported in the form of short

message services (SMS) having 5 tuples.

A typical example is given below.

Passcode#Pin#Rchg#Networkname#V

alue##

The tuple above depict 5

different values. The passcode among

other things consists of the user account

number in the database upon which this

MA architecture is ported. The PIN

(Personal Identification Number) is the

actual value that authenticates the owner

of the account so that even if somebody

else now owns the account number, the

PIN is still expected to be a private

(personal) value.

The admittance or non-

admittance thus rests solely on the

validity of both the PASSCODE and the

PIN. On confirmation of the above, the

mobile agent checks for the operation

required in the RCHG (recharge) value,

in this case, if the holder/user intend to

recharge its account, then it informs it of

how much (monetary value) the user is

intending to recharge with (purchase).

The NETWORKNAME informs the MA

of the type of network the user operates.

The VALUE now confirms the

presentation of the RCHG and then the

MA completes other required operation

leading to account recharge or denial or

other query processing .

Messages passing between the

366

mobile clients and the server must go

through the cloud of communication

network. Each tuple is assumed to take

insignificant time in passing through the

network. It is also expected that once the

message composition and onward

transmission is completed, bandwidth

can be released for other users until the

response to the request is ready before

another connection is ensured.

Once the message has

successfully gotten to the server, the

event monitor of the intelligent agent

structure picks the packet (SMS) and

logs it in the event database. The

housekeeping module gets hold of all

event packets and passed it to the

manager’s module. The manager module

after proper understanding of the request

of the agent’s packet thus pass it down to

the task interface that interconnects the

database where the mobile agent actually

complete its operation.

The following sequence of

operation is performed to facilitate the

agent admittance and performance of

any operation on the database.

Pass the SMS for conformity

Pick the passcode and pin to

authenticate (verify) the user

Collect the user information

from the database (compare with

the SMS)

Confirm the type of recharge

required by the user and value

Confirm that client’s account is

funded enough for transaction

Process vouchers type and

forward voucher number to the

client

Update database accordingly.

4.0 Algorithm For Query

Confirmation And Request Processing

Pass the SMS for conformity

If passcode and pin is valid then

Obtain user information

Confirm type of recharge required and

value

If client’s account balances >RCHG &

Minbal required then

Process voucher type and forward

voucher number to client

Update database accordingly

Else

Generate an invalid passcode or pin

message

End

5.Conclusion

In this paper , we have discussed

about the Global System for mobile

technologies and the architecture and

mechanisms that are supported a mobile

environment.

Mobile Agent technology

provided the increased bandwidth over a

telecommunication networks and how

the sms provided to the server without

any data missing through the networks

after checking the authentication details.

And proceed the algorithm for

query confirmation and request

processing. The objective is to depict the

effectiveness of mobile agent for

intermittent connectivity based queries.

In future work the proposed

system as comprising the of the client

side and facilitates the minimum power

consumption through MA

implementation.

367

REFERENCES

1. Baldi, M. and G.P. Picco (1998).:

Evaluating the Tradeoffs of

Mobile Code Paradigms in

Network Management

Application, in Proceedings of

the 20th International Conference

on Software (ICSE’98), April,

Kyoto, Japan, pp 146 – 155.

2. Adhicandra, I and C. Pattinson,

(2002): Performance Evaluation

of Network Management

Operations, in Proceedings of 3rd

Annual Symposium of

Postgraduate Networking

Conference (PGNET), Liverpool,

UK, pp 210 – 214.

3. Bieszczad, A., S.K. Raza, B.

Pagurek and T. White (1998a).:

Agent-Based Schemes for Plug

and Play Network Components,

in Proceedings of the Third

International Workshop on

Agents in Telecommunications

Applications (IATA ‘98), July 4-

7, Paris, France, pp 89 – 101.

4. Bates, J (1994).: The Role of

Emotion in Believable

Characters, Communications of

ACM, 37(7): pp 122 – 125.

5. Bieszczad, A., S.K. Raza, B.

Pagurek and T. White (1998b).:

Mobile Agent for Network

Management, IEEE

Communications Survey, 1(1) pp

2 – 9.

6. Bieszczad, A. and B. Pagurek

(1998).: Network Management

Application-Oriented Taxonomy

of Mobile Code, in Proceedings

of the EEE/IFIP Network

Operations and Management

Symposium (NOMS’98), Feb. 15

– 20, New Orleans, Louisiana, pp

659 – 669.

7. Baldi, M., S. Gai and G.P. Picco

(1997): Exploiting Code

Mobility in Decentralized and

Flexible Network Management,

in Proceedings of First

International Workshop on

Mobile Agents (MA’97), April,

Berlin, Germany, pp 13 – 26.

8. Appleby, S. and S. Steward

(1994): Mobile Agents for

Control in Telecommunications

Networks, British Technology

Journal, 12(2): pp 104 – 113.

9. Adhicandra, I., C. Pattinson, and

E. Shaghouei (2003): Using

Mobile Agents to Improve

Performance of Network

Management Operations. in

Proceedings of 4th Annual

Symposium of Postgraduate

Networking Conference

(PGNET).

10. Kaveh pahlavan and Prashant

Krishnamurthy - Principles of

wireless networks.

11. http://www.javasoft.com/security

- Java Security, Java Soft (1998)

12. Chess, D., B. Grosof, C.

Harrison, D. Levine, and C.

Parris (1995),: Itinerant Agents

for Mobile Computing, IEEE

Personal Communications, 2(5):

pp 34 – 49.

367

REFERENCES

1. Baldi, M. and G.P. Picco (1998).:

Evaluating the Tradeoffs of

Mobile Code Paradigms in

Network Management

Application, in Proceedings of

the 20th International Conference

on Software (ICSE’98), April,

Kyoto, Japan, pp 146 – 155.

2. Adhicandra, I and C. Pattinson,

(2002): Performance Evaluation

of Network Management

Operations, in Proceedings of 3rd

Annual Symposium of

Postgraduate Networking

Conference (PGNET), Liverpool,

UK, pp 210 – 214.

3. Bieszczad, A., S.K. Raza, B.

Pagurek and T. White (1998a).:

Agent-Based Schemes for Plug

and Play Network Components,

in Proceedings of the Third

International Workshop on

Agents in Telecommunications

Applications (IATA ‘98), July 4-

7, Paris, France, pp 89 – 101.

4. Bates, J (1994).: The Role of

Emotion in Believable

Characters, Communications of

ACM, 37(7): pp 122 – 125.

5. Bieszczad, A., S.K. Raza, B.

Pagurek and T. White (1998b).:

Mobile Agent for Network

Management, IEEE

Communications Survey, 1(1) pp

2 – 9.

6. Bieszczad, A. and B. Pagurek

(1998).: Network Management

Application-Oriented Taxonomy

of Mobile Code, in Proceedings

of the EEE/IFIP Network

Operations and Management

Symposium (NOMS’98), Feb. 15

– 20, New Orleans, Louisiana, pp

659 – 669.

7. Baldi, M., S. Gai and G.P. Picco

(1997): Exploiting Code

Mobility in Decentralized and

Flexible Network Management,

in Proceedings of First

International Workshop on

Mobile Agents (MA’97), April,

Berlin, Germany, pp 13 – 26.

8. Appleby, S. and S. Steward

(1994): Mobile Agents for

Control in Telecommunications

Networks, British Technology

Journal, 12(2): pp 104 – 113.

9. Adhicandra, I., C. Pattinson, and

E. Shaghouei (2003): Using

Mobile Agents to Improve

Performance of Network

Management Operations. in

Proceedings of 4th Annual

Symposium of Postgraduate

Networking Conference

(PGNET).

10. Kaveh pahlavan and Prashant

Krishnamurthy - Principles of

wireless networks.

11. http://www.javasoft.com/security

- Java Security, Java Soft (1998)

12. Chess, D., B. Grosof, C.

Harrison, D. Levine, and C.

Parris (1995),: Itinerant Agents

for Mobile Computing, IEEE

Personal Communications, 2(5):

pp 34 – 49.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

368

MITIGATING ROUTING

MISBEHAVIOUR IN MOBILE AD HOC

NETWORKS

M. Rajesh Babu

1 M. Sangeetha

2 R. Vidhya Prakash

3

1 Senior Lecturer, Department of CSE, PSG College of Technology, Coimbatore, [email protected]

2 PG Student, Department of CSE, PSG College of Technology, Coimbatore, [email protected] 3 PG Student, Department of CSE, PSG College of Technology, Coimbatore,

[email protected]

ABSTRACT

Security in MANETs is of prime

importance in several scenarios of deployment

such as battlefield, event coverage, etc. The

traditional non-secure routing protocols for

MANETs fail to prevent against attacks such as

DoS, spoofing and cache poisoning. One of the

primary goals of designing secure routing

protocols is to prevent the compromised nodes in

the network from disrupting the route discovery

and maintenance mechanisms. In this paper, we

describe a new authentication service for securing

MANET routing protocols. The focus of this work

is on securing the route discovery process in DSR.

Our goal is to explore a range of suitable

cryptographic techniques with varying flavors of

security, efficiency and robustness. This approach

allows the source to securely discover an

authenticated route to the destination using

message authentication codes (MAC) and

signatures.

Key words: AODV, DSR, Mobile Ad hoc Networks, Routing Protocols, Selfish Nodes, Security.

1. INTRODUCTION

Mobile ad hoc networks (MANETs) have

become a prevalent research area over the last couple

of years. Many research teams develop new ideas for

protocols, services, and security, due to the specific

challenges and requirements MANETs have. They

require new

concepts and approaches to solve the networking challenges. MANETs consist of mobile nodes which

can act as sender, receiver, and forwarder for

messages. They communicate using a wireless

communication link. These networks are subject to

frequent link breaks which also lead to a constantly

changing network topology. Due to the specific

characteristics of the wireless channel, the network

capacity is relatively small. Hence, very effective and

resource efficient protocols are needed to use MANETs with many nodes.

Since many nodes communicate over an air

interface, security becomes a very important issue.

Compared to a wired link, the wireless link can be

intercepted or disrupted by an attacker much more

easily, since it is freely accessible and not protected

at all. In addition, the constantly changing topology

makes it hard to determine which node really left the

network, just changed the location, or has been

intercepted or blocked. Due to numerous proposed attack scenarios, mechanisms and protocols have to

be developed to secure MANETs and utilize them,

e.g. in a commercial scenario.

Because of the changing topology special

routing protocols have been proposed to face the

routing problem in MANETs. Since routing is a basic

service in such a network, which is a prerequisite for

other services, it has to be reliable and trustworthy.

Otherwise dependable applications cannot be

provided over MANETs. This brings up the need for

secure routing protocols. A secure routing protocol has to be able to identify trustworthy nodes and find a

reliable and trustworthy route from sender to

destination node. This has to be realized within a few

second or better tenths of seconds, depending on the

mobility of the nodes and the number of hops in the

route.

369

Several security techniques have been

proposed, but the system may misbehave when an

attacker enters into the network and the network

performance is reduced. These nodes do not forward

packets properly. They may drop the packets that

forward across them or declare a faulty routing updates. Other attacks such as the information can be

read by the unauthorized persons in the network or

modified by the attacker nodes.

In this paper, we propose a routing protocol

that mitigates the routing misbehavior in MANETs.

In this protocol, the destination node should

authenticate the source node by using MAC. The

destination node should authenticate each intermediate node by using signatures. The node

sequence has to be verified by the source and

destination nodes.

2. BACKGROUND

Since our work is specific to DSR, this

section provides a brief re-cap of the DSR. DSR is a

purely on-demand ad hoc network routing protocol.

This means that a route is discovered only

when it is needed and no pre-distribution of

connectivity is performed. Since route discovery is

done by flooding, nodes do not accumulate network

topology information except for cached routes. DSR

includes two main mechanisms: route discovery and

route maintenance. Route discovery is used to discover a route from a given source to a given

destination, while route maintenance is used to

manage previously discovered routes. Since our focus

is on route discovery, we do not further discuss route

maintenance. Route discovery is composed of two

stages: route request (RREQ) and route reply

(RREP). Whenever a source needs to communicate to

a destination and does not have a route in its route

cache, it broadcasts a RREQ message to find a route.

Each neighbor receives the RREQ and appends its

own address to the address list in the RREQ and re-

broadcasts the packet. This process continues until either the maximum hop counter is exceeded or the

destination is reached. In the latter case, the

destination receives the RREQ, appends its address

and generates a route reply packet (RREP) back

towards the source using the reverse of the

accumulated route. Unlike RREQ, RREP percolates

towards the source by unicast. When the source

finally receives RREP, it stores the route in its route

cache.

Fig.1illustrates an example of route discovery shows the processing of RREQ and RREP packets.

3. RELATED WORK

Yih-Chun Hu, David B. Johnson and Adrian

Perrig, have designed and evaluated the Secure

Efficient Ad hoc Distance vector routing protocol (SEAD), a secure ad hoc network routing protocol

based on the design of the Destination-Sequenced

Distance-Vector routing protocol (DSDV) [1]. In

order to support use with nodes of limited CPU

processing capability, and to guard against Denial-of-

Service (DoS) attacks in which an attacker attempts

to cause other nodes to consume excess network

bandwidth or processing time, they have used the

efficient one-way hash functions and do not use

asymmetric cryptographic operations in the protocol.

SEAD has performed well over the range of

scenarios they have tested, and it was robust against multiple uncoordinated attackers creating incorrect

routing state in any other node, even in spite of any

active attackers or compromised nodes in the

network.

Tarag Fahad & Robert Askwith concentrate

on the detection phase and present a new mechanism

which can be used to detect selfish nodes in MANET

[3]. The detection mechanism is called Packet

Conservation Monitoring Algorithm (PCMA).

PCMA succeeded in detecting selfish nodes which perform full/partial packets attack. Though the

protocols address the issue of packet forwarding

attacks, it does not address other threats.

Yanchao Zhang, Wenjing Lou, Wei Liu, and

Yuguang Fang, propose a credit-based Secure

Incentive Protocol (SIP) to stimulate cooperation

C A

D F G

E B

RREQ

RREP

370

among mobile nodes with individual interests [4].

SIP can be implemented in a fully distributed way

and does not require any pre-deployed infrastructure.

In addition, SIP is immune to a wide range of attacks

and is of low communication overhead by using a

Bloom filter. Though the protocol addresses the issue of packet forwarding attacks, it does not address

other threats.

Liu, Kejun Deng, Jing Varshney, Pramod K.

Balakrishnan, Kashyap propose the 2ACK scheme

that serves as an add-on technique for routing

schemes to detect routing misbehavior and to

mitigate their adverse effect [5]. The main idea of the

2ACK scheme is to send two-hop acknowledgment packets in the opposite direction of the routing path.

In order to reduce additional routing overhead, only a

fraction of the received data packets are

acknowledged in the 2ACK scheme. But, the

acknowledgement packets are sent even though there

is no misbehavior, which results in unnecessary

overhead.

A. Patwardhan,J. Parker, M. Iorga, A.

Joshi,T. Karygiannis and Y. Yesha, present an

approach of securing a MANET using a threshold

based intrusion detection system and securing routing

protocol [6]. Their implementation of IDS deployed

on hand held devices and in a MANET test bed

connected by a secure version of AODV over IPV6.

While the IDS help detect attacks on data traffic, Sec

AODV incorporates security features of non-

repudiation and authentication, without relaying on

the availability of a Certificate Authority (CA) or a

Key Distribution Center (KDC). They have presented the design and implementation details of their

system, the practical considerations involved, and

how these mechanisms are used to detect and thwart

malicious attacks.

4. SYSTEM DESIGN AND ALGORITHM

OVERFLOW

We begin by stating some

environmental assumptions and summarizing our

notation. We assume bidirectional communication on

each link: if node S is able to send a message to node

D, then node D is able to send to node S. This

assumption is justified, since many wireless MAC-

layer protocols, including IEEE 802.11, require

bidirectional communication.

We assume that a node is aware of the exact

set of its current immediate neighbors.

The cryptographic techniques that we

propose further below fall into two categories:

shared-key MACs and public key signatures.

Although sometimes implemented with the aid of

conventional ciphers, MAC functions are often

constructed using cryptographically suitable keyed hash functions. The most common MAC construct is

the HMAC. For schemes based on MACs, we assume

the existence of a secure key distribution mechanism.

In our approach, each node will keep track

of number of packets forwards to the neighboring

nodes, and number of acknowledgements received

for the forwarded packets. When packet drop is more

in neighboring node, then it assumes that node as

misbehaving node.

In DSR, a route is accumulated incrementally through flooding until the destination

is reached, at which point the route is confirmed by

re-visiting it by unicast in reverse order.

Each node generates the Secret key and

public key, and then it distributes the public key to all

other nodes.

When a node wants to transfer a packet, it

first looks at its cache for the route. If any route is

found, it forwards the packet. If there is no pre-existing route, then the sender constructs the Route

Request.

The sender generates the Route Request

(RREQ) packet. In RREQ, it includes the source id,

destination id and route id. The sender computes

MAC over the route id and adds it in the packet

header and floods the RREQ to its neighboring nodes

except misbehaving nodes.

The intermediate node receives the RREQ

checks for the destination id. If it does not match with its id, then it adds its id and signature with RREQ and

then floods to its neighbors.

When the destination node receives the

RREQ packet, it checks for the identity of source by

computing the MAC over the route id by using the

key shared with the source. If the computed MAC is

same as that in the RREQ packet, then the destination

will verify the integrity of RREQ by using the public

key of each intermediate node. If the verification of

any intermediate node fails, then it discards the RREQ.

371

When the MAC computed over route id

does not matches with the MAC in destination then it

discards the RREQ.

The RREQ process is illustrated below:

SN1 : RREQ[S, D, MAC (Rid)]

N1N2 : RREQ[S, D, MAC (Rid), (N1),

(signn1)]

N2N3 : RREQ[S, D, MAC (Rid), (N1,N2),

(signn1,signn2)]

NM-1D: RREQ[S, D, MAC(Rid),(N1,N2,Nm-1),

(signn1,signn2,signm-1)]

The destination node then constructs the

Route Reply (RREP) by adding its id, source id, route

id and id of all the intermediate nodes. Then it forwards RREP to its neighbor node.

When the sender receives the RREP packet,

it repeats the same steps as the destination node. It

checks for authentication of the destination system by

checking with the MAC computed over route id, then

integrity of the intermediate nodes by using the

public key of intermediate nodes.

The RREP process is illustrated below:

DNM-1: RREP[S, D, MAC (Rid), (N1,N2,NM-1), (signD)]

N2N1: RREP[S, D, MAC (Rid), (N1,N2,NM-1),

(signD,signm-1,signn2)]

N1S : RREP[S, D, MAC (Rid), (N1,N2,NM-1),

(signD,signm-1,signn2,signn1)]

5. CONCLUSION

Mobile Ad Hoc Networks

(MANETs) have been an active area over the past

few years, due to their potentially widespread

application in military and civilian communications. Such a network is highly dependent on the

cooperation of all its members to perform networking

functions. This makes it highly vulnerable to selfish

nodes. One such misbehavior is related to routing.

When such misbehaving nodes participate in the

Route Discovery phase but refuse to forward the data

packets, routing performance may be degraded

severely. For robust performance in an untrusted

environment, it is necessary to resist such routing

misbehavior. A possible extension to DSR to mitigate

the effects of routing misbehavior in ad hoc networks has been proposed. It provides security against

routing misbehavior and attacks. It relies on

cryptographic techniques for authentication. The

security mechanism which is developed is highly

efficient. It also prevents a variety of DoS attacks.

REFERENCES

1. Farooq Anjum and Dhanant

Subhadrabandhu and Saswati Sarkar

“Signature based Intrusion Detection for

Wireless Ad-Hoc Networks: A

Comparative study of various routing

protocols” Vehicular Technology

Conference, 2003. VTC 2003-Fall. 2003 IEEE 58th, Oct. 2003.

2. M.Rajesh Babu and S.Selvan,”A Secure

Authenticated Routing Protocol for Mobile

Ad-Hoc Networks”, CIIT Journal, July

2009.

3. Yih-Chun Hu, David B. Johnson and Adrian

Perrig, "SEAD: Secure Efficient Distance

Vector Routing for MobileWireless Ad Hoc

Networks", in proceedings of IEEE

Workshop on Mobile Computing Systems

and Applications, pp.3-13, 2002.

4. Tarag Fahad & Robert Askwith “A Node

Misbehaviour Detection Mechanism for

Mobile Ad-hoc Networks” The 7th Annual

PostGraduate Symposium on The

Convergence of Telecommunications,

Networking and Broadcasting, 26-27 June

2006.

5. YihChun Hu, Adrian Perrig and David B. Johnson," Ariadne: A Secure On Demand

Routing Protocol for Ad Hoc Networks",

Technical Report, Rice university 2001

6. Yanchao Zhang, Wenjing Lou, Wei Liu, and

Yuguang Fang, “A secure incentive

protocol for mobile ad hoc networks”,

Wireless Networks (WINET), vol 13, issue

5, October 2007.

7. Liu, Kejun Deng, Jing Varshney, Pramod K.

Balakrishnan, Kashyap “An

Acknowledgment-based Approach for the

Detection of Routing Misbehavior in

MANETs” Mobile Computing, IEEE

Transactions on May 2007.

8. A. Patwardhan,J. Parker, M. Iorga, A.

Joshi,T. Karygiannis and Y. Yesha,

372

"Threshold-based intrusion detection in ad

hoc networks and secure AODV", Vol.6,

No.4, pp.578-599, 2008.

9. Katrin Hoeper and Guang Gong,

"Bootstrapping Security in Mobile Ad Hoc

Networks Using Identity-Based Schemes

with Key Revocation", Technical

Report CACR 2006-04, Centre for Applied

Cryptographic Research, January 2006.

10. Gergely Acs, Levente Buttya, and Istvan Vajda, "Provably Secure On- Demand

Source Routing in Mobile Ad Hoc

Networks", IEEE Transactions On

Mobile Computing, Vol. 5, No. 11, pp.1533-

1546, November 2006.

11. Syed Rehan Afzal, Subir Biswas, Jong-bin

Koh, Taqi Raza, Gunhee Lee, and Dong-

kyoo Kim, "RSRP: A Robust Secure Routing Protocol for Mobile Ad hoc

Networks", IEEE Conference on Wireless

Communications and Networking, pp.2313-

2318, April 2008.

AUTOMATIC CALL TRANSFER SYSTEM BASED ON LOCATION PREDICTION USING

WIRELESS SENSOR NETWORK

Vairavamoorthy.A#1, Somasundaram.C#2, Jesvin Veancy#

Department of Electronics and Communication Engineering, Easwari Engineering College

Approved by AICTE Accredited by NBA, Affiliated to Anna University, Chennai and ISO 9001-2000 Certified [email protected]

[email protected]

Abstract—The Public Switched Telephone Network

(PSTN) is the oldest and largest tele-communications

network in existence. The auto call-transfer function

is one of the basic services supported by PSTN that

switches incoming calls to the subscriber. It is likely

for the subscribers to lose incoming calls when they

leave the room. Therefore, we propose a ZigBee

location-awareness and tag identification technique

for the traditional phone system. Our technique

identifies a person’s location as he/she moves around

the office building. Whenever there is an incoming

call, the call will be transferred to the office phone

nearest the person. As a result, the overall service

quality would be enhanced to practical effect.

I. INTRODUCTION

The current telephone systems can be grouped into

PSTN, GSM and the new digital technology SIP

(VOIP) and Skype systems. These systems

intercommunicate through different gateway

interfaces. PSTN has been around for hundreds of

years and remains high in market shares. Therefore,

integration into PSTN is a common goal for the rest

of the systems. PSTN provides an important

function: the auto call-transfer service. This service

switches unanswered incoming calls to a back-up

position after a pre-determined interval of time. (If

a call is not answered at phone X after a set time,

the call will be transferred to Y). Currently, the

GSM system can automatically transfer a PSTN

call to a predestined cell phone line. However, the

service charges an additional fee from the cell

phone customers. A number of studies have sprung

to study RFID (Radio Frequency Identification)

and Zigbee technology and aimed to reach the goal

of positioning. RFID is an auto-identification

technology using radio to identify objects. Its

principal operation theory is to use a reader for

transmitting radio signals and reading the tag

embedded or posted on the objects. RFID uses the

radio data to trace its objects using the reader as a

positioning reference. It has become a common

practice to include these indoor positioning devices

in both wired and wireless networks. However,

because the wireless network is built upon the

backbone network of the wired, the infrastructure

of the wireless cannot be established without the

wired network. The other auto identification

system, Zigbee, has a self-built network that can be

constantly updated. It also has low power

consumption. Zigbee uses the intensity of the

received signals from the laidout sensors to identify

the position of the mobile devices. In this study, we

propose a location-awareness technique in the

traditional phone system that offers a call-transfer

service by locating the person moving around with

an RFID and Zigbee. Our system also supports

Mesh and provides Multi-path functionality to

enhance network stability. The rest of the paper is

organized as follows. In Section II gives a

description of the system architecture. Section III

presents the experiment environment and the

simulation result. Finally, Section IV concludes

this work.

II. SYSTEM ARCHITECTURE The functional block diagram of the cabin side and

system side are shown in the fig. 1 and fig. 2

respectively.

Fig.1 Cabin side

Proceedings of the third National conference on RTICT 2010Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

373

Fig. 2 System side

In cabin side, RFID is used to monitor the

position of person. If person in Cabin1 then system

sends that corresponding information about that

person to system. In the system side decision will

be taken for call. It monitors all persons where they

are present using zigbee communication.

Our system consists of a user’s hand-held

Zigbee tag. It is moved in a pre-defined area, when

the set reference point detects the tag. After the

positioning system concludes its correct position,

the detected position is sent by the Application

System to the DataBase for cross comparison to

acquire the extension number of the position, the

application system sends out transfer instruction by

way of communication interface to PABX for auto-

transfer service. The system architecture is

illustrated in Fig.3.

Fig. 3: System Architecture

A. Location-Awareness

Our system adopts a signal intensity ranging

algorithm. The signal intensity received by the tags

set up at different reference points is used to do

position analysis. The measured data of single

signal intensity could only lock the mobile stations

on a track which will take the reference point as

center position. The radius of a circle could be

confirmed by signal strength, when there are three

reference points, the received signal could

accurately tell the position of the mobile device.

B. Application Systems

We use the SCADA (Supervisory Control And

Data Acquisition) as the main screen in this study.

It monitors both the real time moving status of

different devices and the RSSI value received by

each reference point. With the calculation

positioning system, the position of the tag will be

accurately shown on the graphic interface, and the

tag number will be displayed. At the same time, the

new data of each mobile device are recorded (Tag,

Time and Position). When a “moving event” takes

place, the system will inquire DB (DataBase). This

inquiry begins with the tag. The tag gets the local

phone number and the phone number of the new

position, and then sends out the instructions to the

PABX system by passing through the transfer

service. Every step of the transfer process can be

monitored in real time through the screen.

C. PABX Transfer Services

The telephone system used in this study is Phillip’s

traditional enterprise-type PABX with 60 outbound

lines and 1500 inbound lines. The system provides

Command Line instruction that can be connected

through RS-232. The transmission interface is set

to 9600 bps (bits per second), 8 bits, and non-

parity. The instruction of ChFoME (Change Follow

Me) is used as the phone transfer setup.

Change follow me relation

CHFLME: <FM-TYPE>, <ORIG-DNR> [,<DEST-

NUMBER>];

<FM-TYPE>: 0: when has Ring starts transfer

<ORIG-DNR>:Original telephone number

<DEST-NUMBER>:Destination telephone

number

<;>:Execution

Example:

chflme: 0,503300,503201;

When have dial telephone number is 503300 will

be transfered to 503201.

III. EXPERIMENTAL RESULTS This section evaluates the automatic phone transfer

service based on the location-awareness technology

by setting up actual calls. The parameters of the

test environment are shown in Fig. 4. An additional

phone is used to call into these people’s offices.

Ten calls have been used to connect each of the

five telephones, and each call is unmistakenly

transferred to the correct person in his/her new

location. For example, the person with the tag ID

101 gets the call at the destination office with the

new phone number 503201 that has been

transferred from the original phone line.

374

Fig. 4: Test Environment

We test our system through the transfer

between inbound lines. First, the person who takes

the 101 tag moves from room 101 to room 110.

When the position system senses that 101 has been

moved into room 110, the auto call-transfer service

resets the phone number from 503300 to 503201.

The system was tested by using the phone in room

105 to dial 503300. The phone number 503300

connects room 110 instead of 101.In contrast, when

the person goes back to room 101, the Change

Follow Me service would be cancelled. If we dial

the number 503300 from another line, the phone

would ring in room 101 as usual.

IV. CONCLUSION AND FUTURE WORK In this study, we apply the State-of-the-Art

positioning technology to the telephone auto call-

transfer service. We aim to reduce the chance of

missed calls as one leaves his/her work area. Our

system uses Zigbee to build up a positioning

sensing system that detects any new position when

one leaves his/her office and transfers the phone

call to the phone at the new location. We will

integrate SIP (Session Initiation Protocol) into our

system to provide an analog/digital

communications in the auto call-transfer service in

the future.

REFERENCES

[1] W. Jiang, D. Yu, and Y. Ma, “A Tracking

Algorithm in RFID Reader Network,” Proceedings

of the Frontier of Computer Science and

Technology, pp. 164-171, Nov. 2006.

[2] K. Kaemarungsi, “Design of Indoor positioning

system based on location fingerprint techinique”.

University of Pittsburgh, 2005.

[3] W. Klepal and M. P. Dirk, “Influence of

Predicted and Measured Fingerprint on the

Accuracy of RSSI-based Indoor Location

Systems,” Proceedings of the Positioning,

Navigation and Communication, pp. 145-151, Mar.

2007.

[4] W. C. Park and M. H. Yoon, “The

Implementation of Indoor Location System to

Control ZigBee Home Network,” Proceedings of

the International Joint Conference, pp.2158-2161,

Oct. 2006.

[5] J. Rosenberg and H. Schulzrinne, “SIP: Session

Initiation Protocol,” IETF RFC 3261, 2002.

[6] E. D. Zand, K. Pahlavan, and J. Beneat,

“Measurement of TOA using frequency domain

characteristics for indoor geolocation,” Proceedings

of the IEEE Indoor and Mobile Radio

Communications, pp. 2213-2217, Sep. 2003.

[7] Y. Zhang “SIP-based VoIP network and its

interworking with PSTN,” Electronics &

Communication Engineering Journal, pp.273-282,

OCT. 2002.

[8] J. Zhao, Y. Zhang, and M. Ye, “Research on the

Received Signal Strength Indication Location

Algorithm for RFID System,” Proceedings of

International Symposium on Communications and

Information Technologies, Oct. 2006.

[9] Skype Developer Zone. [online], available at <

https://developer.skype.com/>.

375

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

376

An Energy Efficient relocation of gateway

for improving Timeliness in Wireless Sensor

Networks

S.S.Rahmath Ameena* Dr.B.Paramasivan

**

* Student, Dept. of CSE, National Engineering College, Kovilpatti.

E-Mail: [email protected] **Professor, Dept.of CSE, National Engineering College, Kovilpatti.

E-Mail: [email protected]

ABSTRACT:

In Wireless Sensor Networks (WSN) collected

data are routed to multiple gateway nodes for processing.

Based on the gathered reports the gateways may also

work collaboratively on a set of actions to serve

application level requirements or to better manage the

resource constrained WSN. In this paper an efficient

algorithm of AQM (Active Queue Management) is

presented. AQM strives to maintain buffer management

among the gateways while repositioning individual

gateway to better manage the sensors in their vicinity.

Simulation results have demonstrated the effectiveness of

AQM and positive effect on network longevity.

Keywords: Wireless Sensor Networks, AQM, Timeliness, End

to End delay.

1. INTRODUCTION

Wireless sensor networks(WSN) have

numerous applications in a variety of disciplines, both

military and civilian [1] [2].The ability to remotely

measure ambient conditions and track dynamic events become valuable especially in harsh environments

where human intervention is risky or infeasible.

Sensors are usually battery-operated and have a limited

transmission and processing capacity. Such constraints

have motivated lots of research on data fidelity, latency,

and coverage so that the network can stay functional for

the longest duration.

A typical WSN architecture involves large

numbers of sensor nodes that report their measurements

to locally deployed data collection centers often

referred to as gateway node. Gateway is usually more capable in terms of their energy supply, radio range and

computational resources.

Gateway relocation [3] is one of the

approaches pursued to improve the performance of the

network. By relocating the gateway towards highly

active sensors, the packet will be routed to the gateway

through fewer sensors. The shortened data paths help in

preserving sensors energy, lowering the packet loss rate

and reducing delivery latency. In general, most gateway

relocation techniques identify bottlenecks in the current

network topology and tend to move the gateway close

to the bottleneck positions. However, such

performance-centric relocation may move the gateway

dangerously close to one or multiple targets/events in the environment and thus may expose the gateway to

the risk of getting damaged, captured etc.

In this paper we propose a AQM (Active Queue

Management) algorithm solution to the gateway

relocation problem .The idea is to identify a position for

the gateway to better serve wireless sensor network

which consists of a large number of sensor nodes and

one or more sink nodes scattered over a region of

interest where data gathering and reporting tasks are

performed. A sensor node often has limited power

resources. Therefore it is very crucial to reduce the

energy consumption in certain real-time applications such as fire alarm monitoring, traffic monitoring

information collected is valid only for a limited amount

of time after that it become irrelevant information.

Hence all the packets need to be conveyed to the sink

within a certain deadline period and packet timeliness

requirement also becomes an important issue for such

real-time applications in wireless sensor network.

Sensor networks data are routed towards a

single sink the gateway in our model hops close to that

gateway become heavily involved in packet forwarding

and thus their energy resource gets depleted rather quickly. Such scenario increases total transmission

power and gradually limits sensor coverage in the

network and eventually makes the networks useless. If

the gateway has limited motion capability it will be

desirable to relocate the gateway close to an area of

heavy traffic or near loaded hops in order to decrease

the total transmission power and extend the life of

nodes on the path of heavy packet traffic.

377

A significantly less energy constrained

gateway node than all the sensors is deployed in the

physical proximity of sensors. It is assumed to know the

geographical location of sensors and it is responsible

for organizing the activities at sensor nodes fusing

collected data coordinating communication and interacting with the command node. This model is

constructed for stationary sensor nodes with gateway

having limited mobility. The gateway remains

stationary unless the network operation becomes

inefficient. The gateway relocates to another position

in-order to optimize performance metrics such as

maximizing network life-time, sensors are assumed to

be within the communication range of the gateway node

and capable of operating in an active mode or a low-

power stand-by mode. The transmission power of the

transmitter and receiver can be programmed so that the

transmission can be turned on/off as required. This sensor acts as a relay node to forward data.

2. RELATED WORKS

SPEED [5] protocol is a highly efficient and

scalable protocol for sensor networks where the

resources of each node are scarce. SPEED that supports

soft real-time communication based on feedback

control and stateless algorithms for large-scale sensor

networks. SPEED helps balance the traffic load to increase the system life-time. Speed also utilizes

geographic location to make localized routing

decisions. Speed algorithm is to support a soft real-time

communication service with a desired delivery speed

across the sensor network. SPEED provides three types

of real-time communication services, namely real-time

unicast, real-time area multi-cast, and real-time area

any cast for sensor networks. SPEED satisfies the

following design objectives such as, stateless

architecture, soft real-time, minimum MAC layer

support, QoS routing and congestion management,

Traffic load balancing, Localized behavior, void avoidance. However, the drawback of SPEED protocol

provides only one network-wide speed, which is not

suitable for differentiating various traffic with different

deadlines. It has used to only a single path routing then

on demand algorithm is less suitable for real-time

applications.

MMSPEED[6](Multi-path and Multi-speed)

protocol is designed to provide relies on the premise

that the underlying MAC protocol can perform the

following functions prioritized access to shared medium

depending on the speed layer .Reliable multicast delivery of packets to multiple neighbors supporting

measurement of average delay to individual neighbors.

MMSPEED differentiates packets depending on their

reliability requirement and thus we can drop more

packets with low reliability requirements than SPEED.

However the major drawback of this protocol is

increase the overall power consumption by sensor node

radio modules.

WFQ[7] method is flow based by considering

each imaging sensor node as a source of different real-

time flow with only one real-time queue to

accommodate the real-time data coming from these multiple flows. The service ratio “r” is the bandwidth

ratio set by the gateway and is used in allocating the

amount of bandwidth to be dedicated to the real-time

and non-real-time traffic on a particular outgoing link

.However the drawback of Weighted Fair Queuing

buffer management is very poor. So we cannot reduce

the end-to-end delay the energy consumption is very

high.

MAC layer multicast protocol which uses a

separate RTS/CTS hand shake for each of the

recipients, followed by data transmissions and another

sequence of Request to Acknowledge (RAK)/ACK handshakes to ensure the reliable delivery to all

multicast recipients. However, this incurs a long

sequence of resulting in a long delay for each multicast.

This MAC protocol ensures the reliable frame

transmission only to primary recipient expecting

successful eavesdropping by all other recipients.

3. PROPOSED PROTOCOL

I. Sensor Architecture: The no of sensor nodes deployed are generally

equipped with data processing and communication

capabilities. The sensors send its collected data, usually

via radio transmitter to a command center either

directly or through a base station. The gateway can

perform fusion of the sensed data in order to filter out

erroneous data and anomalies to draw conclusions from

the reported data over a period of time.

Fig 3.1: System Model

The system architecture of sensor network

system is depicted in fig 3.1

Sensor Nodes

G

Gateway Node

Base station

G

G

G

G

378

Fig 3.2: Process flow diagram of gateway

relocation

II. Cluster Formation:

The sensors are uniformly deployed and the

cluster is formed using the location based clustering

algorithm. This algorithm fully utilizes the location

information of routing to reuse the routing cost. The

cluster is formed by following equation

Fns1rs1

snT

))/mod(()(

0 Others

Where s - Percentage of cluster heads

over all nodes in the network

r - The number of rounds of

selection

F - The set of nodes that are not

selected in round 1/s

III. Gateway Election:

Gateway election is more important in wireless

sensor networks. Gateway is also called as cluster head.

Gateway is elected for the context information of all the

sensor nodes which is based on all the sensor node energy and bandwidth coverage.

IV. AQM Implementation:

In order to solve the buffer management

problem and then decrease the end-to-end delay we

propose a Active Queue Management (AQM) based

phenom routing protocol for sensor networks. Active

Queue Management provides congestion signal for flow

control not only based on the queue length but also the

channel condition and the transmission rate. It uses

priority based queue management to improve

performance in congested area. Once the nodes are

deployed the sensor nodes form the cluster members

and elect a cluster head using location based clustering

algorithm. Next elect a cluster head as the gateway and

apply the AQM method where the buffer management is high by comparing the weighted fair queuing method

.Hence avoids more no of packet loss.

v. Relocation Approach to Enhance Performance :

There are three phases in gateway relocation

depicted in fig 3.2. These are when to relocate the

gateway, where to put it and how to handle its motion

without disrupting data traffic and negative effect.

Phase I: The time when gateway should be relocated Step 1: Always relocation is decided based on the

missed ratio.

Step 2: The missed ratio is indirectly related to finding proper r-value in the above equation.

Step 3: Real time routing finds the r-value by

considering all the paths from the source of real time

traffic to the gateway.

Step 4: In cases where a proper r-value between 0 and

1 cannot be found, the connection is simply rejected

and the path is not established

Step 5: Even where proper r-values are found, the miss

ratio may start to decrease.

Step 6: The gateway sets a threshold value for the

maximum level of miss ratio and maintains such statistics periodically when such threshold is reached

the gateway is to a better location.

Step 7: The gateway can be relocated more than once

whenever necessary during the data traffic cycles in fig

3.3

Phase II: The position to which the gateway should

be located

Initially Sensor nodes

are deployed

Last hop node are

assuming gateway

Gateway relocation

should be decided

Based on performance miss ratio.

Gateway can be

relocated more than once

During data traffic cycles

Fig 3.3: When to relocate the gateway

Nodes are connected multihop

path

Sensor Architecture

Cluster Formation

Cluster Head Election

AQM Implementation

Cluster Head Mobility

Find Relocation

Performance Analysis

Context

Information

Traffic

Monitoring

379

Step 1: After deciding that the gateway is to be

relocated then a new location is searched for it.

Step 2: Next the gateway is moved towards the loaded

nodes interms of real time traffic so that the end-to-end

delay can be decreased.

Step 3: Then next searches the last hop nodes. Step 4: The biggest r-value are considered for

relocating the gateway at the position of that hop.

Step 5: Decrease the average end-to-end delay since the

number of hops for data packets to travel will be

decreased.

Step 6: Breaking the ties when multiple alternative

nodes with the same r-value are found

Step 7: If the gateway is still reachable by the nodes Y

and Z they just increase their transmission range of

uninterrupted data delivery.

Phase III: How to relocate the gateway Step 1: After determined the new location, the gateway

explores two options based on the information of

whether it will be out of range at the new location or

not.

Step 2: If the gateway detects that it would go out of

the transmission range of last hop nodes and cannot

receive the data from other relay nodes at the new

location.

Step 3: Then explores the option of employing sensor

nodes to forward the packets.

Step 4: The miss ratio for real time data periodically to detect situations where there is need for relocation.

Step 5: Then final gateway are communicating to

command node.

Step 6: Next gateway checks whether it can still be

reachable by by the last hopes while traveling on the

next stride and inform yht last hop nodes about its

situation.

Step 7: Once,it detects that forwarder nodes are needed,

the routes are extended by those nodes and that

information is sent to relevant nodes.

4. PERFORMANCE RESULTS

In this section the performance results were

shown from fig5.1 to 5.4 with and without repositioning

the gateway. In order to see the effect of repositioning

the gateway on energy consumption, we looked at the

average lifetime of a node, time for first and last node

to die and average energy consumed per packet. The results depicted in fig1 have shown that our approach of

repositioning the gateway decreases energy

consumption significantly especially by increasing the

average lifetime of the node and time for first node to

die. In this approach, since the gateway is relocated

close to heavily loaded nodes, less energy is consumed

for communication thus leading significant amount of

energy savings. On the other hand, the latency of the

packets from those nodes to the gateway will be

decreased, causing the decrease delay per packet to

decrease drastically. There is also a slight increase on

the throughput due to smaller packet drop probability since packets travel shorter distances.

Table1: Simulation Environment Settings

Terrain 500mx500m

Node number 50

Node placement Uniform

Bandwidth 100 kbps

Inter-column spacing 82m

Inter-row spacing 70m

Simulation duration 10s

Transmission rate 10ms/pkt

Table 2: Movable Gateway

ST TRE ARE PDR AE2E

D

10 1069 106.9 98.03 0.157

20 2120 212 98.69 0.109

30 3180 318 99.02 0.085

40 4230 423 99.22 0.071

50 5281 528.1 99.35 0.062

60 6331 633.1 99.44 0.056

70 7390 739 99.51 0.051

80 8435 843.5 99.57 0.047

90 9486 948.6 99.61 0.044

Table 3: Fixed Gateway

ST TRE ARE PDR AE2ED

10 1070 107 97.69 0.1563

20 2129 212.9 97.30 0.1227

30 3179 317.9 97.17 0.1122

40 4230 423 97.5 0.1072

50 5275 527.5 97.69 0.1041

60 6340 634 97.82 0.1017

After increasing packet losses on some

data path

Next continuing packet delivery to

gateway node

Then gateway motion handling

Next involving data transfer

Fig 3.4 :How to relocate the

gateway

380

70 7381 738.1 97.91 0.1007

80 8431 843.1 97.88 0.0995

90 9498 949.8 98.03 0.0983

Where ST - Simulation Time

TRE – Total Remained Energy

ARE - Average Remained Energy

PDR – Packet Delivery Ratio AE2ED –Average End to End Delay

Fig 4.1: Total Remained Energy

Fig 4.2: Average Remained Energy

Fig 4.3: Packet Delivery Ratio

Fig 4.4: Average End to End Delay

5. CONCLUSION In this paper, a repositioning approach is

presented for a gateway in order to enhance the overall

performance of a wireless sensor networks in terms of

popular metrics such as energy, delay and throughput.

The presented approach considers relocation of the

gateway by checking the traffic density of the nodes

that are one-hop away from the gateway and their

distance from the gateway. Once the total transmission

power for such is guaranteed to reduce more than a

certain threshold and the overhead of a moving the gateway is justified, the gateway starts to move to the

new location. The gateway is moved in the routing

phase so that the packet relaying will not be affected.

Simulation results have shown that such repositioning

of the gateway increases the average lifetime of the

nodes by decreasing the average energy consumed per

packet. Moreover, the average delay per packet is

decreased significantly.

REFERENCES

[1] I.F.Akyildiz et al.,“Wireless sensor networks:asurvey”,Computer

Networks,Vol.38,pp.393-422,2002.

[2]Bogdanov,A. Maneva,E. Riesenfeld,S.,“Power Aware Base

Station Positioning for Sensor Networks”, Twenty-third Annual Joint

Conference of the IEEE Computer and Communications Societies,

Vol:1, pp:-585,2004.

[3] English,J. Wiacek,M. Younis,M, “CORE : Coordinated

Relocation of sink nodes in Wireless Sensor networks “,23rd Biennial

Symposium on Communications

pp:320-323,2006.

[4] Waleed Youssef and Mohamed Younis ,”A Cognitive Approach

for Gateway Relocation in Wireless Sensor Networks”,2006

[5] Tian He John A Stankovic Chenyang Lu Tarek

AbdelZaher,”SPEED : A Stateless Protocol for Real-Time

Communication in Sensor Networks “in the proceeding of

ICDCS.Providence,RI,May 2003.

[6] Emad Felemmban,Chang-Gun Lee,Eylem Ekici, ,”MMSPEED :

Multipath Multi-SPEED Protocol for QoS Guarantee of Reliability

and Timeliness in Wireless Sensor Networks” IEEE

381

TRANSACTIONS ON MOBILECOMPUTING,volume:5,No.6,

2006.

[7] Ramon Serna Oliver and Gerhard Fohler,” A Proposal for a

Notion of Timeliness in Wireless Sensor Networks”,2004

[8]Boonma,P. Suzuki,J.,”LeveragingBiologically-inspired Mobile

Agents Supporting Composite Needs of Reliability and Timeliness in

Sensor Applications”, FBIT 2007 Frotiers in the Convergence of

Bioscience and Information Technologies.Pp:851-860,2007.

[9] Tian He Stankovic, J.A. Chenyang Lu Abdelzaher, T,”

SPEED: a stateless protocol for real-time communication in sensor

networks”, 23 rd International Conference on

DistributedComputingSystems,2003,19-22May2003, On page(s):46-5

[10] Kamal Akkaya Mohamed Younis and Meenakshi Bangad,”Sink

Repositioning for Enhanced Performance in Wireless Sensor

Networks”, 2006.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

382

IMPROVING STABILITY OF SHORTEST MULTIPATH

ROUTING ALGORITHM IN MANET

S.Tamizharasu

M.E. Software Engineering

Anna University

Jothipuram, Coimbatore

[email protected]

ABSTRACT

To propose multipath routing

scheme, referred to as shortest multipath

source (SMS) routing based on dynamic source

routing (DSR) is proposed here. To avoid the

root flop by identify the node weight and

transfer the data from source to destination.

The mechanism that compare AOMDV

routing schemes with DSR routing schemes

along with the performance metrics viz:

throughput, end-to-end delay, delivery ratio,

and distance calculation.

The SMS achieves shorter multiple

partial-disjoint paths and allows more rapid

recovery from route breaks. The performance

differentials are investigated using NS-2 under

conditions of varying mobility, offered load

and network size. Results reveal that SMS

provides a better solution than existing source-

based approaches in a truly mobile ad-hoc

environment.

M.Newlin rajkumar M.E.

Lecturer

Dept. of Computer Science and Engineering

Anna University

Jothipuram, Coimbatore

[email protected]

1. AN INTRODUCTION TO MOBILE AD

HOC NETWORK

Mobile ad-hoc networks (MANETs)

are a key part of the ongoing evolution of wireless

communications. MANETs are defined as the

category of wireless networks that utilize multi-

hop radio relaying and are generally capable of

operating without the support of any fixed-line

infrastructure. Each node has the capability to

communicate directly with other nodes, acting not

only as a mobile wireless host but also as a router,

forwarding data packets to other nodes. In other

words, such networks are self-creating, self-

organizing and self-administrating. Key

application areas for these types of networks

include: conference networking, disaster relief,

military networks and sensor networks.The

network topology may be subject to numerous,

frequent and random changes. In a network

consisting of mobile nodes, the connection

between a source–destination pair may break and

require to be updated regularly. Routing should

adapt to such dynamics and continue to maintain

connections between the communication nodes in

the presence of path breaks caused by mobility

and/or node failures. A variety of routing

protocols for MANETs have been previously

proposed.

383

2 SHORTEST MULTIPATH SOURCE

Shortest Multipath Source SMS

computes multiple partial-disjoint paths that will

bypass at least one intermediate node on the

primary path. An example that illustrates the

computation of partial-disjoint paths is shown in

Figure 1. In case of a link failure between A and

B (Figure 2), the source node will search for an

alternate route that does not contain node B. In

this case, an alternative route between source and

destination is S-A-G-C-D. The major advantage

of such a protocol is that the shortest alternative

routes by contrast to the routes computed using

either link-disjoint or node-disjoint paths are

selected.

Figure 1 Partial disjoint multiple paths

Figure 2 Selection of alternative

path in case of link failure

Route Discovery

Route Discovery has two basic phases;

route request propagation and route reply.

Route Request Propagation

When a node seeks to communicate with

another node, it searches its cache to find any

known routes to the destination. If no routes are

available, the node initiates route discovery by

flooding route-request packet, containing an ID

which, along with source address, uniquely

identifies the current discovery. The SMS scheme

introduces a novel approach to reduce broadcast

overhead. Each node records the ID of route-

request packet and the number of hops the route

request traversed from the source node. Nodes use

this information to rebroadcast duplicate route-

request packets only if: number of hops ≤ last

hop-count recorded.

Route Reply Propagation

When the node receives the route

request, it generates a route reply to the source

using the reverse path identified within the route

request. In addition, the node records the reverse

path within its internal cache. Upon receiving

multiple copies of route-request packets of the

same session, the node sends limited replies to

avoid a reply storm. These new paths are also

appended to the internal cache. In SMS, the

source is responsible for selecting and recording

multiple partial disjoint route paths.

Route Maintenance

In the event of a link failure or an

intermediate node moving out of range, a route-

error packet is transmitted. The reception of a

route-error packet will invalidate the route via that

link to that destination and will switch to an

alternative path if available. The source node will

select an alternative path that does not contain the

next node of the one that sent a route-error packet.

When all routes in the cache are marked as

invalid, the node will delete the invalid routes and

a new route discovery is instigated.

384

3. ROUTING OVER HEAD DETAILS

DSDV-SQ are plotted on a the same

scale as each other, but AODVLL and TORA are

each plotted on different scales to best show the

effect of pause time and offered load on overhead.

TORA, DSR, and AODV-LL are on-demand

routing protocols, so as the number of sources

increases, to expect the number of routing packets

sent to increase because there are more

destinations to which the network must maintain

working routes. DSR and AODV-LL, which use

only on-demand packets and very similar basic

mechanisms, have almost identically shaped

curves. Both protocols exhibit the desirable

property that the incremental cost of additional

sources decreases as sources are added, since the

protocol can use information learned from one

route discovery to complete a subsequent route

discovery. However, the absolute overhead

required by DSR andAODV-LLare very different,

with AODV-LL requiring about 5 times the

overhead of DSRwhen there is constant node

motion (pause time 0). This dramatic increase in

AODV-LL‟s overhead occurs because each of its

route discoveries typically propagates to every

node in the ad hoc network.

3. 1 PACKET DELIVERY RATIO DETAILS

The fraction of the originated application

data packets each protocol was able to deliver, as

a function of both node mobility rate (pause time)

and network load (number of sources). For DSR

and AODV-LL, packet delivery ratio is

independent of offered trafficload; with both

protocols delivering between 95% and 100% of

the packets in all cases.DSDV-SQ fails to

converge below pause time 300, where it delivers

about 92% of its packets. At higher rates of

mobility (lower pause times), DSDV-SQ does

poorly, dropping to a 70% packet delivery ratio.

Nearly all of the dropped packets are lost because

a stale routing table entry directed them to be

forwarded over a broken link. The DSDV-SQ

maintains only one route per destination and

consequently, each packet that the MAC layer is

unable to deliver is dropped since there are no

alternate routes.

3.2 SHORTEST MULTIPATH ROUTING

Shortest Multipath Routing (SMR) is an on-

demand routing protocol that builds multiple

routes using request/reply cycles. When the

source needs a route to the destination but no

route information is known, it floods the ROUTE

REQUEST (RREQ) message to the entire

network. Because this packet is flooded, several

duplicates that traversed through different routes

reach the destination. The destination node selects

multiple disjoint routes and sends ROUTE

REPLY (RREP) packets back to the source via

the chosen routes.RREQ Propagation The main

goal of SMR is to build maximally disjoint

multiple paths.To construct maximally disjoint

routes to prevent certain nodes from being

congested, and to utilize the available network

resources efficiently. To achieve this goal in on-

demand routing schemes, the destination must

know the entire path of all available routes so that

it can select the routes. Therefore, by use the

source routing approach where the information of

the nodes that consist the route is included in the

RREQ packet. Additionally, Intermediate nodes

are not allowed to send RREPs back to the source

even when they have route information to the

destination. If nodes reply from cache as in DSR

and AODV it is difficult to establish maximally

disjoint multiple routes because not enough

RREQ packets will reach the destination and the

destination node will not know the information of

the route that is formed from the cache of

intermediate nodes.

3. 3 FINDING LINK DISJOINT PATHS

Path disjointness has the nice property

that paths fail independently. There are two types

of disjoint paths: node-disjoint and link-disjoint.

Node-disjoint paths do not have any nodes in

common, except the source and destination. In

contrast, link-disjoint paths do not have any

common link. Note that link-disjoint paths may

have common nodes. After investigated the issue

of node- vs. linkdisjointness carefully. Generally

speaking, using linkdisjointness is fine as long as

links do fail independently. This is often the case,

except in situations where multiple links from/to a

node fail together as the node moves out of range.

385

4 RELATED WORKS

AOMDV

AOMDV is a multipath routing

protocol which is based on AODV. When each

node receives duplicate copies of RREQ, it checks

whether the packets‟ hop counts are less than the

hop count already in a routing table. This check

avoids routing loops. Also, if this check is passed,

additional route information is added to the

routing table.

Multiple Routes Establishment by RREQ/RREP

Enhancement

To extend RREQ/RREP messages by

forcing source routing information in order to

establish valid multiple routes. Note that, unlike

other source routing protocols like DSR, to use

source routing information only in the case of

RREQ/RREP messages. Any data packets are

forwarded in a hop-by-hop manner without source

routing option similar to AODV.

RREQ Extension

In this extension, each node which

forwards RREQ packets inserts its own IP address

into a packet header. After receiving a RREQ

packet, it also checks source routing information

field in the packet. If there exists its own address,

it detects routing loop and discards the RREQ

packet. Otherwise, it updates its routing table with

the RREQ information. Unlike AODV, duplicate

copies of RREQ are not discarded immediately. If

the second path information of the routing table is

empty, and loop check is passed, the RREQ

information is adopted as the second reverse path.

RREP Extension

Similar to the RREQ extension, each

node that forwards RREP packets inserts source

routing information and updates its routing table

accordingly. The RREP packets are forwarded

along reverse routes. If there are multiple (two)

reverse routes, the node bicasts RREP packets to

each of them. The call direction going to a

destination “upstream”, and direction returning to

a source node “downstream”. When a node

receives a RREP packet, it checks that next hop (1

or 2) of the reverse route (to the node S) exists in

an address information field in the RREP packet.

If the next hop is found, this means that the RREP

packet was forwarded from the downstream nodes

and routing loop occurred. In this case, the node

discards the RREP packet immediately.

5 PROPOSED SYSTEM

To avoid the root flop by use the Link

Score value to identify the node weight and

transmission success rate to transfer the data‟s

from source to destination. The mechanism that

compare AOMDV routing schemes with SMR

routing schemes along with the performance

metrics viz: throughput, end-to-end delay,

delivery ratio, and distance calculation. A metric,

LinkScore,is defined as LinkScore = (PE ラ WE

+PT ラ WT ), where PE – energy level of the next

hop node (0.0 to 100.0), WE– assigned weight for

PE (0.0 to 1.0), PT – transmission success rate

(0.0 to 100.0) and WT – assigned weight for PT

(0.0 to 1.0). Weights, WE and WT, may be

determined empirically but their sum must equal

1. For example, in a low noise environment, the

probability of successful transmission is higher. In

this scenario, WE = 0.7 and WT = 0.3may be

chosen allowing routing decisions to focus more

on energy conservation in path selection.

Conversely, in a noisy environment, WT = 0.7 and

WE = 0.3 could be chosen instead, giving higher

emphasis to the selection of high reliability paths

over energy conservation. LinkScore then takes on

a value from 0 to 100 and a higher value indicates

a better link. An arbitrary value is initially

assigned to PT as the link performance is

unknown. PT rises (or drops) when subsequent

packet transmissions succeed (or fail). PE starts at

100 and drops as a node consumes its energy

resources.

386

LinkScore is used when there are two

links of different routes with the same hub

distance competing to be admitted to the routing

table. When a new link is received and the routing

table is full, link replacement is initiated. The

search ignores blacklisted links and targets the

link with the lowest LinkScore to be replaced.

When there is more than one entry with the same

LinkScore, the entry with the longest length is

chosen to be replaced. This worst route is then

compared against the incoming route and the

shorter route is admitted into the routing table. If

there is a tie in route length, then the route with

the higher LinkScore is admitted. Since each node

stores and maintains the best available RF links in

its routing table, packets travel on the best route

from a node to a hub at the given time. Since the

routing table contains only one entry to the next

hop node, its size scales slowly with network size

and multiple hubs.

Figure 6 Illustration of forwarding based

on link score metric

CONCLUSION

The SMS model is also implemented within

the NS-2 simulator environment and SMS

performance is compared with SMR, which uses

node-disjoint secondary paths for correcting route

breaks. The SMS model is compared with

AOMDV and route flops are reduced by using the

Link score values. With an increase in mobility,

offered load and network size.

REFERENCES

1. BROCH J., MALTZ D.A., JOHNSON

D.B., HU Y., JETCHEVA J.: „A

performance comparison of multi-hop

wireless ad-hoc network routing

protocols‟. Proc. 4th Annual ACM/IEEE

Int. Conf. on Mobile Computing and

Networking, Dallas, October 1998, pp.

85–97

2. JOHNSON D.B., HU Y., MALTZ D.A.:

„Dynamic source routing in ad-hoc

wireless networks‟, IETF RFC 4728,

February2007,http://www.ietf.org/rfc/rfc4

728.txt

3. LEE S., GERLA M.: „Split multipath

routing with maximally disjoint paths in

ad-hoc networks‟. Proc. IEEE Int. Conf.

On Commun. (ICC), Helsinki, June 2001,

pp. 3201–3205

4. LI X., CUTHBERT L.: „A reliable node-

disjoint multipath routing with low

overhead in wireless ad-hoc

networks‟.Proc. 7th ACM/IEEE Int.

Symp. on Modeling, Analysis and

Simulation of Wireless and Mobile

Systems, Venezia, 2004,pp. 230–233

5. MARINA M., DAS S.: „On-demand

multipath distance vector routing in ad-

hoc networks‟. Proc. 9th IEEE Int.Conf.

on Network Protocols, California, 2001,

pp. 14–23

6. NASIPURI A., CASTANEDA R., DAS

S.: „Performance of multipath routing for

on-demand protocols in mobile adhoc

networks‟, ACM/Kluwer Mobile Netw.

Appl. J., 2001, 6, (4), pp. 339–349

7. Performance evaluation of shortest

multipath source routing scheme H.

Zafar1, 2 D. Harle1 I. Andonovic1 Y.

Khawaja2

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

387

BRODCAST SCHEDULING IN WIRELESS NETWORK

Ms.A.Gurushakthi Meera, PG Student, Mrs.G.Prema, M.S, Professor

Electronics and Communication Engineering, Mepco Schlenk Engineering College, sivakasi,virudhunagar dist

[email protected], [email protected]

Abstract – The paper broadcast scheduling in

wireless networks is designed to avoid the

problem of broadcast storm. The broadcast

storm problem tells us that naive flooding is not

practical because it causes severe contention,

collision and congestion. In this paper the

broadcast storm problem is investigated using

two new models. They are 2-disk model and

SINR model.In 2 Disk model, 2 disks are

employed to represent the transmission and the

interference range. The signal to interference

model plus noise ratio model is more realistic

where many far-away nodes could still have the

effect of interfering some nodes if they are

transmitting simultaneously. The 2-disk model

is extended to SINR model because in SINR

model the accumulative interference range of

many nodes outside the interference range will

not be neglected.

Keywords— Broadcast, Scheduling, SINR,

2-Disk, TDMA.

I.INTRODUCTION

Broadcast is a task initiated by a single processor,

called the source, transmitting a single message.

The goal is to have the message reach all processors

in the network. Broadcast is the task initiated by

any of the nodes of a network, called source, whose

goal is to send a message to all other network

nodes. In addition to disseminating data, a

broadcast mechanism represents the starting point

for the implementation of group communication

primitives and various system services in

distributed environments. With the renewed interest

in ad hoc networks, combined with the need to

support multimedia and real-time applications, the

need arises to have in the mobile network a

mechanism for reliable dissemination of control and

data. It is necessary to have broadcast protocols that

meet the combined requirements of these

networks—delivery-guarantee and mobility

transparency. The broadcast problem has been

extensively studied for multi-hop networks. In

particular, several solutions have been presented in

which the broadcast time complexity is

investigated. In detail optimal solutions were

obtained for the case when each node knows the

topology of the entire network.

Broadcast is an important operation in wireless

networks. However, broadcasting by naive flooding

cause severe contention, collision, and congestion,

which is named as the broadcast storm problem.

Many protocols have been proposed to solve this

problem, with some investigations focusing on

collision avoidance yet neglecting the reduction of

redundant rebroadcasts and broadcasting latency;

while other studies have focused on reducing

redundant rebroadcasts yet have paid little attention

to collision avoidance. Due to advances in wireless

communication technology and portable devices,

wireless communication systems have recently

become increasingly widespread. A wireless

388

network is a collection of hosts that communicate

with each other via radio channels. The hosts can be

static such as base stations in packet radio

networks, or mobile, such as notebook computers in

mobile ad hoc networks (MANETs). If all hosts in a

wireless network can communicate directly, the

network is single hop, otherwise it is multi-hop.

Broadcast is an important operation in all kinds of

networks. However, due to the limited transmission

range and bandwidth of a radio channel, the

broadcast protocol must be designed carefully to

avoid serious contention, collision, and congestion,

known as the broadcast storm problem.

References [1, 2, 3, 4, 5, 6, 7, and 8] attempt to

design collision-free neighbouring broadcast

protocols, and model the broadcast scheduling

problem as a graph-colouring problem. The colours

in the graphs represent the channels assigned to the

hosts(which could be slots, frequencies or codes).

Since no host can be assigned the same colour

(channel) as any of its neighbours within two-hop,

collisions can be avoided. The research goal of the

protocols presented in [6, 7] is to minimize the

assigned colours (channels), while the protocols

Presented in [2, 8] aim to increase total through-put.

Using a different approach, two collision-free

protocols for one-to-all broadcast are proposed in

[9], one centralized and the other distributed. In the

centralized scheme, the source host schedules the

transmission sequence using knowledge of global

network topology. Unlike the graph-colouring

problem, a host can simultaneously use the same

channel as its neighbours within two hops, provided

no collisions occur among the receiving hosts.

However, in the distributed scheme, the source host

follows the depth first search tree to pass a token to

every host in the network. After receiving the

token, the host realizes which time slots are

collision-free and can then decide its transmission

sequence based on this information. The above

broadcast protocols are aimed to alleviate the

broadcast storm problem: some works [1, 9, 2, 3, 4,

5, 6, 7, 8] focus on avoiding collisions but pay little

attention to reducing redundant rebroadcasts and

broadcasting latency. Other studies [10and 11] try

to reduce redundant rebroadcasts, but cannot

guarantee a collision-free broadcast.

Our work uses the SINR model, so it is also related

to those who used this model. Moschibroda et al.

[25].They considered the problem of scheduling a

given topology by using the SINR model. In a

network, for any given topology, we may not be

able to realize this topology in one time slot if

interference is considered. In other words, we need

to do scheduling in order to make a topology

feasible, and Moscibroda et al. [12] focused on the

latency issue. This problem is not directly related to

our work, as scheduling a topology is always a one-

hop concept, in which there is no relay. In

broadcast, a non source node cannot transmit a

message, unless it has already received from

another node. This property makes our work

fundamentally different from [12].

II.INTERFERENCE MODELS

In this section, we introduce two interference

models, namely, the 2-Disk and SINR models.

A.2-Disk Model

A wireless network is modelled as a set of nodes

V arbitrarily located in a 2D Euclidean space. Each

node is associated with two radii: the transmission

radius rT and the interference radius rI (where rI

greater than or equal to rT). The transmission

range of a node v is a disk of radius r T centred at

v, and the interference range of v is a disk of radius

rI centred at v. However, the transmission range is

a concept with respect to the transmitting nodes,

while the interference range is a concept with

respect to the receiving nodes. A node u receives a

message successfully from v if and only if u is

within the ―v‖ transmission range and no other

nodes are within ―u‖ interference range. For

simplicity, we assume that all nodes have the same

rT and rI in the 2-Disk model throughout this

paper.

B.SINR MODEL

389

In the SINR model, a wireless network is also

regarded as a set V in a 2D Euclidean space. Each

node is associated with a transmission power P. For

simplicity, we assume that all nodes have the same

P. A node v receives a message successfully in a

time slot from another node u if and only if the

SINR at v is at least a given constant Beta, where

Beta is called the minimum SINR.

III.BROADCAST SCHEDULING

The broadcast scheduling consists of three steps.

A.MIS CONSTRUCTION

We consider the transmission graph G =(V ,ET )

generated by rT and V . To define the broadcast

schedule, we first need to construct a virtual

backbone as follows: We look at G and its Breadth

First Search (BFS) tree and then divide V into

layers L0; L1; L2; . . . ; LR. All nodes of layer i are

thus i hop away from the root. Then, we construct a

layered maximal independent2 set, called BLACK,

as follows: Starting from the zero-th layer, which

contains only s, we pick up a maximal independent

set (MIS), which contains only s as well. Then, at

the first layer, we pick up an MIS in which each

node is independent of each other and those nodes

at the zero-th layer. Note that this is empty, because

all nodes in L1 (layer 1) must be adjacent to s.

Then, we move on to the second layer, pick up an

MIS, and mark these nodes black again. Note that

the black nodes of the second layer also need to be

independent of those of the first layer. We repeat

this process until all layers have been worked on.

Nodes that are not marked black are marked white

at last. Those black nodes are also called the

dominators, and we will use these two terms

interchangeably throughout this paper. The pseudo

code of layered MIS construction is given as

follows.

1 .Construct an MIS in G layer by layer.

2. Input the value of set of nodes, Source node,

Graph

3. Set the value of BLACK is equal to NULL.

4. Find an MIS such that the value of value of

BLACKi is the subset of Li.

5. Set BLACK as Combination of independent Set

and Maximum Independent set.

B.VIRTUAL BACKBONE CONSTRUCTION

Now, we construct the virtual backbone as follows:

We pick some of the white nodes and colour them

blue to interconnect all black nodes. Note that

L0=s and all nodes in L1 must be white. We

simply connect s to all nodes in L1. To connect L1

and L2, we look at L2’s black nodes. Each black

node must have a parent on L1, and this parent

node must be white, since black nodes are

independent of each other. We colour this white

node blue and add an edge between them.

Moreover, we know that this blue node must be

dominated by a black node either on L1 or L0 (in

this case, L0). We then add an edge between this

blue node and its dominator. We repeat this process

layer by layer and finally obtain the desired virtual

backbone (which is a tree) in this manner. Note that

in this tree, each black node has a blue parent at the

upper layer and each blue node has a black parent at

the same layer or the layer right next to it above.

C.SCHEDULING THE BROADCAST

The broadcast scheduling algorithm based on the

virtual backbone in the 2-Disk model is described

as follows: Starting from the zero layer containing

Only the source s, the s is scheduled to transmit in

the first time slot, and obviously, this transmission

causes no collision. After the first time slot, all

nodes of the first layer will receive successfully. A

schedule is designed such that all nodes of the (i+1)

layer receive from the ith layer successfully for i

=1, 2….R. We partition the plane into half-open

half-closed hexagons. Then, the distance between

two hexagons of the same colour is at least r t+ r1,

which guarantees the validity of the proposed

schedule. This schedule has two parts, and in the

first part, each blue node of layer I is scheduled to

transmit in the time slot according to its targeted

black nodes colours. If there is more than one

targeted black node with the same colour, those

blue nodes will need to transmit multiple times.

Transmission

Graph

Tessellation of

layers

390

Figure 1: Block Diagram of Proposed System

IV.RESULT

In Figure2 the transmission latencies remain

almost the same when the number of nodes in

network n is larger than 1,000 for each set of X and

Y. This is actually expected: the increase in n does

not change the transmission tessellation and its

depth significantly (as discussed in Section 7). As

the network size increases, the transmission latency

becomes longer. This is because of the increased

depth of the virtual backbone.

Figure 2: Transmission Latency for different

network sizes

Figure 3: Transmission Latency for different

number of nodes.

V.CONCLUSION

Many highly theoretical models were used in all

previous works on broadcast scheduling. Instead, in

these paper two more practical models for

reinvestigating this problem has been used. It has

been found that the same method can be applied to

Both models and obtain low-latency schedules.

Although the proposed algorithms used in this

paper are centralized, the minimum latency problem

could not be solved. The main reason for that is the

difficulty of representing the condition that ―a node

can only transmit if it has successfully received

from another node.‖ So far, there is still no good

formulation for representing this condition, even in

the general graph model, so we believe that it is

more difficult to represent it in our more

complicated 2-Disk and SINR models.

IV.REFERENCES

[1] S. Basagni, D. Bruschi, and I. Chlamtac. A

mobility transparent deterministic broadcast

mechanism for ad hoc networks. IEEE/ACM

Transactions on Networking, 7:799–807, December

1999.

[2] A. Ephremides and T. V. Truong. Scheduling

broadcasts in multihop radio networks. IEEE

Virtual

backbone

networks

Broadcast

Scheduling

391

Transactions on Communications, 38:456 –460,

April 1990.

[3] M. L. Huson and A. Sen. Broadcast scheduling

algorithms for radio networks. In IEEE Military

Communications Conference, pages 647–651, July

1995.

[4] J. H. Ju and V. O. K. Li. An optimal topology-

transparent scheduling method in multihop packet

radio networks.IEEE/ACM Transactions on

Networking, 6:298 –306,June 1998.

[5] C. K. Lee, J. E. Burns, and M. H. Ammar.

Improved randomized broadcast protocols in multi-

hop radio networks.In Proceedings of International

Conference on Network Protocols, pages 234–241,

October 1993.

[6] S. Ramanathan and E. L. Lloyd. Scheduling

algorithms for multihop radio networks.IEEE/ACM

Transactions on Networking, 1:166 –177, April

1993.

[7] R. Ramaswami and K. K. Parhi. Distributed

scheduling of broadcasts in a radio network. In

INFOCOM, volume 2, pages 497–504, 1989.

[8] A. Sen and J. M. Capone. Scheduling In Packet

Radio Networks - A New Approach. In Proceedings

of Global Telecommunications Conference, pages

650–654,1999.

[9] I. Chlamtac and S. Kutten. Tree-based

broadcasting in multihop radio networks. IEEE

Transaction on Computers, pages 1209–1225,

October 1987

[10] S. Y. Ni, Y. C. Tseng, Y. S. Chen, and J. P.

Sheu. TheBroadcast Storm Problem in a Mobile Ad

Hoc Network.In Proceedings of the fifth annual

ACM/IEEE international conference on Mobile

computing and networking, pages 151–162, August

1999.

[11] W. Peng and X.-C. Lu. On the reduction of

broadcast redundancy in mobile ad hoc networks.

In Proceedings of Mobile and Ad Hoc Networking

and Computing, pages 129–130, 2000.

[12] T. Moscibroda, R. Wattenhofer, and A.

Zollinger, ―Topology Control Meets SINR: The

Scheduling Complexity of Arbitrary Topologies,‖

Proc. ACM MobiHoc ’06, pp. 310-321, 2006.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

392

Abstract— Quality of Service (QoS) is the important parameter

to achieve the better performance of the network. In Previous

work increase the security that decreases the QoS. No solution

has been provided when changing available network resources

due to traffic, mobility etc., which results in security features

producing too much overhead such that it significantly impacts

routing QoS performance. To achieve the objective of this

project proposes the three routing protocols like Destination

Sequenced Distance Vector (DSDV), Dynamic Source Routing

(DSR) and Ad Hoc On-Demand Distance Vector Routing

(AODV) and the performance have been implemented. This

project aims to propose a security and QoS self- optimization for

mobile ad hoc network which can automatically adapt the

network security level and QoS with minimum requirements

while aiming to provide more than the minimum security and

QoS. A simulation model describes DSR perform better than the

DSDV and AODV protocol. A variety of parameters, as

characterized by mobility, and size of the ad hoc network were

simulated. The performance differentials are analyzed using

varying mobility, and network size. These simulations are

carried out based on the ns -2 network simulator to run ad hoc

simulations.

Index Terms— MANET, AODV, DSR, DSDV, NS2.

I. INTRODUCTION

A mobile Ad-hoc network (MANET) is a kind of

wireless ad-hoc network, and is the collection of mobile nodes

where the nodes will self configure and self optimize

themselves of mobile routers (and associated hosts)

connected by wireless links. The routers are free to move

randomly and organize themselves arbitrarily; thus, the

network's wireless topology may change rapidly and

unpredictably. Such a network may operate in a standalone

fashion, or may be connected to the larger Internet. It is also

expected to play an important role in civilian forums such as

campus recreation, conferences, and electronic classrooms,

military, earthquake etc. Due to the mobility of nodes, the

topology of the network may changes. In MANETs, the nodes

must be able to relay traffic since communicating nodes

might be out of range. The inherent feature of

communications quality in a MANET makes it difficult to

offer fixed guarantees on the services offered to a device.

Providing different quality of service levels in changing

environment will be a challenging issue. One of the main

factors in Ad-hoc network is to develop a routing protocol

which must be capable of handling large number of nodes in

the network with secure communication among the nodes and

improving the quality of service in the network. The existing

routing techniques for network services, and poses a number

of challenges ensuring the security of communication. Many

ad hoc routing protocols have been proposed to solve the

security issues to route packets among participating nodes. In

this paper we mainly discuss the performance analysis of

security and quality of service of three wireless multi-hop

routing protocols; reactive protocols like DSR and AODV,

and proactive protocol, DSDV. We focus on the performance

of both proactive and reactive protocols specifically AODV,

DSR (reactive) and DSDV (proactive) protocols in dynamic

environment. By this motive, we intend to compare Ad Hoc

routing protocols and analyze results of these protocols. We

need this work to study the behaviour of each protocol and

how they performing in different scenarios and to find, which

protocol performs better under a particular simulation.

The rest of this paper is organized as follows, System

model is given in Section 2. In Section 3, we present our

routing protocol. Simulation results are presented in Section

4. Section 5 concludes this System Model

II. SYSTEM MODEL

Mobile Ad Hoc Networks are wireless networks which do

not need any infrastructure support for transferring data

between nodes. In these networks, nodes also work as a

router, which means that they also route packets for other

nodes. Thus a node can be a host or router at different times

depending upon the situation i.e. if the node wants to send or

receive data, then it will act as a host and if it has to just

transfer some data packet to other node, then it will act as a

router. Nodes are free to move, independent of each other.

Topology of such networks keeps on changing dynamically

which makes routing much difficult. Therefore routing is one

of

Performance Analysis of Security and QoS

Self-Optimization in Mobile Ad Hoc Networks. R.R.Ramya, [email protected] T. Sree Sharmila

Dept of Information technology

SSN COLLEGE OF ENGINEERING,

KALAVAKKAM, CHENNAI.

393

the most concerned areas in these networks. Normal routing

protocols which work well in fixed networks do not show the

same performance in Mobile Ad Hoc Networks because the

requirements differ in the two scenarios. In wireless

networks, routing protocols should be more dynamic so that

they quickly respond to topological changes which occur

frequently in these networks.

The dynamic nature of MANETs is attributed to several

inherent characteristics, such as variable link behaviour, node

movements, changing network topology and variable

application demands. Providing QoS in such a dynamic

environment is very difficult. The routing protocols fall into

two categories:

1. Reactive

2. Pro-active

In Re-active routing protocol does not use bandwidth

except when needed. It establishes routes “on demand” by

flooding a query. Much network overhead is present in the

flooding process when querying for routes. In Pro-active routing protocols, consistent and up-t o-date routing

information to all nodes is maintained at each n ode.

III. AD-HOC ROUTING PROTOCOLS DESCRIPTION

Many QoS routing algorithms represent an extension of

existing classic best-effort routing algorithms. Many routing

protocols have been developed which support establishing and

maintaining multi-hop routes between nodes in MANETs.

These algorithms can be classified into two different

categories: on-demand (reactive) such as DSR and AODV

and table-driven (proactive) such as Destination Sequenced

Distance Vector protocol (DSDV).

Table-driven routing protocols (DSDV) attempt to

maintain up-to-date routing information from each node to

every other node in the network. Every node in this network

maintains the route table to store route information. On the

other hand, on-demand routing protocols create route only

when needs araise. When a source needs a route to a

destination, it starts route discovery and maintenance.

For our performance analysis study, we pick up two on

demand protocols (AODV, DSR) and one table-driven

protocol, DSDV.

A. DSR- Dynamic Source Routing Protocol

DSR is one of the most well known routing algorithms for

ad hoc wireless networks. DSR uses source routing, which

allows packet routing to be loop free. It increases its

efficiency by allowing nodes that are either forwarding route

discovery requests or overhearing packets through

promiscuous listening mode to cache the routing information

for future use. DSR is also on demand, which reduces the

bandwidth use especially in situations where the mobility is

low. It is a simple and efficient routing protocol for use in ad

hoc networks. It has two important phases, route discovery

and route maintenance

The main algorithm works in the following manner. A node

that desires communication with another node first searches

its route cache to see if it already has a route to the

destination. If it does not, it then initiates a route discovery

mechanism. This is done by sending a Route Request

message. When the node gets this route request message, it

searches its own cache to see if it has a route to the

destination. If it does not, it then appends its id to the packet

and forwards the packet to the next

node; this continues until either a node with a route to the

destination is encountered (i.e. has a route in its own cache)

or the destination receives the packet. In that case, the node

sends a route reply packet which has a list of all of the nodes

that forwarded the packet to reach the destination. This

constitutes the routing information needed by the source,

which can then send its data packets to the destination using

this newly discovered route. Although DSR can support

relatively rapid rates of mobility, it is assumed that the

mobility is not so high as to make flooding the only possible

way to exchange packets between nodes.

B. AODV - The Ad Hoc On-demand Distance-Vector

Protocol

AODV is another routing algorithm used in ad hoc

networks. Unlike DSR, it does not use source routing, but like

DSR it is on-demand. In AODV, to initiate a route discovery

process a node creates a route request (RREQ) packet. The

packet contains the source node’s IP address as well as the

destination’s IP address. The RREQ contains a broadcast ID,

which is incremented each time the source node initiates a

RREQ. The broadcast ID and the IP address of the source

node form a unique identifier for the RREQ. The source node

then broadcasts the packet and waits for a reply. When an

intermediate node receives a RREQ, it checks to see if it has

seen it before using the source and broadcast ID’s of the

packet. If it has seen the packet previously, it discards it.

Otherwise it processes the RREQ packet.

To process the packet the node sets up a reverse route entry

for the source node in its route table which contains the ID of

the neighbour through which it received the RREQ packet. In

this way, the node knows how to forward a route reply packet

(RREP) to the source if it receives one later. When a node

receives the RREQ, it determines if indeed it is the indicated

destination and, if not, if it has a route to respond to the

RREQ. If either of those conditions is true, then it unicasts a

route reply (RREP) message back to the source. If both

conditions are false, i.e. if it does not have a route and it is

not the indicated destination, it then broadcasts the packet to

its neighbours. Ultimately, the destination node will always

be able to respond to the RREQ message. When an

intermediate node receives the RREP, it sets up a forward

path entry to the destination in its routing table. This entry

contains the IP address of the destination, the IP address of

the neighbour from which the RREP arrived, and the hop

count or distance to the destination. After processing the

RREP packet, the node forwards it toward the source. The

node can later update its routing information if it discovers a

better route. This could be used for QoS routing support to

choose between routes based on different criteria such as

reliability and delay. To provide

394

such support additional QoS attributes would need to be

created, maintained, and stored for each route in the routing

table to allow the selection of the appropriate route among

multiple routes to the destination.

C. DSDV - The Destination Sequenced Distance Vector

Protocol

DSDV is one of the most well known table-driven routing

algorithms for MANETs. It is a distance vector protocol. In

distance vector protocols, every node maintains for

each destination a set of distances for each node j that is a

neighbor of node i treats neighbor k as a next hop for a

packet. The succession of next hops chosen in this manner

leads along the shortest path. In order to keep the distance

estimates up to date, each node monitors the cost of its

outgoing links and periodically broadcasts to all of its

neighbors its current estimate of the shortest distance to every

other node in the network. The distance vector which is

periodically broadcasted contains one entry for each node in

the network. DSDV is a distance vector algorithm which uses

sequence numbers originated and updated by the destination,

to avoid the looping problem caused by stale routing

information. In DSDV, each node maintains a routing table

which is constantly and periodically updated (not on-demand)

and advertised to each of the node’s current neighbors. Each

entry in the routing table has the last known destination

sequence number. Each node periodically transmits updates,

and it does so immediately when significant new information

is available. The data broadcasted by each node will contain

its new sequence number and the following information for

each new route: the destination’s address, the number of hops

to reach the destination and the sequence number of the

information received regarding that destination, as originally

stamped by the destination. No assumptions about mobile

hosts maintaining any sort of time synchronization or about

the phase relationship of the update periods between the

mobile nodes are made. Following the traditional distance-

vector routing algorithms, these update packets contain

information about which nodes are accessible from each node

and the number of hops necessary to reach them. Routes with

more recent sequence numbers are always the preferred basis

for forwarding decisions. Of the paths with the same sequence

number, those with the smallest metric (number of hops to the

destination) will be used. The addresses stored in the route

tables will correspond to the layer at which the DSDV

protocol is operated.

IV. SIMULATION RESULTS

Simulations have been conducted with varying the node

density and Source-Destination pairs. Following metrics

are used for evaluation.

1. Routing overhead: The total number of routing

packets transmitted by sending or receiving node

which involved in the session during the simulation.

2. Throughput: It is the amount of data moved

successfully from one place to another in a given

time period.

3. Packet Drop ratio: It describes how many packets

were lost in transmit between the source (or input)

and destination (or output).

4. Packet Received ratio: It describes that no. Of

packets received and no. Of packets sent.

Routing Overhead is an important metric for comparing

these protocols. The degree to which congested and low

bandwidth environments which should be secure transmits

the packets and maintains the quality of service. The

protocols which send large number of routing packets can

increase with minimum delay time and minimize the

dropping of packets.

Fig.1 shows overhead of routing protocols, where it

contains AODV, DSR, DSDV plots. And it desides that DSR

is having less overhead than others. Fig 2 shows the

comparision of Throughput analysis, whereas DSR

Throughput is increased to that of AODV.Fig 3 Received

ratio of AODV,DSR.DSDV, and it clearly shows that DSR

have high receiving packets. Fig 4 Drop ratio of

AODV,DSR,DSDV where DSR shows less drop ratio thn

others.

Fig 1:Comparision of Routing Protocols overheads

395

Fig 2 shows Throughput of AODV and DSR

Fig 3 shows Received Ratio of DSR, AODV and DSDV

Fig 4 shows the Drop ratio of DSR, AODV, DSDV

V. CONCLUSION

This work presents a detailed comparative analysis of

Routing protocols i.e., AODV, DSR, and DSDV for ad hoc

networks using ns- 2 simulations.. This concludes that the

security and QoS is increased using the DSR routing

protocols performance compared with the other two routing

protocols. The overhead and the drop ratio of the DSR is

better performance compared with the other two routing

protocols.

REFERENCES

[1] Chenxi Zhu and M. Scott Corson. “QoS Routing for

Mobile Ad Hoc Networks”. In the Proc. IEEE Infocom, June

2001.

[2] Demetris Zeinalipour. “A Glance at QoS in MANETs”.

University of California, Tech. Rep., 2001.

[3] David B. Johnson and David A. Maltz. “Dynamic Source

Routing in Ad Hoc Wireless Networks”. In Ad Hoc Wireless

Networks, Mobile Computing, T. Imielinski and H. Korth

(Eds.), Chapter 5, pp 153-181, Kluwer Academic Publishers,

1996.

[4] Andreas Tønnesen. “Mobile Ad-Hoc Networks”

[5] Ahmed Al-Maashri and Mohamed Ould-Khaoua.

“Performance Analysis of MANET Routing Protocols in the

Presence of Self-Similar Traffic”. IEEE, ISSN- 0742-1303,

First published in Proc. of the 31st IEEE Conference on

Local Computer Networks, 2006.

[6] Imad Jawhar and Jie Wu. “Quality of Service Routing in

Mobile Ad Hoc Networks”

[7] Jaroslaw Malek. “Trace graph - Network Simulator NS-2

trace files analyzer”

http://www.tracegraph.com

[8] Tutorial for the network simulator “ns”.

http://www.isi.edu/nsnam/ns/tutorial/

Proceedings of the Third National Conference on RTICT 2010 Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

396

Security in Mobile Adhoc Network

Amit Kumar Jaiswal Kamal Kant Pardeep Singh

[email protected] [email protected] [email protected]

Department of Computer Science and Engineering

National Institute of Technology, Hamirpur (H.P)

Abstract: In this paper, we discuss security issues and

their current solutions in the mobile ad hoc network. Due

to the vulnerable nature of the mobile ad hoc network,

there are numerous security threats that disturb the

development of it. We first analyze the main

vulnerabilities in the mobile ad hoc networks, which have

made it much easier to suffer from attacks than the

traditional wired network. Then we discuss the security

criteria of the mobile ad hoc network and present the

main attack types that exist in it. Finally we survey the

current security solutions for the mobile ad hoc network.

I. INTRODUCTION

The Mobile Ad Hoc Network is one of the wireless

networks that have attracted most concentrations

from many researchers. A Mobile Ad hoc NETwork (MANET) is a system of wireless mobile nodes that

dynamically self-organize in arbitrary and temporary

network topologies. People and vehicles can thus be

internetworked in areas without a preexisting

communication infrastructure or when the use of such

infrastructure requires wireless extension [1]. In the

mobile ad hoc network, nodes can directly

communicate with all the other nodes within their

radio ranges; whereas nodes that not in the direct

communication range use intermediate node(s) to

communicate with each other. In these two situations,

all the nodes that have participated in the communication automatically form a wireless

network, therefore this kind of wireless network can

be viewed as mobile ad hoc network. The mobile ad

hoc network has the following typical features [2]:

Unreliability of wireless links between nodes.

Because of the limited energy supply for the wireless

nodes and the mobility of the nodes, the wireless

links between mobile nodes in the ad hoc network are

not consistent for the communication participants.

The rest of the paper is organized as follows: In

Section 2, we discuss the main challenges that make the mobile ad hoc networks unsecure. In Section 3,

we discuss security goals for the mobile ad hoc

networks, in Section 4 and 5 we will discuss the few

existing solution for security in Adhoc network and

at last we have given some open challenges for

mobile adhoc network.

II. CHALLENGES

Lack of secure boundaries makes the mobile ad hoc

network susceptible to the attacks. The mobile ad hoc

network suffers from all-weather attacks, which can

come from any node that is in the radio range of any

node in the network, at any time, and target to any

other node(s) in the network. The attacks mainly

include passive eavesdropping, active interfering,

leakage of secret information, data tampering,

message replay, message contamination, and denial

of service Secondly, in many situations the nodes may be left unattended in a hostile environment. This

enables adversaries to capture them and physically

attack them. Proper precautions (tamper resistance)

are required to prevent attackers from extracting

secret information from them. Even with these

precautions, we cannot exclude that a fraction of the

nodes may become compromised. This enables

attacks launched from within the network. Thirdly, an

ad hoc network is dynamic because of frequent

changes in both its topology and its membership (i.e.,

nodes frequently join and leave the network). Trust relationship among nodes also changes, Fourthly,

many wireless nodes will have a limited energy

resource (battery, solar panel, etc.). This is

particularly true in the case of ad hoc sensor

networks. Security solutions should be designed with

this limited energy budget in mind. lack of

centralized management machinery will impede the

trust management for the nodes in the ad hoc network

[2]. In mobile ad hoc network, all the nodes are

required to cooperate in the network operation, while

no security association (SA2) can be assumed for all

the network nodes. Thus, it is not practical to perform an a priori classification, and as a result, the usual

practice of establishing a line of defense, which

distinguishes nodes as trusted and nontrusted, cannot

be achieved here in the mobile ad hoc network.

III. SECURITY GOALS

To secure an adhoc network, we consider the

following attributes: availability, confidentiality,

integrity, authentication, and non-repudiation.

397

Availability ensures the survivability of network

services despite denial of service attacks. A denial of

service attack could be launched at any layer of an ad

hoc network.

Confidentiality ensures that certain information is

never disclosed to unauthorized entities. Network transmission of sensitive information, such as

strategic or tactical military information, requires

confidentiality. Leakage of such information to

enemies could have devastating consequences.

Integrity guarantees that a message being transferred

is never corrupted. A message could be corrupted

because of benign failures, such as radio propagation

impairment, or because of malicious attacks on the

network.

Authentication enables a node to ensure the identity

of the peer node it is communicating with. Without

authentication, an adversary could masquerade a node, thus gaining unauthorized access to resource

and sensitive information and interfering with the

operation of other nodes.

Non-repudiation ensures that the origin of a message

cannot deny having sent the message. Non-

repudiation is useful for detection and isolation of

compromised nodes. When a node A receives an

erroneous message from a node B, non-repudiation

allows A to accuse B using this message and to

convince other nodes that B is compromised.

IV. SECURE ADHOC ROUTING PROTOCOLS

The secure adhoc routing protocols enhance the

existing ad hoc routing protocols, such as DSR and

AODV with security extensions. In these protocols,

one node signs its routing messages using some

cryptographic authentication method like digital

signature so that each node can authenticate the legal

traffic efficiently and distinguish the unauthenticated message packets from attackers and correct packets.

However, there are still chances that an authenticated

node has been compromised and controlled by the

malicious attacker. Therefore, we have to further

ensure that a node obeys the routing protocols

properly even it is an authenticated node. In the

following, we describe how different types of routing

protocols are secured.

A. Source Routing

For source routing protocols such as DSR, the main

challenge is to ensure that none of intermediate

nodes can remove already existing nodes from or add

extra nodes into the route. The basic idea is to attach

a per-hop authenticator for the source routing

forwarder list so that any altering of the list can be

detected right away. A secure extension of DSR is

Ariadne, which is described in [3].

B. Distance Vector Routing

For distance vector routing protocols such as

AODV, the main challenge is that the information

about the routing metric has to be advertised

correctly by each intermediate node. For example, if we use hop count as the routing metric, each node has

to increase the hop count only by one exactly. A hop

count hash chain is used so that none of the

intermediate nodes can decrease the hop count in a

routing update to achieve benefits, which is described

in [4].

C. Secure Packet Forwarding

The routing message exchange is only one part of the

network-layer protocol which needs to be protected.

It is still possible that malicious nodes deny

forwarding packages correctly even they have acted

correctly during the routing discovery phase. The basic idea to solve this issue is to ensure that each

node indeed forwards packages according to the

protocol. Reactive methods should be used instead of

proactive methods since attacks on package

forwarding cannot be prevented.

V. MULTICAST SECURITY IN MANET

Integrity plays an important role in ad-hoc

networks. To overcome man-in-the-middle attack in

mobile-ad-hoc networks, SHA-1 algorithm is used. Normally, hop count field is mutable in nature. To

protect this hop count value, hash values are found by

using SHA-1 algorithm for those fields. Here, the

packets are sent along with the hashed values of hop

count field. Now, the malicious nodes, which

forwards the false routing information, can be

effectively defended [5].This algorithm takes input as

source address, destination address and hop count

with a maximum length of less than 264 bits and

produces output as a 160-bits message digest. The

input is processed in 512-bits blocks. This algorithm

includes the following steps [6]. Padding —The purpose of message padding is to

make the total length of a padded message congruent

to 448 modulo 512 (length = 448 mod 512). The

number of padding bits should he between 1 and 512.

Padding consists of a single 1-bit followed by the

necessary number of 0-bits.

Appending Length— The 64-bit binary representation

of the original length of the message is appended to

the end of the message.

Initialize the SHA-1 buffer —The 160-bit buffer is

represented by five four-word buffers (A, B, C, D, E) used to store the middle or finally results of the

message digests for SHA-I functions and they are

initialized to the following values in hexadecimal.

Low-order bytes are put first:

398

Word A: 67 45 23 01

Word B: EF CD AB 89

Word C: 98 BA DC EF

Word D: 10 32 54 16

Word E: C3 D2 El FO

Process message in 16-word blocks The heart of the algorithm is a module that consists of four rounds

of processing 20 steps each. The four rounds have a

similar structure, but each uses a different primitive

logical function. These logical functions are defined

as follows:

Initialize hash value :

a := A

b := B

c := C

d := D

e := E

Main loop: for i from 0 to 79

if 0 ≤ i ≤ 19 then

f := (b and c) or ((not b) and d)

k := 0x5A827999

else if 20 ≤ i ≤ 39

f := b xor c xor d

k := 0x6ED9EBA1

else if 40 ≤ i ≤ 59

f := (b and c) or (b and d) or (c and d)

k := 0x8F1BBCDC

else if 60 ≤ i ≤ 79 f := b xor c xor d

k := 0xCA62C1D6

The output of the fourth round is added to the input

of the first round, and then the addition is modulo 232

produce the ABCDE value that calculate next 5l2-

bits block.

Output After all 512-bits blocks have been processed,

the output of the last block is the 160-bits message

digest. These message digest values are sent along

with the packets .So, the packets which are sent by

malicious nodes are suppressed. Thus, the integrity is

ensured.

VI. OPEN CHALLENGES

The research on MANET security is still in its early

stage. The existing proposals are typically attack-

oriented in that they first identify several security

threats and then enhance the existing protocol or

propose a new protocol to thwart such threats.

Because the solutions are designed explicitly with

certain attack models in mind, they work well in the

presence of designated attacks but may collapse under unanticipated attacks. Therefore, a more

ambitious goal for d hoc network security is to

develop a multifence security solution that is

embedded into possibly every component in the

network, resulting in in-depth protection that offers

multiple lines of defense against many both known

and unknown security threats. This new design

perspective is what we call resiliency-oriented

security design. We envision the resiliency-oriented

security solution as possessing several features. First, the solution seeks to attack a bigger problem space. It

attempts not only to thwart malicious attacks, but also

to cope with other network faults due to node

misconfiguration, extreme network overload, or

operational failures. In some sense, all such faults,

whether incurred by attacks or misconfigurations,

share some common symptoms from both the

network and end-user perspectives, and should be

handled by the system. Second, resiliency-oriented

design takes a paradigm shift from conventional

intrusion precertain degrees of intrusions or

compromised/captured nodes are the reality to face, not the problem to get rid of, in MANET security.

The overall system has to be robust against the

breakdown of any individual fence, and its

performance does not critically depend on a single

fence. Even though attackers intrude through an

individual fence, the system still functions, but

possibly with graceful performance degradation.

Third, as far as the solution space is concerned,

cryptography-based techniques just offer a subset of

toolkits in a resiliency-oriented design. The solution

also uses other noncrypto-based schemes to ensure resiliency. For example, it may piggyback more

“protocol invariant” information in the protocol

messages, so that all nodes participating in the

message exchanges can verify such information. The

system may also exploit the rich connectivity of the

network topology to detect inconsistency of the

protocol operations. In many cases, routing messages

are typically propagated through multiple paths and

redundant copies of such messages can be used by

downstream nodes. Fourth, the solution should be

able to handle unexpected faults to some extent. One

possible approach worth exploring is to strengthen the correct operation mode of the network by

enhancing more redundancy at the protocol and

system levels. At each step of the protocol operation,

the design makes sure what it has done is completely

along the right track. Anything deviating from valid

operations is treated with caution. Whenever an

inconsistent operation is detected, the system can

raise a suspicion flag and query the identified source

for further verification. This way, the protocol tells

right from wrong because it knows right with higher

confidence, not necessarily knowing what is exactly wrong. The design strengthens the correct operations

and may handle even unanticipated threats in runtime

operations. Next, the solution may also take a

collaborative security approach, which relies on

399

multiple nodes in a MANET to provide any security

primitives. Therefore, no single node is fully trusted.

Instead, only a group of nodes will be trusted

collectively. The group of nodes can be nodes in a

local network neighborhood or all nodes along the

forwarding path. Finally, the solution relies on multiple fences, spanning different devices, different

layers in the protocol stack, and different solution

techniques, to guard the entire system. Each fence

has all functional elements of prevention,

detection/verification, and reaction. The above

mentioned resiliency-oriented MANET security

olution poses grand yet exciting research challenges.

How to build an efficient fence that accommodates

each device’s resource constraint poses an interesting

challenge. Device heterogeneity is one important

concern that has been largely neglected in the urgent

security design process. However, multifence security protection is deployed throughout the

network, and each individual fence adopted by a

single node may have different security strength due

to its resource constraints. A node has to properly

select security mechanisms that fit well into its own

available resources, deployment cost, and other

complexity concerns. The security solution should

not stipulate the minimum requirement a component

must have. Instead, it expects best effort from each

component. The more powerful a component is, the

higher degree of security it has. Next, identifying the system principles of how to build such a new-

generation of network protocols remains unexplored.

The state-of-the-art network protocols are all

designed for functionality only. The protocol

specification fundamentally assumes a fully rusted

and well-behaved network setting for all message

exchanges and protocol operations. It does not

anticipate any faulty signals or ill behaved nodes. We

need to identify new principles to build the next-

generation network protocols that are resilient to

faults. There only exist a few piecemeal individual

efforts. Finally, evaluating the multifence security design also offers new research opportunities. The

effectiveness of each fence and the minimal number

of fences the system has to possess to ensure some

degree of security assurances should be evaluated

through a combination of analysis, simulations, and

measurements in principle. However, it is recognized

that the current evaluation for state-of-the-art

wireless security solutions is quite ad hoc. The

community still lacks effective analytical tools,

particularly in a large scale wireless network setting.

The multidimensional trade-offs among security strength, communication overhead, computation

complexity, energy consumption, and scalability still

remain largely unexplored. Developing effective

evaluation methodology and toolkits will probably

need interdisciplinary efforts from research

communities working in wireless networking, mobile

systems, and cryptography.

VII. CONCLUSION

In this paper we have discussed the several security

issues in Mobile Adhoc Network. We have also given

some existing solution to these security problem but

the solutions available are not enough to cover all the

security threats. At last we have also given the open

challenges that exists in the field of Mobile Adhoc

Network and need to be solved.

REFERENCES

[1] M.S. Corson, J.P. Maker, and J.H. Cernicione,

Internet-based Mobile Ad Hoc Networking, IEEE Internet Computing, pages 63–70, July-August 1999.

[2] Amitabh Mishra and Ketan M. Nadkarni, Security

in Wireless Ad Hoc Networks, in Book The

Handbook of Ad Hoc Wireless Networks (Chapter

30), CRC Press LLC, 2003.

[3] Y. Hu, A. Perrig, and D. Johnson, Ariadne: A

Secure On-demand Routing Protocol for AdHoc

Networks, ACM MOBICOM, 2002.

[4] M. Zapata, and N. Asokan: Securing Ad Hoc

Routing Protocols, ACM WiSe, 2002.

[5] Junaid Arshad and Mohammad Ajmal Azad,“ Performance Evaluation of Secure on Demand

Routing Protocols for Mobile Adhoc Networks”,

IEEE Network, 2006, pp 971-975.

[6] Dai Zibin and Zhou Ning, “FPGA

Implementation of SHA-1 Algorithm”, IEEE 2003,

pp 1321-1324 .

[7] Yongguang Zhang and Wenke Lee, Security in

Mobile Ad-Hoc Networks, in Book AdHoc Networks

Technologies and Protocols (Chapter 9), Springer,

2005.

[8] Panagiotis Papadimitraos and Zygmunt J. Hass,

Securing Mobile Ad Hoc Networks, in Book The Handbook of Ad Hoc Wireless Networks (Chapter

31), CRC Press LLC, 2003.

[9] Yi-an Huang and Wenke Lee, A Cooperative

Intrusion Detection System for Ad Hoc Networks, in

Proceedings of the 1st ACM Workshop on Security of

Ad hoc and Sensor Networks, Fairfax, Virginia, 2003,

pp. 135 – 147.

[10] Data Integrity, from Wikipedia, the free

encyclopedia,http://en.wikipedia.org/wiki/Data_integ

rity.

[11] P. Papadimitratos and Z. J. Hass, Secure Routing for Mobile Ad Hoc Networks, in Proceedings of SCS

Communication Networks and Distributed Systems

Modeling and Simulation Conference (CNDS), San

Antonio, TX, January 2002

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

400

Abstract—Sensor networks consist of large number of tiny

nodes. Each node is operated by battery of limited energy.

Hence the major design challenge in sensor network is to

increase their operational lifetime as much as possible. The

existing topology management schemes improve the lifetime at

the cost of either latency or packet loss. The aim of our work is

to improve the energy efficiency of the sensor network without

packet loss. The STEM (Sparse Topology and Energy

Management) reduces energy consumption in the monitoring

state, by switching the radio off when they are not in use.

Geographic Adaptive Fidelity (GAF) reduces the energy

consumption of the sensor nodes by dividing the sensor

network into small virtual square grids and keeps only one

node in active state in each grid which will take part in the

traffic forwarding process. Since STEM and GAF are

orthogonal to each other, more energy can be saved by

integrating these two schemes but the packet loss due to

unreachable corner arises. To overcome this problem we have

replaced the square grids in GAF by hexagonal grids (GAF-h)

and integrate STEM with GAF-h. Our proposed work has

improved the energy efficiency of the sensor network without packet loss.

Keywords— sensor networks, topology management, energy efficiency and network lifetime

I. INTRODUCTION

sensor network is composed of a large number of sensor nodes that are densely deployed either inside the area to be monitored or very close to it. The position of

sensor nodes need not be predetermined. Hence random deployment in inaccessible terrains or disaster relief operations is possible. On the other hand, this also means that sensor network protocols and algorithms must possess self-organizing capabilities. Sensor nodes consist of sensing unit, processing unit, transceiver unit and power unit. The sensing unit consists of sensors which will sense the interested parameter. The output data of the sensor unit is partially processed by the onboard processor in the sensing unit. The partially processed data is then transmitted to the sink through multi hop wireless path. The power unit consists of a battery source of limited energy, which will supply power to all other units in the node. Such sensor network has the application in wildlife observation and office

buildings. The sensor networks are used in battlefield and disaster area monitoring.

The major design challenge in sensor network is to increase the operational lifetime of the sensor node as much as possible, because the sensor nodes are powered by the battery of limited energy. The available energy should be utilized efficiently. In terms of energy consumption, the wireless exchange of data between the nodes will strongly dominates the other node functions such as sensing and processing. The radio in transceiver unit consume almost the same energy in receive and idle states. Most of the time, the sensor network is in monitoring state and the radio is being idle. Since the significant energy saving can be obtained by switch off the radio when it is idle.

The aim of the topology management scheme is to maintain the sufficient network connectivity between the nodes such that the data can be forwarded efficiently to the sink. The existing topology management scheme called Geographic Adaptive Fidelity (GAF) reduces the energy consumption of the sensor node at the cost of packet loss. GAF divides the sensor field into number of virtual square grids and kept only one node in active state in each grid. All other nodes in each grid are kept in sleep mode and hence it will reduce the energy consumption of the sensor nodes.

Sparse Topology and Energy Management (STEM) is a topology management scheme which will reduce the energy consumption of the sensor node at the cost of latency. STEM turns off the radio and it will wakes up its radio when it is needed to transmit data to the sink. Now, the problem is that the radio of the next hop in the path to the data sink is still turned off. To overcome this problem, each node periodically turns on its radio for a short time to check whether any other nodes want to communicate with it. In order to avoid the interference between the wake up protocol and the data packet transmission, dual radio with separate radios operating at different frequency bands are used.

GAF controls the node density to save energy without affecting the data forwarding capacity. STEM saves energy by putting the radio of the sensor node into the sleep mode when they are not in use. We can get better result by utilizing the advantages of both schemes. Fortunately STEM and GAF are orthogonal to each other and hence the resulting energy gain will be the full potential of both schemes.

In GAF-h, the sensor field is divided into number of virtual grids and in each grid only one node is kept in active

Energy Efficient Hybrid Topology Management

Scheme for Wireless Sensor Networks

S. Vadivelan #1

, A. Jawahar #2

#Electronics and Communication Engineering, SSN College of Engineering, Chennai, TamilNadu, India 603110

[email protected] [email protected]

A

401

state. The node in active state will act as the grid head and takes part in the routing function. In combined scheme, the GAF-h manages one of the nodes in each grid as grid head based on its available energy. STEM controls grid head between the sleep mode and the active mode.

II. GEOGRAPHIC ADAPTIVE FIDELITY (GAF)

Geographic Adaptive Fidelity (GAF) [3] divides the sensor field into number of virtual square grids of side r. The size of the square grid r is chosen such that all the nodes in a grid can be reached by all the nodes in its horizontal and vertical adjacent grids.

Figure 1. GAF grid size.

In order to make all the nodes in the adjacent grid can be

reachable for all the nodes in the grid the size of the grid is

set as shown in Figure 1. By applying Pythagoras theorem, the value of R is given

by

25rR

To make all the nodes in the adjacent grids are reachable, the sides of the virtual square grid r must be

5

Rr

Where R is radio range and r is the sides of the virtual square grid.

Since all the nodes in a grid can be reached by all the nodes in its adjacent grids, it is possible to perform routing function with single node in each grid. GAF keeps only one node in active mode in each grid and all other nodes in the sleep mode. In sleep mode the sensing and processing units are kept on and the radio unit is switched off. Whenever the node has data to send, it will turn on its radio unit. The active node in each grid will act as the grid head and take part in the data forwarding process. In order to balance the energy consumption of the nodes, the responsibility of grid head is rotating among the nodes in the grid. The grid head is elected based on the remaining energy available in the battery source. Therefore the energy consumption of the node is shared equally by all the nodes in the grid. If we have ns nodes in the square grid, the nodes in the grid will be in active mode for only 1/ns

th of the time. Therefore the energy

consumed by the node is reduced by ns times and hence the lifetime of the node is increased by ns times.

The average number of nodes in each grid is given by

2

2r

L

Nns

Where, ns is the average number of nodes in the grid, N is the total number of nodes deployed, L2 is the area of the sensor field and r2 is the area of a square grid.

A. Unreachable Corner Problem

GAF sizes its virtual grid based on the radio range R such that the nodes in any adjacent grids must be able to reach each other with single hop. To evaluate the upper bound of life time, r is set such that

5

Rr

In GAF, nodes in a grid is reachable to all the nodes in the vertical and horizontal grids using single hop and considered as neighbors but not all the nodes in the diagonal grid are reachable [5] as shown in Figure 2.

Figure 2. Unreachable corner in GAF.

In order to reach the nodes in the diagonal grid, the transmission range can be increased. But if we increase the transmission range, the transceiver consumes more energy. If we keep the transmission range constant and reduce the grid size to overcome the unreachable corner node problem, the number of grid increases and hence the number of active node increases. This leads to more energy consumption and decreases the lifetime of the network.

The Probability of a node is not reachable by another node

is the probability of distance between the nodes is greater

than R.

Punreach=Pdistance between nodes >R (5)

r r

r R

402

B. Diagonal Forwarding Probability

Consider a packet originated from a node and its destinations are evenly distributed through the space. The choice of the next hop cell depends on the position of the destination. When the next hop node falls into a diagonal grid, it may be unreachable.

Figure 3. Adjacent grids in GAF.

The probability that a packet is routed via diagonal grid is

Pgridsadjacent ofnumber Total

gridsadjacent diagonal ofNumber diagonal via

5.08

4 diagonal Pvia

C. Loss Probability in GAF

When the node is not reachable to the node in the adjacent diagonal grid, then the data packets send to that node will be lost or dropped. So the retransmission of the lost packets will be required which will take alternate path to the destination. The alternate path may have more number of hops and hence more number of nodes spends their energy to forward the packets to the sink. It also causes congestion in traffic flow and hence the throughput of the network is reduced.

The probability of packet loss due to unreachable nodes

can be written as

P packet loss = P un reach × P via diagonal

To overcome the problem of packet loss due to unreachable nodes, we have replaced the square virtual grid in GAF by the hexagonal virtual grid [5]. In hexagonal virtual mesh each hexagonal grid has six neighbor grids. The GAF with hexagonal grids is called as GAF-h. The size of the hexagon is based on the radio range of the transmitter. The size of the hexagon is chosen such that a node in the hexagonal grid can communicate with all nodes in its six adjacent hexagonal grids within single hop.

The virtual hexagonal grid is formed such that all the nodes in a grid can reach all the nodes in the neighbor grids

within the single hop. To evaluate the upper bound of life time of the sensor nodes, the side r of the hexagon is set to

13

Rr

Where r is the side of the hexagonal grid and R is the transmission range.

Figure 4. Adjacent grids in GAF-h.

Therefore the area a of one hexagonal grid is

2

2

33ra

The area of hexagonal grid in terms of radio transmission range is obtained by applying (9) in (10).

2

26

33Ra (11)

The average number of nodes in a hexagonal grid is

aL

Nnh 2

Where, nh is the average number of nodes in a hexagonal grid and a is the area of one hexagonal grid.

The cell placement of GAF-h topology management scheme is shown in the figure 5.

The horizontal distance between the centers of the hexagonal grid to one half of the side of the hexagon along the straight line is

O

D D

D D

403

rd2

3

The perpendicular distance from the centre of the hexagon to the side of the hexagon in vertical direction is

rh2

3

Figure 5. Hexagonal grid formation.

D. Proposed Model

Our proposed model with hexagonal grid is shown in Figure 6. We have assumed that all the nodes having two radios. The sensor nodes are deployed uniformly random over the sensor field of size 100 × 100. The nodes in sleep mode are indicated by hollow circles and the solid circles indicate the nodes in active mode.

The sensor field is divided into a number of hexagonal virtual grids and in each grid only one node is kept in active mode and all other nodes are kept in sleep mode. The node in active mode will acts as a grid head and take part in the routing function. The responsibility of being as a grid head is rotated among the nodes in the grid based on the available energy level. When the energy level of the grid head reduces below the threshold energy level, the responsibility of grid head is transferred to another node in the grid. In our simulation we have set the threshold energy level as 25%. The grid head is responsible for forwarding the data packets originated from other nodes. When a node is forwarding the data packets, the choice of the next hop node is based on the direction in which the destination sink node is present. When Sparse Topology and Energy Management (STEM) is

integrated, the dual radio concept of STEM is implemented in all the nodes including the grid head.

Figure 6. Proposed model.

The basic idea of STEM is to only turn on the sensing

and some processing circuitries. The transceiver unit is kept in off state. When the sensor node detects an event, it will wake up its main processor to process the data in detail. It will wake up its radio when it is needed to transmit data to the sink. Now, the problem is that the radio of the next hop in the path to the data sink is still turned off. To overcome this problem, each node periodically turns on its radio for a short time to check whether any other nodes want to communicate with it.

The node that wants to communicate is called initiator node and the node it is trying to wake up is called target node. The initiator node wakes up its target node by sending a beacon. When the target node receives the beacon, it will send a response to the initiator node and both the node turn on their radio. Once both the nodes turned its radio on, a link is established between them and data packet is transferred between them. If that packet is intended for other node, the target node will act as the initiator node and send packet to the node in the next hop towards destination and the process is repeated.

In order to avoid the interference between the wake up beacon and the data packet transmission, transceiver uses dual radio and each radio is operating at different frequency bands. The frequency band f1 with low duty cycle is used to transmit the wakeup messages and hence it is called wakeup plane. Once the target node has received a wakeup message, both the nodes will turn on its radio operating at frequency band f2.The original data packets are transmitted in this plane and hence called as data plane.

E. Integration of STEM and GAF-h

GAF-h controls the number of active nodes to save energy without affecting the data forwarding capacity. STEM saves energy by putting the radio of the sensor node into the sleep mode when they are not in use. We can get better result by utilizing the advantages of both schemes. Fortunately STEM and GAF-h are orthogonal to each other and hence the resulting energy gain will be the full potential of both schemes.

h

h

d d

r

404

In GAF-h, the sensor field is divided into number of virtual hexagonal grids and in each grid only one node is kept in active state. The node in active state will act as the grid head and takes part in the routing function. In combined scheme, the GAF-h manages one of the nodes in each grid as grid head based on its available energy. STEM controls grid head between the sleep mode and the active mode.

In GAF-h, the energy consumption of a sensor node is

hn

EE 0

Where, E0 is the energy consumption of a node at normal condition and nh is the number of nodes in that grid. .

When the STEM is interacted with GAF-h, the power

consumption of the node become

h

wakeupdata

n

EEE

Where E is the energy consumption of a node with

hybrid scheme, Edata is the energy consumption of a node in

data plane, Ewakeup is the energy consumption of a node in

wakeup plane and nh is the number of nodes in the grid.

III. PERFORMANCE EVALUATION

We verify our algorithm through simulations. We deploy N number of nodes in an uniformly random fashion over the square field of size L × L and each node has transmission range R. For a network having large number of deployed nodes N, the probability for a node to have n neighbor is approximated by Poisson distribution.

x

n

en

xnP

!)(

The average number of neighbors of a node is denoted by the parameter x.

2

2R

L

Nx

Where, x is the average number of neighbor which

denotes the average number of nodes within the coverage

area of node and R2 is the coverage area of the transmitter of a node.

A. Simulation Setup

In our simulation we have chosen the transmission

range R = 20 m, which has the radio characteristics as

shown in the table 1.

TABLE I. RADIO CHARECTERISTICS

Radio mode Power consumption (mW)

Transmit (Tx) 14.88

Receive (Rx) 12.50

Idle 12.36

Sleep 0.016

We divide the square field of size 100 × 100 into number of square grids of size r × r. We have simulated the

GAF scheme and found out the number of unreachable

corner nodes and the percentage of packets lost. The

simulation is repeated by increasing the number of deployed

nodes of the order of 100.

B. Simulation Results

Figure 7 shows the number of unreachable corner nodes as the function of number of the number of nodes deployed. In GAF the number of unreachable corner nodes increases as the number of deployed nodes increases but in GAF-h the number of unreachable corner nodes is zero. In our simulation the number of unreachable corner nodes is 78 when the total number of deployed nodes is 1000.

Figure 7. Unreachable corners nodes.

Figure 8 shows the percentage of packet lost due to unreachable corners as the function of number of nodes deployed. In GAF the percentage of packet lost increases as the number of nodes deployed increases but in GAF-h, the percentage of packets lost is zero.

Figure 9 shows the normalized energy consumption of sensor node as the function of average number of neighbors in GAF-h, STEM, our proposed scheme and the normalized energy consumption of a sensor node without any schemes. We have plotted the energy consumption of node in STEM

405

scheme with the wake up interval of 600 ms over the observation time of 1000 ms.

Figure 8. Packet loss.

Figure 9. Energy Consumption of a node.

In STEM, the energy consumption of the node is constant with respect to the average number of neighbors,

because STEM works on the time domain. The energy

consumption of a node in GAF-h decreases with average

number of neighbors increases, because the GAF-h

algorithm is related to the density domain. When STEM is

integrated with GAF-h, we got further reduction in the

energy consumption of the sensor node. When the average

number of neighbors is 130, the normalized energy

consumption of a node is 0.16 in GAF-h, 0.38 in STEM and

the normalized energy consumption of a node is 0.06 in the

hybrid scheme of GAF-h and STEM.

IV. CONCLUSION

In this paper we have analyzed the GAF, a topology

management scheme that reduces the energy consumption

of the sensor node by using the node density. But in GAF

the diagonal adjacent grids are not taken into the

consideration. Hence there is the possibility of packet loss

due to unreachable corner nodes. In our work we replace the

square virtual grids by hexagonal virtual grids to overcome

the packet loss due to unreachable corner nodes and hence it

is called as GAF-h. GAF-h reduces the energy consumption of the sensor node by six.

On the other hand, STEM reduces the energy

consumption of the sensor node by switching its radio off

when they are idle. In this work we simulate STEM with

wakeup interval of 600 ms which reduce the energy

consumption of the node by the factor of about 2.5

We have integrated STEM with GAF-h and the hybrid

scheme reduces the energy consumption of the sensor node

by 18 times when the average number of neighbor is 130.

In our proposed scheme the energy conservation of about 18

times is achieved at the cost of latency. So it is worth

investigating how the energy efficiency can be improved with reduced latency.

REFERENCES

[1] I. F. Akyildiz, W. Su, Y. Sankarasubramaniam and E. Cayirci, “A

Survey on Sensor Networks,” IEEE Communications Magazine, vol. 40, pp. 102 - 114, 2002.

[2] I Mo Li, Baijian Yang, “A A Survey on Topology issues in Wireless

Sensor Network,” SPIE – The International Society for Optical Engineering, Orlando, FL, pp. 229-237, April 1999.

[3] Y. Xu, J. Heidemann, D. Estrin, “Geography-informed energy conservation for ad hoc routing,” MobiCom 2001, Rome, Italy, pp.

70-84, July 2001.

[4] Curt Schurgers, Vlasios Tsiatsis, Mani B. Srivastava, “STEM: Topology Management for Energy Efficient Sensor Networks,”

IEEEAC paper #260, Updated Sept 24, 2001.

[5] Ren Ping Liu*, Glynn Rogers, Sihui Zhou, John Zic, “Topology Control with Hexagonal Tessellation,” Feb 16, 2003.

[6] Curt Schurgers, Vlasios Tsiatsis, Saurabh Ganeriwal, Mani B.

Srivastava, “Optmimizing Sensor Networks in the Energy-Latency-Density Design Space,” IEEE Transactions on mobile computing,

vol. 1, No. 1, January-March 2002.

[7] Susumu Matsumae, Vlasios Tsiatsis, Mani B. Srivastava, “Energy-Efficient Cell Partition for Sensor Networks with Location

Information,” Network Protocols and Algorithms, Vol. 1, No. 2., 2009.

[8] Majid I. Khan,Wilfried N. Gansterer, Günter Haring, “Congestion

Avoidance and Energy Efficient Routing Protocol for Wireless Sensor Networks with a Mobile sink,” Journal of Networks, vol. 2,

No. 6, December 2007.

[9] Ya Xu, Solomon Bien, Yutaka Mori, John Heidemann, Deborah

Estrin “Topology Control Protocols to Conserve Energy in Wireless Adhoc Networks,” , January 23, 2003.

[10] Sankalpa Gamwarige, Chulantha Kulasekere, “A Cluster Based

Energy Balancing Strategy to Improve Wireless Sensor Network Lifetime,” Second International Conference on Industrial and

Information Systems, ICIIS 2007, Sri Lanka, 8 – 11 August 2007.

[11] Chuan-Yu Cho, Cheng-Wei Lin and Jia-Shung Wang, “Reliable Grouping GAF Algorithm using Hexagonal Virtual Cell Structure,”

3rd International Conference on Sensing Technology, Tainan, Nov. 30 – Dec. 3, 2008.

[12] Annette Bohm, “State of Art on Energy-Efficient and Latency-

Constrained Networking Protocol for Wireless Sensor Networks ,” Technical Report, IDE0749, June 2007.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

406

RATE LESS FORWARD ERROR CORRECTION USING LDPC

FOR MANETS

N.Khadirkumar

Student, IInd

M.tech(IT Department), Anna University Coimbatore, Coimbatore-22. [email protected]

Abstract The Topology-transparent scheduling for mobile wireless ad hoc networks has been treated as a theoretical

curiosity. This makes two contributions towards its practical deployment. One is cover-free family and another one is rateless forward error correction. As a result from cover free family, a much wider number and variety of constructions for schedules exist to match network conditions. In simulation, I closely match the theoretical bound on expected throughput by using rateless forward error correction (RFEC). Since the wireless medium is inherently unreliable, RFEC also offers some measure of

automatic adaptation to channel load. These contributions renew interest in topology-transparent scheduling when delay is a principal objective.

Index Terms:

Mobile ad hoc Networks, rateless forward error correction, Topology-transparent scheduling

Introduction

Mobile ad hoc Networks is a collection of mobile

nodes that are dynamically communicating without

centralized supervision. It is self-creating, self-

organizing and self-administrating network. Absence

of the base station from the network necessitates the

functionality of the network nodes to include routing

as well. This task becomes more complex as the network nodes change randomly their positions. An

efficient routing protocol that minimizes the access

delay and power consumption while maximizing

utilization of resources remains a challenge for the

ad-hoc network design.

For these reasons we have considered efficient

routing protocols and we have evaluated their

performances on a different MAC layers.

Each device in a MANET is free to move

independently in any direction, and will therefore

change its links to other devices frequent. The

medium access protocol attempts to order to

minimum delay and maximum throughput on a per

hop basis on each nodes

A Vehicular Ad-Hoc Network, or VANET, is a

form of Mobile ad-hoc network, to provide

communications among nearby vehicles and between vehicles and nearby fixed equipment, usually

described as roadside equipment

Intelligent vehicular ad hoc networks

(InVANETs) are a kind of artificial intelligence that

helps vehicles to behave in intelligent manners during

vehicle-to-vehicle collisions, accidents, drunken

driving etc.

Scheduled approaches to channel access provide

deterministic rather than probabilistic delay

guarantees. This is important for applications

sensitive to maximum delay. Furthermore, the control overhead and carrier sensing associated with

contention MAC protocols can be considerable in

terms of time and energy [1].

Two approaches have emerged in response to

topology changes. Topology-dependent protocols

alternate between a contention phase in which

neighbor information is collected, and a scheduled

phase in which nodes follow a schedule constructed

using the neighbor information [2],[3]. Topology-

transparent protocols are to design schedules that are

independent of the detailed network topology. The

schedules do not depend on the identity of a node’s neighbors, but rather on how many of them are

transmitting. Even if a node’s neighbors change its

schedule does not change. The schedule is still

succeeds when the number of neighbors does not

exceed the designed bound.

407

The main objective of this paper is to reduce

delay and maximum throughput for mobile nodes.

Part 2, defines a cover-free family and examines time

division multiplexing and as I also derive the bound

on expected throughput. Part 3, discusses

acknowledgment schemes including RFEC for this purpose and overviews the LDPC process. In part 4,

the simulation environment is explained and in part 5,

the conclusion is stated.

RELATED WORK

The extensive work related to this paper can be

categorized into cover-free family and rateless

forward error correction

Cover Free Family

In designing a topology-transparent

transmission schedule with parameters N and D we

are interested in the following combinatorial property. For each node, we must guarantee that if a

node ʋi has at most D neighbors its schedule Si

guarantees a collision-free transmission to each

neighbor.

This is precisely a D cover-free family. These are

equivalent to D disjoint matrices and to superimposed

codes of order. As a result, there is also equivalence

between the no-tation used for cover-free families,

disjunct matrices, and superimposed codes. Such

combinatorial designs arise in many other

applications in networking. As we showed existing constructions for

topology- transparent schedules correspond to, time

division multiple access giving cover-free families.

Since this is essential to the provision of topology-

transparent schemes of sufficient variety and number

for practical applications, we outline this connection

in more detail.

TDMA-based MAC protocol developed for low

rate and reliable data transportation with the view of

prolonging the network lifetime,

Adapted from LMAC protocol. Compared to conventional -based protocols, which depend on

central node manager to allocate the time slot for

nodes within the cluster, our protocol uses distributed

technique where node selects its own time slot by

collecting its neighborhood information. The protocol

uses the supplied energy efficiently by applying a

scheduled power down mode when there is no data

transmission activity.

The protocol is structured into several frames,

where each frame consists of several time slots. As

shown in each node transmits a message at the beginning of its time slot, which is used for two

purposes; as synchronization signal and neighbor

information exchanges.

By using this message, the controlled node

informs which of its neighboring nodes will be

participating in the next data session. The intended

nodes need to stay in listening mode in order to be

able to receive the intended packet, while other nodes

turn to power down mode until the end of the current time. TDMA calculates collision frequency for each

nodes and automatically send packets from the source

to destination. By allocating time we can easily find

collision and thus it reduces time and the throughput

increases.

The operation of time slot assignment in A-MAC

is divided into four states; initial, wait, discover, and

active. As illustrated in the Fig. 4 below, a new node

that enters a network starts its operation in initial

state where node listens to the channel for its

neighbor’s beacon message in order to synchronize

with the network. Node starts synchronization when it receives a beacon message from one of its

neighbors and adjusts its timer by subtracting the

beacon received time with beacon transmission time.

Node remains in this state for a Listen frames in

order to find the strongest beacon signal. This is

important as to continuously receive the signal from

the synchronized node.

Else, a potential synchronization problem with

the rest of neighboring nodes might arise due to the

resulted drift problem caused by imprecision of

microcontroller’s timer. analyze the effects of MAC protocols, four ad hoc routing protocols are selected

for study. First, the Dynamic Source Routing (DSR)

and Ad hoc On-Demand Distance Vector Routing

protocol are included as examples of on demand

protocols. On-demand protocols only establish routes

when they are needed by a source node, and only

maintain these routes as long as the source node

requires them.

Structure of A-MAC frame

408

Next Destination-Sequenced Distance-Vector

(DSDV) and Wireless Routing Protocol (WRP) that

are distance vector table driven protocols. Table-

driven protocols periodically exchange routing table

information in an attempt to maintain an up-to-date

route from each node to every other node in the network at all times.

ACKNOWLEDGMENT SCHEMES

To approach the theoretical bound for expected

throughput of topology-transparent scheduling in

practice an acknowledgment scheme is required.

Without an acknowledgment, a node must transmit

the same packet in each of its assigned slots to

guarantee reception to a specific neighbor. This is

because while the schedule guarantees a collision-

free slot to each neighbor by the end of the frame, it

is not known which of its slots is successful to a specific neighbor; this depends on the schedules of

the nodes currently in its neighborhood.

With forward error correction (FEC), the source

includes enough redundancy in the encoded packets

to allow the destination to decode the message. Most

FEC schemes require knowledge of the loss rate on

the channel. Determining a suitable rate for the code

in practice is not easy. If the rate is chosen

conservatively to account both for collisions and for

communication errors, as well as allowing for the

maximum number of permitted active neighbors, many additional packets are sent containing

redundant information. Too low a rate decreases

throughput, while too high a rate fails to deliver

enough information to decode. Worse yet, adapting to

a more suitable rate requires an agreement between

transmitter and receiver to change the encoding in

use.

With backward error correction, the destination

explicitly returns feedback to the source. These

techniques may require the source to wait an entire

frame for receipt of the feedback, even if both

transmitter and receiver have at most D neighbors. In the pathological case that the transmitter is densely

surrounded by neighbors while receiver is not,

acknowledgment can cause collisions at the

transmitter and result in total loss; this may result in

stalling for many frames. Further, these techniques

require window buffer, and timer management, not to

mention that packets suffering collision need

retransmission

Rateless Forward Error Correction

Rateless FEC overcomes numerous concerns with acknowledgment in topology-transparent schemes.

Among the rateless FEC codes currently available I

use LDPC code. The LDPC process is capable of

generating a potentially infinite number of equally

useful symbols from a given input, giving the codes

immunity to tolerate arbitrary losses in the channel.

This makes LDPC codes an effective coding technique for wireless channels. for a symmetric

memory-less channel. The noise threshold defines an

upper bound for the channel noise up to which the

probability of lost information can be made as small

as desired. Using iterative belief propagation

techniques, LDPC codes can be decoded in time

linear to their block length.

low-density parity-check (LDPC) code

A low-density parity-check (LDPC) code is a

linear error correcting code, a method of transmitting

a message over a noisy transmission channel, and is constructed using a sparse bipartite graph. LDPC

codes are capacity-approaching codes, which means

that practical constructions exist that allow the noise

threshold to be set very close to the theoretical

maximum for a symmetric memory-less channel.

The noise threshold defines an upper bound for the

channel noise up to which the probability of lost

information can be made as small as desired. Using

iterative belief propagation techniques, LDPC codes

can be decoded in time linear to their block length.

LDPC code using Forney's factor graph notation.

In this graph, n variable nodes in the top of the graph

are connected to (n–k) constraint nodes in the bottom

of the graph. This is a popular way of graphically

representing an (n, k) LDPC code. The bits of a valid

message, when placed on the T's at the top of the

graph, satisfy the graphical constraints. Specifically,

all lines connecting to a variable node have the same

value, and all values connecting to a factor node must sum, modulo two, to zero.

409

SIMULATION

This work is implemented using the Network Simulator Ns-2. The simulation environment is

chosen with the following parameters:

1. Number of nodes : 100 2. Antenna Directional : Omni

3. Network Area : 1500 * 1500 m

4. MAC Layer : IEEE 802.11 CSMA/CD

5. Routing Protocol : DSR protocol

6. Node Max Speed : 5 m/s

7. Mobility Model : Random Waypoint

8. Data rate : LDPC rate

9. Wireless interface : 11 MBPS

When compared to 802.11,the TDMA has minimum

delay. Because the data transfer in 802.11 consumes

more time.

PACKET DELIVERY RATIO

Packet received

The performance of topology-transparent

scheduling using schedules generated from an

TDMA is measured by two metrics, throughput and

delay. I define throughput as the average number of

successful transmissions by a node in a frame. In the

best case a node can have as many as successes. If

the degree of the node is at most , at least one success

is guaranteed. The delay incurred at the MAC layer is

defined as the amount of time taken on average for a packet to reach its next-hop destination; this includes

queuing delay

Conclusion

The combinational characterization leads not

only to more general construction schemes but also to

analysis results suggesting that topology-transparent

schemes retain strong throughput and delay

performance even when in an environment with

neighborhoods larger than anticipated. The fundamental problem, from the beginning, has been

to develop a realistic acknowledgment model that

realizes the performance indicated by a theory based

on omniscient acknowledgment (OMN) and in which

collision is the only cause of erasures. Rate less

forward error correction (RFEC) has been proposed

here as a solution, and a practical implementation

using LT codes described. We emphasize that LT

codes is just one of a number of schemes that could

be used. The simulation results examine the case of

unicast traffic when every packet follow a single router. The technique opens the door for a true

multicast and reliable broadcast and in this cases

RFEC appears to be not just the best, but perhaps the

only currently viable acknowledgement scheme. To

validate this solution, experiments have been

conducted using topology-transparent schedules

based on orthogonal arrays, to compare OMN and

RFEC, and to explore the analytical model developed

earlier.

REFERENCES 1. Rate less Forward Error Correction for Topology-

Transparent Scheduling Violet R. Strontium,

410

Member, IEEE, Charles J. Coburn, and South

Yellamraju

2. J.-H. Ju and V. O. K. Li, “An optimal topology-

transparent scheduling method in multihop packet

radio networks,” IEEE/ACM Trans. Networking, vol.

6, no. 3, pp. 298–306, Jun. 1998.

3. C. J. Colbourn, A. C. H. Ling, and V. R. Syrotiuk,

“Cover-free families and topology-transparent

scheduling for MANETs,” Designs, Codes, and

Cryptography, vol. 32, no. 1–3, pp. 35–65, May 2004.

4. V. R. Syrotiuk, C. J. Colbourn, and A. C. H. Ling,

“Topology-transparent scheduling for MANETs

using orthogonal arrays,” in Proc. DIAL-M/POMC

JointWorkshop on Foundations of Mobile

Computing, Sep. 2003, pp. 43–49.

5. D. R. Stinson, R. Wei, and L. Zhu, “Some new

bounds for cover-free families,” J. Combinatorial

Theory, ser. A, vol. 90, pp. 224–234, 2000.

6. M. Ruszinkó, “On the upper bound of the size of

the -cover-free families,” J. Combinatorial Theory,

ser. A, vol. 66, pp. 302–310, 1994.

7. D.-Z. Du and F. K. Hwang, Combinatorial Group

Testing and Its Applications,

ser. Series on Applied Mathematics, 2nd ed.

Singapore:World Scientific, 2000, vol. 12.

8. A. D’yachkov, V. Rykov, and A. M. Rashad,

“Superimposed distance

codes,” Problems Control and Information Theory,

vol. 18, pp.

237–250, 1989.

9.C. J. Colbourn, J. H. Dinitz, and D. R. Stinson,

“Applications of combinatorial designs to

communications, cryptography, and networking,” in

Surveys in Combinatorics, ser. Lecture Note Series

267, J. D. Lamb

and D. A. Preece, Eds. London, U.K.: London

Mathematical Society,

1999, pp. 37–100.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

411

APPLICATION OF GENETIC ALGORITHM FOR THE DESIGN OF CIRCULAR MICROSTRIP ANTENNA

Mr. Rajendra Kumar Sethi,Sr. Lect,GIET,Gunupur,Orissa,[email protected] Ms. Nirupama Tripathy,Lect.,GIET,Gunupur,Orissa,[email protected]

Abstract

Microstrip antennas are widely used

in the telecommunication field. It has

advantages like low profile, low cost, ease

of construction, conformal geometry and

flexibility in terms of radiation pattern, gain

and polarization etc. The MoM based

Zealand IE3D software has been opted as an

efficient tool for microstrip antenna design

and analysis. On the other hand evolutionary

techniques like Genetic Algorithm, Particle

Swarm Optimization, Bacterial Foraging

Optimization and Artificial Neural Networks

etc have gained a great importance in the

field of antenna application. These are all

based on biological concepts. In this paper

we have applied Genetic Algorithm for

optimization of Circular microstrip antenna.

A genetic algorithm is natural selection

procedure, i.e survival of the fittest or

Darwinian principle. It is a population to

population global search technique based on

statistics.

Keywords

MoM, GA, IE3D, Crossover, Mutation

Introduction

Microstrip antenna has very low

bandwidth. Hence it is very important to

find an accurate dimension and its feed

position to efficiently operate such antenna.

There are many empirical formulae for

different regular structure microstrip patch

antenna for calculating the resonant

frequency. However, resonant frequency

being a non-linear function of parameters

like the physical dimensions and material

property of the antenna, it is quite difficult

to adjust all these parameters simultaneously

to design a microstrip patch antenna for a

particular operating frequency. Therefore,

Genetic Algorithm is applied in this thesis to

optimize such problems. Genetic Algorithm

performs its searching process through

population to population instead of point-to-

point search. The most favorite advantage of

Genetic Algorithm is its parallel architecture.

They use probabilistic and deterministic

rules.

Genetic Algorithm Overview

412

Generate initial population G(0) at random, i=0

Determine fitness of every individual in population G(i)

Select parents from G(i), based on

their fitness, and a genetic operator.

Apply genetic operator on selected parents to create one or more

individuals for G(i+1)

G(i+1) is filled up?

i=i+1

Termination criteria satisfied?

Yes

NO

Yes NO

Find the fittest individual in G(i), And Quit

Yes

a) A Genetic Algorithm is a global search

technique based on Darwinian principle,

i.e., survival of the fittest.

Figure. 1 Flow Chart of Genetic Algorithm

It performs following six basic

tasks:-

b) Encode the solution parameters as

genes,

c) Create a string of genes to form a

chromosome,

d) Initialize a starting population,

e) Evaluate and assign fitness values to

individuals in the population,

f) Perform reproduction through the

fitness-weighted selection of individuals

from the population, and

g) Perform crossover and mutation to

produce members of new generation

Design of Circular Microstrip Patch

Antenna using Genetic Algorithm

Circular microstrip antenna, due to

its simple design features is popular in

industrial and commercial applications.

However, due to inherent narrow bandwidth,

the resonant frequency or the dimension of

the patch antenna must be predicted

accurately.

Genetic Algorithm (GA) has been

applied to calculate the optimized radius of

Circular Microstrip Antennae. Resonant

frequency (f) in the dominant TM11 mode,

dielectric constant (r) and thickness of the

substrate (h) are taken as inputs to GA,

which gives the optimized radii (a) of the

antennae. We have applied Method of

Moment (MOM) based IE3D software of the

Zealand Inc., USA to validate experimental

results with that of GA outputs. It is shown

in the table that the GA results are more

accurate while taking less computational

413

time. The results are in good agreement with

experimental findings.

The circular patch antenna with its

design parameters i.e. thickness of substrate

„h‟ and radius of circular patch „a‟, is shown

in figure 2.

Fig. 2 Circular Patch Antenna

The resonant frequency of circular

microstrip antenna is expressed as:

2/1

0

))]65.1268.0()77.144.1()2

(ln(2

1[2

84118.1

rr

r

eff

r

a

h

h

a

a

ha

cf

[1]

Where, the effective dielectric constant

( eff), is given by

2/112

12

1

2

1

a

hrr

eff

[2]

And c0 is the velocity of light.

Equation (1) is used as the fitness

function of GA to optimize radius of the

patch of the antennae. The population size is

taken 20 individuals, and 200 generations

are produced. The probability of crossover is

set at 0.25, while the probability of mutation

was equal to 0.01. Resonant frequency (fr),

dielectric constant (r) and thickness of the

substrate (h) are given as inputs to GA,

which gives the optimized radii (a) of the

antennae. The comparisons of GA and

results obtained by IE3D software are listed

in table 1 for nine different fabricated

circular microstrip antennae. The optimized

radii (a) obtained using GA are in good

agreement with the experimental results as

listed in column „VII‟ of table 1.

Using these calculated radius (a) in

IE3D simulation software, resonant

frequencies are calculated which almost

match with the input resonant frequencies

used as input, thus, validating the results of

GA. The percentage of error in calculation

of radius using GA, are listed in column VI.

Average error obtained using GA is only

0.65.

Table 1 Comparison of Genetic Algorithm Result

with Experimental Result

I II III IV V VI VII

a A

n te

nna

No. rf

In

GHz

r

h

In

mm

a

In

mm

By GA

Error

In %

DIEf 3

In

GHz

1 4.945 4.55 2.35 7.6742 0.306234 4.94

2 3.75 4.55 2.35 10.3837 0.156731 3.735

3 2.003 4.55 2.35 20.0659 0.3295 2.02

4 1.03 4.55 2.35 39.5602 0.477484 1.05

5 0.825 4.55 2.35 49.502 0.0040404 0.82

414

6 1.51 2.33 3.175 35.2043 0.785285 1.53

7 4.07 2.33 0.794 13.0196 2.51654 4.12

The return loss plots calculated using IE3D

simulation software for antenna number

1and antenna number 2 are shown in figure

3 and figure 4 respectively.

Fig. 3 Return Loss Plot for Antenna No. 1

Fig. 4 Return Loss Plot for Antenna No. 2

Seven antennae are optimized to

validate the developed code using GA. IE3D

software and experimental results are used

to compare and hence, to validate the

obtained results by GA. Design parameter as

obtained using GA are used to simulate the

antenna using IE3D. Return loss plots are

presented for simulated antennas. As seen,

the results obtained using GA are more close

to experimental results. Thus, a highly

selected fitness function in GA will give

much accurate result. Application of GA to

microstrip antenna design seems to be an

accurate, computationally simple and cost

effective method, which may go a long way

in antenna design.

Conclusion & Future Scope

Recent trends of miniaturized microstrip

antenna design and computational

electromagnetic analysis demands efficient

soft computing tools that give high accuracy

with less computational time. In this

scenario, genetic algorithm has become a

good tool for the engineers. In this paper,

GA has been used efficiently for the design

of microstrip antennas. Genetic algorithm

may be improved by taking new crossover

415

and mutation operators in continuous micro-

genetic algorithms. The proposed methods

may be applied in wide range applications

like calculation of radiation pattern, gain

array factor correction etc. of microstrip

antennas.

References

1. J. Watkins, “Circular Resonant

Structures in Microstrip,” Electronic

Letters, Vol. 5, No. 21, Oct. 1969, p.

524.

2. R. K. Mishra and A. Patnaik,

“Design of Circular Microstrip

Antenna using Neural Networks,”

IETE Journal of Research, vol. 44,

pp. 35-39, Jan-Apr 1998.

3. X. Gang, “On the resonant

frequencies of microstrip antennas,”

IEEE Trans. Antennas Propagat., pp.

245-247, Feb. 1989.

4. G. A. Deschamps, “Microstrip

Microwave Antennas,” The Third

Symposium on The USAF Antenna

Research and Development

Program, University of Illinois,

Monticello, Illinois, October 18-22,

1953.

5. R. Garg, P. Bhartia, I. Bahl and A.

Ittipiboon, “Microstrip Antenna

Design Handbook,” Artech House,

Inc, 2001.

6. Pozar DM, Schaubert DH.

Microstrip Antennas: The Analysis

and Design of Microstrip Antennas

and Arrays. IEEE Press: New York,

1995.

7. G. Derneryd, “New Technologies

for Lightweight, Low Cost

Antennas,” Workshop on

lightweight antennas, 18th European

Microwave Conference, Stockholm,

Sweden, 1988.

8. G. Kumar and K. P. Ray,

“Broadband Microstrip Antennas,”

Artech House, Inc, 2003.

9. D. S. Weile and E. Michielssen,

“Genetic algorithm optimization

applied to electromagnetics: A

review,” IEEE Trans. Antennas

propagat., vol. 45, Mar. 1997, pp.

343-353.

10. J. M. Johnson and Y. Rahmat-Samii,

“Genetic algorithms in engineering

electromagnetics,” IEEE Antennas

Propagat. Mag., vol. 39, Aug. 1997,

pp. 7-21.

11. Goldberg, D. E., "Genetic

Algorithms in Search, Optimization

and Machine Learning," Addison-

Wesley, 1989.

12. IE3D Software with its

manual(student version) from

www.Zealand.com

Bhaarath Ramesh Vijay

Department of Electronics and Communication

Sri Venkateswara College of Engineering

Pennalur, Sriperumbudur, India.

E-Mail id: [email protected]

Abstract:

This is a research paper proposes a design of Water

Quality monitoring System Using GSM Service for the

Aqua - Culture based Industries. This design, if

implemented, helps in monitoring the water quality

remotely, via GSM (by SMS). It is compulsory for an every

officer from his industry to visit the ponds at a designated

time and perform manual testing on finding the purity

level of the water. But it is also known, that these kind of

technique will consume lot time and effort. This research

project focuses on developing a prototype that can

evaluate data collected through three bases: Oxygen

dissolved in the water, Level of pH in water,

Temperature level of the water. This design also has the

capability conducting the tests automatically with a timer

present in it. It also sends the degra- adation of water

quality in the pond via SMS (Short Messaging Service).

1. Introduction

During the country’s fresh water resources consist

of 195210 kilometers of rivers and canals, 2.9 million hectares

of Minor and major reservoirs, 2.4 million hectares of ponds

and Lakes and about 0.8 million hectares of flood plain lakes

and derelict water bodies. During the ten-year period of 1995-

2004 inland capture production grew from 600,000 tons to

800,000 tones and at present contributes to 13% of total fish

production of the country [1].

Among issues faced by industry is monitoring

by conducting water quality for their ponds. Maintaining

water quality is important in farming aquaculture organisms, as

they are sensitive to water condition.

In the past, aqua farmers monitor quality of water by

conducting colorimetric test. The test is conducted to measure

ammonia level, pH level and dissolved oxygen (DO) level in

the water. The test is conducted by trained and skilled staff

Kapil Adhikesavalu Department of Electronics and Communication

Sri Venkateswara College of Engineering

Pennalur, Sriperumbudur, India.

E-Mail id: [email protected]

Furthermore, the process is very tedious to execute. To solve

the problem, water quality monitoring system was introduced to

farm [2].

The aims of the research project are to implement an

automated remote water quality monitoring system via SMS as

well as an alert system to inform the degradation of water

quality.

The objectives of this paper are to disclose why the three

criteria namely Dissolved Oxygen level, pH level and

temperature level are used as the parameters to monitor water

conditions, to describe the development process that will take

place in implementing this system and lastly to explain the

architecture of the remote automated water monitoring system.

2. Literature Review

2.1 Importance of Quality in water for aquaculture.

Water is a ‘Universal Solvent’ where various chemical

dissolved in the water, as well as all physical attributes affecting

them combines to form water quality. Good water quality level

determined by all attribute present in the water at an appropriate

level and often aquaculture water quality does not equal to

environmental water quality criteria differ from species to

species [3].

Physical, chemical, and biological properties are

interrelated and it affects survival, growth and reproduction of

aquaculture. Aquaculture also can have reverse effect to the

environment as aquatic organism, consume oxygen and

produce and ammonia. Important water quality parameters to

be considered are; temperature, salinity, pH, dissolved

oxygen, ammonia, nitrite/nitrate, hardness, alkalinity, and

turbidity.

Proceedings of the third National conference on RTICT 2010Bannari Amman Insitute of Technology , Sathyamangalam 638 401

9-10 April 2010

416

2.2 Water Quality Parameter

P. Fowler, et al. [4] in their study recommended that

temperature, DO, and pH be monitored directly on a

continuous basis since they tend to change rapidly and have a

significant adverse effect on the system if allowed to operate

out-of- range.

Temperature refers to degree of hotness or coldness and

it can be measured in degree Celsius. Temperature of water

needs to be monitored regularly as outside tolerable temperature

range, disease and stress become more prevalent. Among effect

of temperature changes are; photosynthetic activity, diffusion

rate or gases, amount of oxygen that can be dissolved, and

physiological process of the prawn and level of other

parameters [5].

pH refers to the hydrogen ion concentration or how acidic

or basic as water is pH is defines as log[H+]. pH value range

form 0-14; pH = 7 is neutral, ph < 7 is acidic, pH > 7 is basic.

Fig.1 explains the effects of pH to prawn.

Figure 1: Effect of pH to Prawn

Dissolved oxygen describes the concentration of oxygen

molecular in the water and it dependent on the temperature of

the water and the biological demand of the system [3].

Dissolved oxygen is used in aerobic decomposition of organic

matter, respiration of aquatic organism, and chemical oxidation

of mineral. As many organisms in the water use dissolved

oxygen, therefore it tends to change rapidly. Dissolved oxygen

is supplied to water through several method direct diffusion of

oxygen from the atmosphere, wind and wave action; and

photosynthesis [3].

2.3 Automated Water Quality Monitoring System

Aquaculture is the farming of aquatic organism in natural

or controlled marine or freshwater environments [4].

Aquaculture history has started during 19th century in Tonle Sap

Lake, Kampuchea and Indonesia. Farmers used cages made by

bamboo and later the technique was improved by using steel

tubes and later the technique was improved by using steel tubes

and floating drums t o act as cage. Since 20 years ago,

aquaculture industry is blooming rapidly and it is most popular

method in fishery sector [5].

‘Intelligent Aqua Farm System via SMS’ is an example

of remote monitoring and maintaining water quality system of

aquaculture ponds. The system automatically monitors and

records real-time data of two parameters; pH level and DO

level, which are reported through Short Messaging Service [5].

2.4 Evaluating Water Quality Degradation

A study has been done by Faculty of Engineering, Akdeniz

University using fuzzy logic to assess ground water pollutions

levels below agriculture field. The study was conducted by

taking into consideration four inputs; nitrite, nitrate,

orthophosphate, and seepage index value. Water quality index

is then determined using fuzzy logic. Water quality index was

originally designed to make an integrated assessment of water

quality conditions to meet utilization goals. In the study, a fuzzy

logic system was developed to assess the ground water pollution

of Kumluca plain at previously selected nine sampling stations.

The applied water pollution evaluation system involves the

selection of water quality parameters and index values to form

Water Pollution Index (WPI) [7].

Other example of research project using fuzzy logic is a

case study by Virgili University. The study used fuzzy logic

due to limitation of current water quality index. The study

indicated the need for more appropriated techniques to manage

the importance of water quality variables, the interpretation of

an acceptable range for each parameter, and the method used to

integrate dissimilar parameters involved in the evaluation

process is clearly recognized. Therefore they propose fuzzy

logic inference to solve the problem. The study proposed

dissolve oxygen and organic matters as input for the evaluation.

Several sets of rule were define to help them with the evaluation

[8].

Meanwhile, a study conducted by et al [6] used for

model; Dissolved Oxygen predicting model, unilinear

temperature model, BOD-DO multi linear model, and one

dimension zoology model in evaluating and predicting water

quality of ponds [6].

2.5 Relationship among the Water Quality Parameters

P.Fowler, et al in their study recommended that temperature,

DO, and pH be monitored directly on a continuous basis since

they tend to change rapidly and have a significantly adverse

effect on the system if allowed to operate out of range [12].

417

During day, aquatic organism respiration uses oxygen

and produces carbon dioxide. The carbon dioxide is then used

by aquatic plant through photosynthesis and it produces oxygen

as it by product. The cycle continues during daytime. However,

during nighttime, aquatic organisms keeps using oxygen but

carbon dioxide produced dissolve in water as photosynthesis do

not occurs during night. As for that, concentration of oxygen in

water reduces [13].

Table 1 and Table 2 below describe relation of dissolved

oxygen, carbon dioxide and pH in ponds over 24 hours, as well

as summarized tolerable range of water quality parameter for

prawn farming [14].

2.6 Data Telemetry

Remote monitoring system is an important technology in

today’s life. Our society for instance is pervaded by computer-

controlled devices. Everything from Digital Clock to Hi-fi

Robot can have one or often several microcomputer devices and

can be controlled remotely [9]. Remote monitoring system

employs a technology called data telemetry. It allows remote

measurement and reporting information of interest to the system

designer or operator [5].

The most wide spread, wide area, wireless, digital network is

the GSM network. GSM is a digital cellular network,

originally intended for voice telephony [9]. SMS (Short

messaging service) is an application that can be utilized from

GSM network. Message sent, which is in text only, from the

sending mobile is stored in a central short message centre

(SMC) which will then be forwarded to the destination mobile

[5].

3. Research Project Development

Figure 2 depicts the development flow of the

research project. Development of the research project starts

with preliminary research to obtain adequate

information regarding the problem was identified and agreed,

design phase takes place.

After that, development of the research project

starts by configuring hardware to be used, followed by

developing the software required. After both hardware and

software are completed, the system is then tested. The

development of the research project follows the figure above

where at each phases; evaluation will be done to determine

whether the design/system developed fulfills the requirement.

Figure 2

Flowchart Diagram for Development of Research project

418

4. System Architecture

The system architecture, which represents skeleton of the

research project, consists of four components; data acquisition

System, data telemetry, data processing and output and is

depicted in Figure 3.

4.1 Power supply and UPS

This block helps the monitoring system to be in activation

mode 24×7. In case of any power failure, the UPS which can

withstand for long hours can help in continuing the activation of

the monitoring system.

4.2 Timer Count down

This equipment controls the device regularly. It has a timer

in it. When the count is done the test is done and the acquired

information is transmitted. The tests can be performed regularly

without any manual help. They are programmed that the test

can be performed once in a day and send the details. Also the

staff in-charge can also program the timing for the tests to be

conducted.

4.3 Data Acquisition System

Data Acquisition system is consists of two components;

sensors and data acquisition kit. Sensors are needed to pick

particular reading from the pond and the reading would be

gathered by data acquisition kit. Data acquisition kit is

responsible to convert analog signal received and converts it

into digital signal. The kit is also responsible to prepare

message for data telemetry purposes.

4.4 Data Telemetry

Data telemetry is consisting of GSM modem. The modem

is connected to the data acquisition kit. The modem will receive

prepared messages by the kit and send the message to data

processing component.

4.5 Data Processing

Data processing component would be the heart of the

system. The component consists of four sub-component; data-

base, knowledge base, inference, and evaluation. The first sub-

component stores the message received from data acquisition kit

through short message service. Knowledge base, meanwhile,

stores data related to water quality level for inference purpose.

Inference component would use Mamdani fuzzy logic style.

Real time data received would be fed into the inference to

determine water quality of the pond. Evaluation will be done

using the inference result. This sub-system enables proactive

measure in the monitoring system. Monitored data would not

have to be in intolerable range before it could trigger the

system.

4.6 Output

Result of the water quality evaluation would be used by this

component. Two type of output would be produced; first, the

data received and the output of the evaluation would be

displayed on central computer screen, second, alert message

would be sent to farmers upon detecting degradation of the

water quality.

4.7 Data flow Diagram

Data flow diagram (DFD) is a tool that depicts the flow of

data through as system and the work or processing performed

by the system [11]. Figure 4 shows the data flow diagram for

the project.

419

The system is gets activated from the timer. Then it is

triggered with inputs reading by sensors placed in water. The

reading will be converted to its corresponding voltage value by

the sensors. Then the inputs will be sent to data acquisition kit;

CRS-844, where the analog reading received will be converted

into digital signal. The output will be sent to central system or

the data processing part using GSM Modem connected to the

CRS -844 board.

As the input received is in the voltage form, the input need to

be converted to its actual value and this will be done by the

data processing system built using visual basic.

Inference and Evaluation will be done by the data processing

using fuzzy logic. Output of the evaluation will be compared to

knowledge base. Two outputs will be produced by the system;

evaluation result will be sent through short message service

upon detecting any degradation of the water quality.

5. Conclusion

This research paper proposes system architecture for

proactive automated water quality monitoring system. It is

believed that by having such system, tedious, and cost-

consuming jobs of manual monitoring can be eliminated. The

architecture proposed is consists of four components; data

acquisition, data telemetry, data processing and output.

Three parameters will monitored and used to evaluate the

water quality namely; temperature, pH, and dissolved oxygen.

Fuzzy logic meanwhile is used to evaluate the fuzziness of the

Mamdani style of inference to create the decision over water

quality level of the pond.

Further work includes developing the prototype of the

system and testing the system in actual aqua farm for more

accurate results.

420

6. Reference

[1] FAO fisheries and aquaculture of INDIA

http://www.fao.org/fishery/countrysector/naso_india/en

[2] Handheld Professional Plus Instrument Saves

Time When There’s No Time To Spare. : YSI

Environmental, 2007.

[3] "Water Quality Management and Treatment for

Prawn Hatchery." Marine Aquaculture Institute, Kota

Kuala Muda, Kedah. 13 Jun. 2005.

[4] C. S. Hollingsworth, R. Baldwin, K. Wilda, R.

Ellis, S. Soares, , et al. "Best Management Practices

For Finfish Aquaculture in Massachusetts." UMass

Extension Publication AG-BPFA (2006)

[5] Sharudin, Mohd Shahrul Zharif . "Intelligent

Aqua-Farm System via SMS." Diss. Universiti

Teknologi PETRONAS, 2007.

[6] W. Ruimei, H. Youyuan, F. Zetian., "Evaluation

of the Water Quality of Aquaculture Pond Expert

System." Institute of Agriculture Engineering, Beijing

(2004): 390-397.

[7] A. Muhammetoglu, and A. Yardimci .

"Evaluation of the A Fuzzy Logic Approach to

Assess Groundwater Pollution Level Below

agricultural Fields." Environmental Monitoring and

Assessment, Springer 118 (2005): 337-354.

[8] W. O. Duque, N. F. Huguet, J. L. Domingo, M.

Schuhmacher, , et al. "Assessing Water Quality in

Rivers with Fuzzy Inference Systems: A Case Study."

Environment International 32.6 (2006): 733-742.

[9] Mikael , Sj¨odin. "Remote Monitoring and

Control Using Mobile Phones." Newline Information

(2001)

[10] D. E. Avison, and G. Fitzgerald . "Where now

for development methodologies?." Communications

of the ACM 46.1 (2003): 78-82.

[11] J. L. Whitten, L. D. Bentley, K. C. Dittman,

“System Analysis and Design Methods 5th Edition”,

McGraw Hill, 2001

[12] P. Fowler, D. Baird, R. Bucklin, S. Yerlan, C. Watson &

F. Chapman2, , et al. "Microcontrollers in

Recirculating Aquaculture Systems1." EES 326

(1994): 1-7.

[13] N. Hashim, M. Ibrahim, T. Murugan, . Personal

Interview. 9 Apr. 2008.

[14] . Cadangan Perniagaan Projek Ternakan Udang

Galah. Kuala Lumpur: Department of Fisheries,

Ministry of Agriculture and Agro-Based Industry,

2004.

421

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

422

U Slotted Rectangular Microstrip Antenna for Bandwidth

Enhancement

D.Sugumar 1, Shalet sydney

2

1 & 2Department of Electronics and Communication

Karunya University, Coimbatore 641114, Tamil Nadu, India

[email protected],

[email protected]

Abstract: - A new design technique of microstrip patch antenna is presented in this paper. The

proposed antenna design consists of direct coaxial probe feed technique and the novel u

slotted shaped patch.The composite effect of integrating these techniques and by introducing

the new slotted patch, offer a low profile, high gain, broadband, and compact antenna

element. Parameters like return loss, radiation pattern and bandwidth are analysed using

FDTD algorithm in MATLAB and compared for a non slotted microstrip patch antenna , a

rectangular slotted microstrip patch antenna and a u slotted microstrip patch antenna.

Bandwidth for the u slot design is found to be as large as about 1.785 times that of a

corresponding unslotted rectangular microstrip antenna and also better directivity is achieved.

Details of the antenna design and simulated results are presented and discussed.

Key-Words: -Microstrip Patch Antenna, Bandwidth Enhancement, FDTD Algorithm,MATLAB.

1.Introduction

With the wide spread proliferation of

wireless communication technology in

recent years, the demand for compact, low

profile and broadband antennas has

increased significantly. A number of new

developed techniques to support high data

rate wireless communication for the next

generation technologies have been rapidly

increasing. Basically, the maximum

achievable data rate or capacity for the

ideal band-limited additive white Gaussian

noise (AWGN) channel is related to the

bandwidth and the signal-to-noise ratio

through Shannon-Nyquist criterion :

C=Blog2(1+SNR) (1)

where C denotes the maximum transmit

data rate, B stands for the channel

bandwidth, and SNR is the signal-to-noise

ratio. From this principle, the transmit data

rate can be enhanced by increasing either

the bandwidth occupation or the

transmission power. However, the

transmission power cannot be readily

increased since many portable devices are

battery powered and the potential

interference should also be avoided. Thus,

a large frequency bandwidth seems to be

the proper solution to achieve a high data

rate. To meet the requirement, the

microstrip patch antenna has been

proposed because of its low profile, light

weight and low cost . However,

conventional microstrip patch antenna

suffers from very narrow bandwidth. This

poses a design challenge for the microstrip

antenna designer to meet the broadband

techniques.

There are several well-known methods to

increase the bandwidth of patch antennas,

such as the use of thick substrate, cutting a

resonant slot inside the patch, the use of a

low dielectric substrate, multi- resonator

stack configurations, the use of various

impedance matching and feeding

423

techniques, and the use of slot antenna

geometry. However, the bandwidth and the

size of an antenna are generally mutually

conflicting properties, that is,

improvement of one of the characteristics

normally results in degradation of the

other.

Several techniques have been proposed to

enhance the bandwidth in the state-of-the

art antenna research. A novel single layer

wide-band U slotted rectangular patch

antenna with improved impedance

bandwidth has been demonstrated. By

using a U-slot patch, and coaxial feed

patch antennas, wideband which is

electrically small in size have been

reported .

2.FDTD analysis

Fig. 1. The 3-D FDTD unit cell.

The FDTD algorithm solves Maxwell’s

time-dependent curl equations by first

filling up the computation space by a

number of “Yee cells”. The relative spatial

arrangements of the E fields and the H

fields on the “Yee cell” enables the

conversion of Maxwell’s equations into

finite-difference equations. These

equations are then solved in a time

matching sequence by alternatively

calculating the electric and magnetic fields

in an interlaced spatial field.

The first step in designing an antenna with

an FDTD code is to grid up the object. A

number of parameters must be considered

in order for the code to work successfully.

The grid size must be small enough so that

the fields are sampled sufficiently to

ensure accuracy. Once the grid size is

chosen, the time step is determined such

that numerical instabilities are avoided,

according to the courant stability

condition.

A Gaussian pulse voltage with unit

amplitude, given by

V (t)=exp(−(t−t0)2/T2) (2)

where T denotes the period and t0

identifies the center time, is excited in the

probe feed.In this case, the frequency

domain FDTD near to far field

transformation are used because it is both

memory efficient and time efficient.

3.Design considerations of the

proposed antenna

The geometries of the proposed antennas

are given below:

Fig. 2. Geometry of a broad-band rectangular microstrip antenna

without a slot and having coaxial feed. The dimensions given in

thefigure are in millimeters.

Fig. 3. Geometry of a broad-band rectangular microstrip antenna

with a rectangular slot and coaxial feed. The dimensions given

in thefigure are in millimeters.

424

Fig. 4. Geometry of a broad-band rectangular microstrip antenna

with modified U-shaped slot and coaxial feed. The dimensions

given in the figure are in millimeters.

Fig. 2, shows the geometry of a

rectangular patch antenna without any slot

in it and Fig. 3, shows Geometry of a

broad-band rectangular microstrip antenna

with a rectangular slot. As shown in Fig. 4,

a U shaped slot is placed on the patch. The

rectangular patch has dimensions of

57.5055 mm 47.6975 mm and is printed on

a grounded substrate of thickness(h) 1.6

mm, relative permittivity(ᵋ) 3.2 and size 60

mm ×50 mm. The values of thickness(h) ,

relative permittivity (ᵋ) and resonant

frequency are fixed previously and patch

length and width are determined using

transmission line model. The patch is fed

by a coaxial probe along the centreline(x

axis) at a distance of L6 from the edge of

the patch as shown in figures 2,3, and 4.

TABLE 1

Parameter Value[mm]

L 47.6975

W 57.5055

L1 3.97479

L2 11.924375

L3 3.179839

L4 19.07900

L5 11.92437

L6 26.233625

L7 4.76975

W1 18.68928

W2 21.56457

W3 11.5011

W4 6.46936875

W5 10.06347

d(relative

permitivity)

3.2

4.Results and conclusions

Coding is done in matlab and radiation

patterns for a non slotted microstrip patch

antenna, a rectangular slot microstrip

patch antenna and a u slot microstrip patch

antenna is obtained as follows. It is

observed that for the U slot design , the

side lobes are reduced and the major lobe

has become more prominent increasing the

directivity of the patch antenna.

0

20

40

60

80

0

20

40

60-60

-40

-20

0

20

40

60

Position (microns)Ex

Fig. 5. Radiation pattern of rectangular microstrip patch antenna

without a slot.

0

20

40

60

80

0

20

40

60-40

-20

0

20

40

60

Position (microns)Ex

Fig. 6. Radiation pattern of rectangular microstrip patch antenna

with a rectangular slot.

425

0

20

40

60

80

0

20

40

60-40

-20

0

20

40

60

80

Position (microns)Ex

Fig. 7. Radiation pattern of rectangular microstrip patch antenna

with a U-slot.

Fig. 8. Simulated return loss against frequency for proposed

antennas. Parameters for antennas 1–3 are described in Table

I.and Table 2.

Bandwidth can be said as the frequencies

on both the sides of the centre frequency in

which the characteristics of antenna such

as the input impedance, polarization, beam

width, radiation pattern etc are almost

close to that of this value. It can be

defined as the range of suitable frequencies

within which the performance of the

antenna, w.r.t some characteristic,

conforms to a specific standard.

The bandwidth is the ratio of the upper and

lower frequencies of an operation. It can

be obtained as:

Using equations (2) and (3) bandwidth for

each of the three considered antennas are

calculated .

(3)

(4)

TABLE 2

PERFORMANCES OF THE PROPOSED BROAD-BAND ANTENNAS; Fc IS THE CENTER FREQUENCY , FL AND FH ARE THE

LOWER AND HIGHER FREQUENCIES WITH 10-dB RETURN LOSS IN THE OPERATING

BANDWIDTH, AND THE ANTENNA BANDWIDTH IS DETERMINED FROM FH-FL

FL (MHz)

Fc (MHz)

FH (MHz)

Bandwidth (MHz,%)

Antenna 1 (antenna

without slot)

1636

1666

1721

85,5.1020

Antenna 2 (antenna

with rectangular slot)

1661

1789

1822

161,8.99

Antenna 3 (antenna

with u slot)

1677

1800

1841

164,9.1111

426

The fundamental resonant frequency

of the unslotted rectangular patch

antenna is at about 1.66 GHz, with

an operating bandwidth of 5.102 %.

Since the obtained antenna

bandwidths are as large as 8.99–9.11

%, the proposed antennas show a

much greater operating bandwidth,

more than 1.7 times that of an

unslotted rectangular patch

antenna.Thus a new technique for

enhancing the gain and bandwidth

of a microstrip patch antenna has

been developed and implemented

successfully. The experimental

results demonstrate that it has a

better impedance bandwidth of

9.1111% at 10 dB return loss,

covering from 1.67 to 1.84 GHz

frequency. Techniques for

microstrip broadbanding, size

reduction and stable radiation

pattern are applied with significant

improvement in the design by

employing the proposed u-slotted

patch shaped design and coaxial

probe.

4.References [1]Mohammad Tariqul Islam., Mohammed Nazmus Shakib.,

Norbahiah Misran., Baharudin

Yatim, Aug. 2000,” Analysis of L-Probe Fed Slotted Microstrip Patch

Antenna” IEEE Transactions on

Antennas and Propagation, Volume 48, Issue 8, Page(s):1149 – 1152

[2] Mohammad Tariqul Islam.,

Mohammed Nazmus Shakib., Norbahiah Misran., 2009,” multi-slotted

microstrip patch antenna for wireless

communication” Progress In Electromagnetics Research Letters,

Vol. 10, Page(s):11-18

[3] Mohammad Tariqul Islam., Mohammed Nazmus Shakib.,

Norbahiah Misran., Baharudin Yatim,

25-27 December, 2008, Khulna, Bangladesh 2008,” Analysis of

Broadband Slotted Microstrip Patch

Antenna” Proceedings of 11th

International Conference on Computer

and Information Technology (ICCIT 2008), Vol. 32, Page(s):121-148

[4] Jia-Yi Sze., Kin-Lu Wong, Aug.

2000,”Slotted rectangular microstrip antenna for bandwidth enhancement”

IEEE Transactions on Antennas and

Propagation, Volume 48, Issue 8, Page(s):1149 – 1152

[5] Jeun-Wen Wu., Jui-Han Lu, June

2003,” Slotted circular microstrip antenna for bandwidth enhancement”

IEEE Antennas and Propagation

Society International Symposium,Volume 2, Page(s):272 –

275.

[6] Jui-Han Lu , May 2003,”

Bandwidth enhancement design of

single-layer slotted circular microstrip

antennas” IEEE Transactions on Antennas and Propagation, Volume:

51, Issue:5, page(s): 1126- 1129

[7] J. A. Ansari ., R. B. Ram, May

2008,” Broadband stacked U-slot microstrip patch antenna”, Progress In

Electromagnetics Research Letters,

Vol. 4, Page(s):17–24. [8] Baharudin Yatim ., Mohammed N.

Shakib., Mohammad T. Isla.,

Norbahiah Misran , 2009,” Design Analysis of a Slotted Microstrip

Antenna for Wireless Communication”

World Academy of Science,

Engineering and Technology,Page: 49 . [9] Balanis constantine,Antenna theory,

wiley-india, 2008.

Abstract— Finite impulse response (FIR) filter is the key

functional block in the field of digital signal processing. But in the VHDL implementation, the hardware complexity of the FIR

filter is directly proportional to the tap length and the bit-width of input signal. But in Digital signal processing (DSP) is the fastest growing technology this century and, therefore, it poses tremendous challenges to the engineering community. Faster additions and multiplications are of extreme importance in DSP for convolution, discrete Fourier transforms digital filters, etc. The core computing process is always a multiplication routine; therefore, DSP engineers are

constantly looking for new algorithms and hardware to implement them. The whole of Vedic mathematics is based on 16 sutras (word formulae) and manifests a unified structure of mathematics. The exploration of Vedic algorithms in the DSP domain may prove to be extremely advantageous. Engineering institutions now seek to incorporate research-based studies in Vedic mathematics for its applications in various engineering processes. Further research prospects may include the design

and development of a Vedic DSP chip using VHDL technology.

Keywords—VHDL, FPGA, FIR, CAD

I. INTRODUCTION

A finite impulse response (FIR) filter is a type of a

digital filter. The impulse response is finite because it

settles to zero in a finite number of sample intervals.

This is in contrast to infinite impulse response (IIR)

filters, which have internal feedback and may continue

to respond indefinitely. The impulse response of an Nth-

order FIR filter lasts for N+ 1 sample, and then dies to

zero. Firstly Time Domain input data is transformed into the Frequency Domain using a Fast Fourier Transform

(FFT), filtered directly in the Frequency Domain by

passing or amplifying desired frequencies and zeroing

out unwanted ones, and inverse transformed back into

Arun Bhatia is with the Department of Electronics and

Communication Engineering, Mharishi Markandeshwar Engineering

College,Mullana,Ambala, Haryana, India (e-mail:

[email protected] Contact No: 09996820693).

the Time Domain using an Inverse Fast Fourier Transform (IFFT). The outputs of the IFFT are then

concatenated together if necessary as shown in the fig 1.

[1]

Figure 1 Block diagram Of Basic filter

FIR filters have some basic properties:

FIR Filters are inherently stable. This is

due to the fact that all the poles are located at the origin and thus are located within the

unit circle.

They require no feedback. This means that

any rounding errors are not compounded

by summed iterations. The same relative

error occurs in each calculation. This also

makes implementation simpler.

They can easily be designed to be linear

phase by making the coefficient sequence symmetric; linear phase, or phase change

proportional to frequency, corresponds to

equal delay at all frequencies.

II. INTRODUCTION

TO VEDIC MATHEMATICS

Vedic mathematics is the name given to the system of

mathematics to be precise, a unique technique of

calculations based on simple rules and principles with

which any mathematical problem can be solved it

arithmetic, algebra, geometry or trigonometry. The

system is based on 16 Vedic sutras or aphorisms, which

Arun Bhatia, Student, Electronics & Communication Engineering Department Of M.M

Univerisity,Mullana(AMBALA). E-Mail Id: - [email protected] OR

[email protected] Contact NO:09996820693

Implementation of FIR filter using VEDIC Mathematics in VHDL

Proceedings of the Third National Conference on RTICT 2010Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

427

are actually word formulae describing natural ways of

solving a whole range of mathematical problems. Vedic

mathematics was rediscovered from the ancient Indian

scriptures between 1911 and 1918 by Sri Bharati

Krishna Tirthaji (1884-1960), a scholar of Sanskrit,

mathematics, Vedic mathematics is the name given to the ancient system of mathematics, or, to be precise, a

unique technique of calculations based on simple rules

and principles with which any mathematical problem

can be solved – be it arithmetic, algebra, geometry or

trigonometry. The system is based on 16 Vedic sutras or

aphorisms, which are actually word formulae describing

natural ways of solving a whole range of mathematical

problems. Vedic mathematics was rediscovered from the

ancient Indian scriptures between 1911 and 1918 by Sri

Bharati Krishna Tirthaji (1884-1960), a scholar of

Sanskrit, mathematics, renewed interest in Vedic

mathematics, and scholars and teachers in India started taking it seriously. According to Mahesh Yogi, The

sutras of Vedic Mathematics are the software for the

cosmic computer that runs this universe. A great deal of

research is also being carried out on how to develop

more powerful and easy applications of the Vedic sutras

in geometry, calculus and computing. Conventional

mathematics is an integral part of engineering education

since most engineering system designs are based on

various mathematical approaches. All the leading

manufacturers of microprocessors have developed their

architectures to be suitable for conventional binary arithmetic methods. The need for faster processing

speed is continuously driving major improvements in

processor technologies, as well as the search for new

algorithms. The Vedic mathematics approach is totally

different and considered very close to the way a human

mind works. [2]

III. FPGA IMPLIMENTATION

OF FIR FILTERS

FPGAs are being used for a variety of computationally

intensive applications, mainly in Digital Signal

Processing (DSP) and communications [3]. Due to rapid

increases in the technology, current generation of

FPGAs contain a very high number of Configurable

Logic Blocks (CLBs), and are becoming more feasible

for implementing a wide range of applications. The high non-recurring engineering (NRE) costs and long

development time for ASICs are making FPGAs more

attractive for application specific DSP solutions. DSP

functions such as FIR filters and transforms are used in a

number of applications such as communication and

multimedia. These functions are major determinants of

the performance and power consumption of the whole

system. Therefore it is important to have good tools for

optimizing these functions. Equation (1) represents the

output of an L tap FIR filter, which is the convolution of

the latest L input samples. L is the number of

coefficients h(k) of the filter, and x(n) represents the input time series.

y[n] = Σ h[k] x[n-k] k= 0, 1, ..., L-1 --- (1)

The conventional tapped delay line realization of this

inner product is shown in Figure 2. This implementation

translates to L multiplications and L-1 additions per

sample to compute the result. This can be implemented

using a single Multiply Accumulate (MAC) engine, but

it would require L MAC cycles, before the next input sample can be processed. Using a parallel

implementation with L MACs can speed up the

performance L times. A general purpose multiplier

occupies a large area on FPGAs. Since all the

multiplications are with constants, the full flexibility of

a general purpose multiplier is not required, and the area

can be vastly reduced using techniques developed for

constant multiplication. Though most of the current

generation FPGAs have embedded multipliers to handle

these multiplications, the number of these multipliers is

typically limited. Furthermore, the size of these multipliers is limited to only 18 bits, which limits the

precision of the computations for high speed

requirements. The ideal implementation would involve a

sharing of the Combinational Logic Blocks (CLBs) and

these multipliers. In this paper, we present a technique

that is better than conventional techniques for

implementation on the CLBs.

Figure 2. A MAC FIR filter block diagram

The coefficients in most of DSP applications for the

multiply accumulate operation are constants. The partial

products are obtained by multiplying the coefficients by

multiplying one bit of data at a time in AND operation. These partial products should be added and the result

depends only on the outputs of the input shift registers.

The AND functions and adders can be replaced by Look

Up Tables (LUTs) that gives the partial product. This is

shown in Figure 3.

428

Figure 3. A serial DA FIR filter block diagram

Input sequence is fed into the shift register at the input sample rate. The serial output presented to the RAM

based shift registers (registers are not shown in Figure

for simplicity) at the bit clock rate which is n+1 times (n

is number of bits in a data input sample) the sample rate.

The RAM based shift register stores the data in a

particular address. The outputs of registered LUTs are

added and loaded to the scaling accumulator from LSB

to MSB and the result which is the filter output will be

accumulated over the time. For an n bit input, n+1 clock

cycles are needed for a symmetrical filter to generate the

output. In conventional MAC method with a limited number of MAC engines, as the filter length is

increased, the system sample rate is decreased. This is

not the case with serial DA architectures since the filter

sample rate is decoupled from the filter length. As the

filter length is increased, the throughput is maintained

but more logic resources are consumed. Though the

serial DA architecture is efficient by construction, its

performance is limited by the fact that the next input

sample can be processed only after every bit of the

current input samples is processed. Each bit of the

current input samples takes one clock cycle to process.

Therefore, if the input bit width is 12, then a new input can be sampled every 12 clock cycles. The performance

of the circuit can be improved by modifying the

architecture to a parallel architecture which processes

the data bits in groups. Figure 4 shows the block

diagram of a 2 bit parallel DA FIR filter. The tradeoff

here is performance for area since increasing the number

of bits sampled has a significant effect on resource

utilization on FPGA. For instance, doubling the number

of bits sampled, doubles the throughput and results in

the half the number of clock cycles. This change doubles

the number of LUTs as well as the size of the scaling accumulator. The number of bits being processed can be

increased to its maximum size which is the input length

n. This gives the maximum throughput to the filter. For

a fully parallel implementation of the DA filter (PDA),

the number of LUTs required would be enormous. In

this work we show an alternative to the PDA method for

implementing high speed FIR filters that consumes

significantly lesser area and power.

Figure 4. A 2 bit parallel DA FIR filter block diagram

A popular technique for implementing the transposed

form of FIR filters is the use of a multiplier block,

instead of using multipliers for each constant as shown

in Figure 5. The multiplications with the set of constants

hk are replaced by an optimized set of additions and

shift operations, involving computation sharing. Further

optimization can be done by factorizing the expression

and finding common subexpressions. The performance

of this filter architecture is limited by the latency of the

biggest adder and is the same as that of the PDA.

Common sub expression elimination algorithm that minimizes the number of adders as well as the number

of latches. [4-7]

Figure 5. Replacing constant multiplication by multiplier block

Multiplications with constants have to be performed in many signal processing and communication applications

such as FIR filters, audio, video and image processing.

Using a general purpose multiplier for performing

constant multiplication is expensive in terms of area.

Though embedded multipliers are very popular in

modern FPGAs, the size (18x18) and the number of

these multipliers is a limitation. The partial product

429

Constant Coefficient Multiplier called KCM in is a

popular approach for performing constant

multiplications in FPGAs. The partial products are

generated by table look-ups and are summed up using

fast adders. This KCM is about one quarter the size of a

fully functional multiplier [8]. Distributed Arithmetic has been used extensively for implementing FIR filters

on FPGAs. In both Serial and Parallel Distributed

Arithmetic designs are used to implement a 16-tap FIR

filter. The author notes the superior performance

achieved by using PDA over using a traditional DSP

processor. The XilinxTM CORE Generator has a highly

parameterizable, optimized filter core for implementing

digital FIR filters [9]. It generates synthesized core that

targeting a wide range of Xilinx devices. The Serial

Distributed Arithmetic (SDA) architecture is very area

efficient for serial implementations of filters. The fully

Parallel Distributed Arithmetic (PDA) structure can be used for fastest sample rates, but it suffers from

excessive area. In FIR filters are implemented using the

Add and Shift method. Canonical Signed Digit (CSD)

encoding is used for the coefficients to minimize the

number of additions. The critical path of the filter can be

reduced considerably by using registered adders, at

minimal increase in area. This is because of the

availability of a D flip-flop at the output of the LUT.

This is illustrated in Figure 6, which compares a non-

registered and a registered output. The registered adders

and the performance of filters are limited by the latency of the biggest adder. Common sub expression

elimination is used for reducing the number of adders.

Furthermore, the use of sub expression elimination

increases the number of latches that are required to

make the filter fully parallel sub expression elimination

algorithm considers both the cost of the latches as well

as the cost of an adder.[10]

Figure 6. Registered adder at no additional cost

IV. FILTER ARCHITECTURE Filter architecture on the transposed form of the FIR

filter as shown in Figure 2. The filter can be divided into

two main parts, the multiplier block and the delay block,

and is illustrated in Figure 5. In the multiplier block, the

current input variable x[n] is multiplied by all the

coefficients of the filter to produce the y outputs. These

y outputs are then delayed and added in the delay block

to produce the filter output y[n].The constant multiplications are decomposed into registered additions

and hardwire shifts. The additions are performed using

two input adders, which are arranged in the fastest tree

structure. Performing subexpression elimination can

sometimes increase the number of registers

substantially, and the overall area could possibly

increase. Consider the two expressions F1 and F2 which

could be part of the multiplier block.

F1 = A + B + C + D—(2)

F2 = A + B + C + E--- (3)

Figure 7 shows the original un optimized expression trees. Both the expressions have a minimum critical path

of two addition cycles. These expressions require a total

of six registered adders for the fastest implementation,

and no extra registers are required. From the expressions

we can see that the computation A + B + C is common

to both the expressions.

Figure 7. Unoptimized expression trees

If we extract this subexpression, we get the structure

shown in Figure 8. Since both D and E need to wait for

two addition cycles to be added to (A + B + C), we need

to use two registers each for D and E, such that new

values for A,B,C,D and E can be read in at each clock

cycle. Assuming that the cost of an adder and a register with the same bit width are the same, the structure

shown in Figure 7 occupies more area than the one

shown in Figure 6. A more careful subexpression

elimination algorithm would only extract the common

subexpression A + B (or A+C or B + C). The number of

adders is decreased by one from the original, and no

additional registers are added. This is illustrated in

Figure 9. The algorithm for performing this kind of

optimization is described in the next section. [11]

Figure 8. Extracting common expression (A + B + C)

430

Figure 9. Extracting common subexpression (A+B)

V. OPTIMIZATION ALGORITHM

Optimization is to reduce the area of the multiplier block

by reducing the number of adders and any additional

registers required for the fastest implementation of the

FIR filter. In the common sub expression elimination

methods polynomial transformation of constant

multiplications. Given a representation for the constant

C, and the variable X, the multiplication C*X can be

represented as a summation of terms denoting the

decomposition of the multiplication into shifts and additions as

C*X= Σ+XL—(4)

The terms can be either positive or negative when the

constants are represented using signed digit

representations such as the Canonical Signed Digit

(CSD) representation. The exponent of L represents the

magnitude of the left shift. In this method minimum

number of registers required the fastest tree structure

with three addition steps, and one register to

synchronize the intermediate values, such that new values for A,B,C,D,E,F can be read in every clock cycle.

This is illustrated in Figure 10. [12]

Figure 10. Calculating registers required for fastest evaluation

VI. CAD SYSTEM FOR FIR FILTERS

To design FIR filters with the CAD tool that generates various optimum architectures for FIR filters needed in

different applications in terms of speed, power and area

in the IC implementation. CAD system for filter

generation in VHDL. The CAD methodology for this

filter compiler tool centers around a VHDL-based

functional compiler which accepts functional parameters

and automatically generates optimized IC

implementation for FIR filters. Previous functional compilers are extremely rigid in architecture generation

and technology retargeting because of their dependence

on specific low-level layout/floor plan generation tools.

In contrast, our approach generates a VHDL solution

which feeds into logic synthesis tools and achieves

technology independence. The design flow achieved

with the functional compiler is shown in Figure 11. This

design flow leverages advances made in high level

synthesis tools and provides multiple levels of

simulations from VHDL to layout. User specifies

parameters for a design to the functional compiler which

produces an optimum architecture in VHDL format. Logic synthesis tools accept the generated VHDL and

target the design to different IC technologies. Design

verification may be performed at system, VHDL and

layout levels. [13]

Figure 11 CAD tool design flow with Functional Compilers

VII. FIR FILTER ARCHITECTURE

DERIVATION An FIR filter implements the following non-recursive

difference equation:

N-1

Y[n] = ∑ h[k]x [n-k] --- (5)

k = 0

where h[k] are the coefficients of the FIR filter and N is

the order of the filter. This algorithm can be expressed

as a DG shown in Figure 12. The DG of the FIR shows

the dependence of the computations in the FIR filtering

algorithm. The diagonal arcs in the DG intersect the DG

nodes needed to compute a given output sample. The

horizontal arcs pass the current input sample to DG nodes where it is required for computation and the

vertical arcs pass the coefficients to the appropriate DG

nodes. A detailed DG node consists of an adder and a

multiplier. Each node in the DG adds a partial sum

along its input diagonal to the product of the coefficient

431

and the current input sample and outputs the resulting

sum along its output diagonal. Based on different

mappings of the DG onto an array structure, a number of

SFGs (signal $ow graphs) can be derived from the DG.

An SFG is closer to the architecturehardware

implementation of the algorithm and is obtained by PE (processor element) allocation followed by operation

scheduling [14].

Figure.12 Transpose-form architecture derivation using the DG-SFG

technique

A. Architecture Derivation for Decimation Filters

The decimation FIR filter is described by the following

equation:

N-1

y[m] = ∑ h[k]x[mD-k]--- (6)

k = O

where D is the decimation factor. Two approaches have been used to implement decimation FIRS. The first

approach uses a decimator immediately following a full

fledged FIR filters. The second approach computes

decimation FIR output as sum of outputs of D polyphase

filters. These implementations do not exploit decimation

to reduce complexity and as a result can be highly non-

optimal for high decimation factors.

B. Impact of decimation on the DG of an FIR

filter

Figure. 13 show the effect of decimation on the DG of

an FIR filter. The DG nodes are shown shaded on the

diagonal arcs which produce the decimated output

samples. For an FIR filter with N = 4 and D = 2 (Figure

13), computations for every alternate output sample

need to be performed. For D = 2, only half of the DG nodes, which are shown shaded, need to be computed in

each cycle. An alternative interpretation is the division

of the DG into 2 sections as shown in Figure.13. Only

one of the 2 DG nodes from each section is computed

every clock cycle This can be generalized to any

arbitrary filter order, N and decimation factor D. For a

given N and D, the number of nodes to be computed per

clock cycle is [N / D] , where (r) represents the largest

integer contained in r. In other words, decimation

divides the DG into [N/D] each containing D DG nodes,

of which only one is computed every cycle.[15]

Figure.13 Effect of decimation on the DG of the FIR filter

VIII. CONCLUSION

FPGA architecture is proposed that is tailored for FIR

filters realization. The novelty of this architecture lies in

several fields that are associated with the

interconnection scheme of the 169 CALUs within a fine-grained architecture and the way the coefficients are

retrieved from the same output, using paths with

different delay time based on a certain configuration.

Additionally, by using the POF technique and designing

heterogeneous, configurable blocks, the power

calculation of our design has been proved to be

competitive compared with general purpose FPGAs,

although the associated built-in flexibility that implies

high degree of robustness and the additional logic

circuitry, which is introduced in order the system to

cope efficiently with overflow. The primary results,

which have been obtained, are related with the hardware realization and power consumption of this architecture.

Finally, our future research is going to be focused on the

efficient realization of medium and high order FIR

filters on this architecture by using evolutionary

technique.

432

IX. REFRENCES

[1]. http://en.wikipedia.org/wiki/Finite_impulse_response

[2].Jagadguru Swami Sri Bharati Krisna Tirthaji Maharaja, Vedic

Mathematics: Sixteen Simple Mathematical Formulae from the Veda.

Delhi (1965).

[3] K.D.Underwood and K.S.Hemmert, "Closing the Gap: CPU and

FPGA Trends in Sustainable Floating- Point BLAS Performance",

International Symposium on Field-Programmable Custom Computing

Machines, California, USA, 2004.

[4] Y.Meng, A.P.Brown, R.A.Iltis, T.Sherwood, H.Lee, and

R.Kastner, "MP Core: Algorithm and Design Techniques for Efficient

Channel Estimation in Wireless Applications", Design Automation

Conference (DAC), Anaheim, CA, 2005.

[5] B. L. Hutchings and B. E. Nelson, "Gigaop DSP on FPGA",

Acoustics, Speech, and Signal Processing, 2001. Proceedings.

(ICASSP '01). 2001 IEEE International Conference on, 2001.

[6] A.Alsolaim, J.Becker, M.Glesner, and J.Starzyk, "Architecture and

Application of a Dynamically Reconfigurable Hardware Array for

Future Mobile Communication Systems", International Symposium on

Field Programmable Custom Computing

[7] S.J.Melnikoff, S.F.Quigley, and M.J.Russell, "Implementing a

Simple Continuous Speech Recognition System on an FPGA",

International Symposium on Field-Programmable Custom Computing

Machines (FCCM), 2002.

[8] K.Chapman, "Constant Coefficient Multipliers for the XC4000E,"

Xilinx Technical Report 1996.

[9] "Distributed Arithmetic FIR Filter v9.0," Xilinx Product

Specification 2004.

[10] M. Yamada and A. Nishihara, "High-speed FIR digital filter with

CSD coefficients implemented on FPGA", Design Automation

Conference, 2001. Proceedings of the ASP-DAC 2001. Asia and South

Pacific, 2001.

[11] K. Wiatr and E. Jamro, "Constant coefficient multiplication in

FPGA structures", Euromicro Conference, 2000. Proceedings of the

26th, 2000.

[12] G.R.Goslin, "A Guide to Using Field Programmable Gate Arrays

(FPGAs) for Application-Specific Digital Signal Processing

Performance," Xilinx Application Note, San Jose 1995.

[13]R.Jain, P.Yang, and T. Yoshino, FIRGEN: A Computer-Aided

Design System for High Performance FIR Integrated Circuits, IEEE

trans. on Signal Processing, vol. 39, no. 7, pp. 1655-1658

[14]S.Y. Kung, VLSIArray Processors, Prentice Hall, 1988 J.

Laskowski, A Silicon Compiler for Linear-Phase FIR Digital jilters,

M.S. Thesis, UCLA, 1991

[15]Thu-Ji Lin and Henry Samueli, A 200-MHz CMOS x/sin(x)

Digital filter for Compensating D/A Comerter Frequency Response

Distorrion, IEEE JSSC, vol. 26, no. 9, Sept. 1991 Robert Hawley, A

Silicon Compiler for High-speed CMOS Multirate FIR Digital Filters,

M.S. Thesis, UCLA, 1991

X. BIOGRAPHIES

Arun Bhatia born at 4, Dayal Bagh, Bharat Nagar Ambala

Cantt (HARYANA) on June 11, 1984 received B. Tech. in

(Instrumentation and Control Engineering) from Haryana

Engineering College Jagadhri, affiliated to Krukshetra

University krukshetra in the year 2007. Pursuing M.TECH in

the field of Electronics & Communication Engg from M.M

Engineering College Affiliated to M.M University Mullana

(AMBALA).

433

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

434

TWO MICROPHONE ENHANCEMENT OF

REVERBERANT SPEECH BY SETA-TCLMS

ALGORITHM

Authors: Manikandan A1, Ashok MB2, Johnny M3, Kashi Vishwanath RG4

1-Lecturer, KLN College of Information Technology

2, 3, 4-Students, KLN College of Information Technology

Mail id: [email protected]

Abstract- In this letter we propose a novel approach for

implementing two microphone enhancement of

reverberant speech using Selective Tap filtering. Our

approach involves capturing the reverberant speech using

two microphones in the proposed SIMO model and then to

restore the corrupted speech by computing essential filter

tap coefficients which impacts on the corrupted

reverberant speech and updating the tap coefficients

further on each subsequent iteration based on SETA-

TCLMS. Experimental results show that the results are in

correspondence with theoretical results and better than its

full update counterpart (MCLMS algorithm).

Index Terms – Multi Channel Least Mean Square

(MCLMS), Selective Tap Two Channel Least Mean Square

(SETA-TCLMS), Single Input Multiple Output model

(SIMO), Perceptual Evaluation of Speech Quality (PESQ)

I. INTRODUCTION

The speech is subjected to distortion by acoustic

reverberation especially when a speaker is distant from a

microphone such as in hands-free telephony environment and teleconferencing. Reverberation is harmful to speech

intelligibility since it blurs temporal and spectral cues,

flattens formant transitions, reduces amplitude

modulations associated with the fundamental frequency

of speech and increases low-frequency energy, which in

turn results in masking of higher speech frequencies [1].

In that case, dereverberation techniques should be applied

to corrupted speech to restore clean speech. Thus,

dereverberation has become an important topic because

of its high potential for wide applications including

speech enhancement and recognition. Since clean speech

and reverberant environments are usually unknown, blind dereverberation methods have been proposed to

estimate an inverse filter of the reverberation from

microphone speech signal.

In this paper, we derive a novel blind single-

input two-output reverberant speech enhancement

strategy, which stems from the multichannel least-mean-

square (MCLMS) algorithm. The proposed method uses

second-order statistics to identify the acoustic paths in

the time-domain and relies on a novel selective-tap

criterion to update only a subset of the total number of filter coefficients in every iteration. Therefore, it

substantially reduces computational requirements with

only minimal degradation in dereverberation

performance. The potential of the proposed low-

complexity algorithm is verified and assessed through

numerical simulations in realistic acoustical scenarios.

II. PROBLEM FORMULATION

Consider the paradigm shown in Fig. (A)

Where the speech is picked up by two

microphones of a hearing aid device. Let )(ks

represents clean speech signal, )(1 kh and

)(2 kh represents the impulse response of the two

acoustic paths modeled using finite impulse

response (FIR) filters. And )(1 kx and )(2 kx be

the reverberant signals captured by the two microphones of the device such that one

microphone is directional front microphone and

the other is Omni directional microphone

respectively. In the noiseless two-microphone

scenario, we exploit the correlation between the

output signals of each microphone

)()()( kskhkx ii and )()()( kskhkx jj (1)

Where * denotes linear convolution and i, j=1, 2. From (1) and for all i ≠ j, follows that

)()()()()( kskhkhkxkh ijij

)()( kxkh ji (2)

435

Fig (A) -The above mentioned model is an SIMO system, i.e., Single input with Two Microphone system

III. ALGORITHM

A. MULTICHANNEL Least Mean Square

ALGORITHM (MCLMS)

As shown in [10], an intuitive way to ’blindly’ calculate

the unknown acoustic paths is to minimize a cost

function that penalizes correlated output signals between

the th and jth sensors, such that

1

1 1

2)1(minarg)1(

M

i

M

ijijh kkJ e (3)

After rewriting (2) in vector notation, the error function

eij (k+1) in (3) can be further expanded to

)()1()()1()1(~~

khkxkhkxke i

T

jjT

iij (4)

Defined for all i, j = 1, 2 and i ≠ j, where ( . )T denotes

vector transpose.

Then

Tiiii Lkxkxkxkx )1(),...,1(),()( (5)

is the corrupted (reverberant) speech picked up by the ith

microphone . Accordingly, the update equation of the

time-domain multichannel least-mean-square

(MCLMS) algorithm is given by [10]

)1()()1(~~

kJkhkh (6)

)(

)1()1(

~

kh

kJkJ

(7)

2~

~~~~

)(

)()1()()1(2

kh

khkJkhkR x

(8)

Where 0< µ < 1 is the learning parameter controlling the

rate of convergence and speed of adaptation.

The channel coefficient vectors are such that

TTTkhkhkh )](),([)( 2

~

1

~~

(9)

The autocorrelation of the microphone signals is equal to

)()(

)()()(

1121

1222

~~

~~~

kRkR

kRkRkR

xxxx

xxxxx (10)

with )]()([)(~

kxkxEkRT

jxxx ji valid for all i,j =1,2

B. SELECTIVE-TAP TWO-CHANNEL Least Mean

Square ALGORITHM (SETA-TCLMS)

We formulate the new selective-tap approach

based on the two-microphone configuration

depicted in Fig(A).By inspecting (6)-(8) we can

see that as the adaptive algorithm approaches

convergence, the cost function J(k+1) diminishes and its gradient with respect to estimated filter

coefficients which becomes

)()1(2)(~~~

khkRkh x (11)

After removing the unit-norm constraint from

(8). From the above equation, it readily becomes

evident that the convergence behavior of the

MCLMS algorithm depends solely on the

magnitude (element-wise) of the

autocorrelation matrix estimated at each iteration . The computational complexity and

slow convergence of the MCLMS algorithm can

be therefore reduced substantially by employing

a simple tap-selection criterion to update only M

out of L coefficients containing the largest

values of the autocorrelation matrix [11]. The

subset of the filter coefficients updated at

iteration can be determined from the M x M

matrix Q(k), which is coined the tap-selection

matrix

)1(000

0

)1(0

0...0)0(

)(

Mq

q

q

kQ

ij

ij

ij

(12)

Where each element is given by

T

ijijijij Mkqkqkqkq )]1(),...,1(),([)(

(13)

Such that

otherwise

RMlkRlkq xxij

||max|)(|

,0

,1)(

~~

(14)

Where in a two-channel setup (12)-(14) are defined for

i,j =1, 2 and for all lags l = 0,1,2 …. M-1 and the

operator | . | denotes absolute value. In order to calculate

the different filter coefficients that are to be updated at

436

different time instants, a fast sorting routine (e.g., see

SORTLINE [12]) is executed at every iteration. After

sorting, each block of the tap-selection matrix contains

coefficients equal to one in the positions (or indices)

calculated from (14) and zeros elsewhere, such that M <

L , with M = tr | Q(k)| where tr | . | denotes the sum of the diagonal elements of matrix Q(k). To update only M

taps of the equalizer, we write the selective-tap two-

channel least-mean-square (SETA-TCLMS) algorithm as

follows:

)()1(2)()1(~~~~

khlkRkhkh lxll (15)

Where the update is carried out with learning rate λ

only if L corresponds to one of the first M maxima of

autocorrelation, whereas when qi,j ( k-l ) = 0, then (15) becomes

)()1(~~

khkh ll (16)

Note that for M = L, the SETA-TCLMS reduces to the

full-update algorithm described in (6)-(8).

IV. EXPERIMENTAL RESULTS

The performance of the Algorithm is evaluated using

sentences which are phonetically balanced each

consisting of seven to twelve words. These speech

sources are 10s in duration and sampled at 8 kHz.

This algorithm is executed with L=128, whereas the tap

selection length is set to 128, 64, 32.also these

phonetically balanced sentences are pre-recorded and

collected in an IEEE database.

A. PERFORMANCE EVALUATION:

NORMALISED PROJECTION MISALIGNMENT

(NPM):

Since, in our experimental setup the acoustic channel

impulse responses are known a priori the channel

identification accuracy is calculated using the

normalized projection misalignment (NPM) metric.

h

kkNPM

)(log20)( 10

(17)

With h = [h1T, h2

T]T and the projection misalignment ε(k)

)(

)()(

)()(

~

~~

~

kh

khkh

khhhk

T

T

(18)

Where for perfectly identified acoustic paths ε (k) tends

to 0.

Learning Curve as per SETA-TCLMS for L=128 and

M=128

0 50 100 150 200 250 300-0.2

0

0.2

0.4

0.6

0.8

1

1.2

Learning curve as per SETA-TCLMS for L=128 and

M=64

0 20 40 60 80 100 120 140 160 180 200-0.2

0

0.2

0.4

0.6

0.8

1

1.2

Learning Curve as per SETA-TCLMS for L=128 and

M=32

0 20 40 60 80 100 120 140 160 180-0.2

0

0.2

0.4

0.6

0.8

1

1.2

TABLE-I

NPM computed after Convergence of SETA-TCLMS

algorithm

L | M Speaker NPM(dB) PESQ

437

128|128

A

B

C

-11.47

-10.21

-14.32

4.22

4.09

4.35

128|64

A

B C

-10.83

- 9.55 -12.09

3.83

3.62 3.92

128|32

A

B

C

- 8.26

- 9.21

-8.07

3.29

3.48

3.12

Where A, B, C are three different speakers pronouncing

phonetically balanced sentences as per the IEEE

subcommittee database.

PERCEPTUAL EVALUATION OF SPEECH QUALITY

(PESQ):

Although the NPM metric can measure channel

identification accuracy reasonably well, it might not

always reflect the output speech quality. For that reason,

we also assess the performance of the proposed

algorithm using the PESQ [15]. The PESQ employs a

sensory model to compare the original (unprocessed) with

the processed signal, which is the output from the

Dereverberation algorithm, by relying on a perceptual model of the human auditory system. In the context of

additive noise suppression, PESQ scores have been

shown to exhibit a high Pearson’s correlation

coefficient of ρ = 0.92 with subjective listening quality

tests [16]. The PESQ measures the subjective

assessment quality of the dereverberated speech rated as a

value between 1 and 5 according to the five grade mean

opinion score (MOS) scale. Here we use a modified

PESQ measure [16], referred to as mPESQ, with

parameters optimized towards assessing speech signal

distortion, calculated as a linear combination of the average disturbance value Dind and the average

asymmetrical disturbance values Aind [15], [16].

indAaaamPESQ 210 (19)

Such that

191.0,959.4 10 aa and 006.02 a (20)

By definition, a high value of mPESQ indicates low

speech signal distortion, whereas a low value suggests

high distortion with considerable degradation present.

In effect, the mPESQ score is inversely proportional to

reverberation time and is expected to increase as

reverberant energy decreases.

A. DISCUSSION

Table I contrasts the performance of the SETA-

TCLMS algorithm relative to the performance of its

full-update counterpart (MCLMS algorithm). As it can

be seen in Table I, the full-update TCLMS yields the

best NPM performance and the highest mPESQ scores.

Still, the degree of Dereverberation remains largely

unchanged when updating with the SETA-TCLMS using only M = 128 filter coefficients. In fact, even

when employing just M = 32taps, which accounts for a

75% reduction in the total equalizer length (with a

processing delay of just 64 ms at 8 kHz) the algorithm

can estimate the room impulse responses with

reasonable accuracy.In terms of overall speech quality

and speech distortion, the mPESQ score for the

reverberant (unprocessed) speech signals, averaged

across all five speakers and in both microphones, is

equal to 2.72, which suggests that a relatively high

amount of degradation is present in the microphone

inputs. In contrast, after processing the two-microphone reverberant input signals with the SETA-

TCLMS algorithm, the average mPESQ scores

increase to 4.17, 3.62 and 3.40 when using, M=128,

64 and 32 taps, respectively. The estimated mPESQ

values suggest that the proposed SETA-TCLMS

algorithm can improve the speech quality of the

microphone signals considerably, while keeping signal

distortion to a minimum.

V. CONCLUSION

We have developed a selective-tap blind identification

scheme for reverberant speech enhancement using a

two-microphone model. Numerical experiments

carried out on MATLAB with speech signals in a

reverberant setup, indicate that the proposed two-

channel Dereverberation technique is capable of

equalizing fairly long acoustic echo paths with

sufficient accuracy and no degradation. The

proposed adaptive algorithm exhibits a low

computational overhead and therefore is amenable to

real-time implementation in portable devices like

Mobile phones, Hearing aids.

REFERENCES

[1] A. K. Nabelek and J. M. Picket, “Monaural and

binaural speech perception through hearing aids under

noise and reverberation with normal and hearing-

impaired listeners,” J. Speech Hear. Res., vol. 17, pp.

724-739, 1974.

[2] S. Haykin, Ed., Unsupervised Adaptive Filtering

New York, Wiley, 2000, vol. II, Blind Deconvolution. [3] S. Gannot and M. Moonen, “Subspace methods for

multi-microphone speech dereverberation,” EURASIP

J. Appl. Signal Process., vol. 11, pp. 1074-1090, 2003.

[4] T. Nakatani and M. Miyoshi, “Blind dereverberation

of single channel speech signal based on harmonic

structure,” in Proc. ICASSP, 2003, vol. 1, pp. 92-95.

438

[5] M. Wu and D. L. Wang, “A two-stage algorithm for

one-microphone reverberant speech enhancement,”

IEEE Trans. Audio, Speech, Lang. Process., vol. 14, no.

3, pp. 774-784, May 2006.

[6] J.-H. Lee, S.-H. Oh, and S.-Y. Lee, “Binaural semi-blind dereverberation of noisy convoluted speech

signals,” Neurocomputing, vol. 72, pp. 636-642, 2008.

[7] E.A.P.Habets,“Multi-channel speech dereverberation

based on a statistical model of late reverberation,” in

Proc. ICASSP, 2005, vol. 4, pp.173-176.

[8] H. W. Löllmann and P. Vary, “A blind speech

enhancement algorithm for the suppression of late

reverberation and noise,” in Proc. ICASSP, 2009, pp.

3989-3992.

[9] T. Yoshioka, T. Nakatani, T. Hikichi, and M. Miyoshi,

“Maximum like lihood approach to speech enhancement

for noisy reverberant signals,” in Proc. ICASSP, 2008,

pp. 4585-4588.

[10] Y. Huang and J. Benesty, “Adaptive multi-channel

least mean square and Newton algorithms for blind

channel Identification,”Signal Processing, vol. 82, pp.

1127-1138, 2002.

[11] T. Aboulnasr and K. Mayyas, “Complexity

reduction of the NLMS algorithm via selective

coefficient update,” IEEE Trans. Signal Processing., vol.

47, pp. 1421-1424, 1999.

[12] I. Pitas, “Fast algorithms for running ordering and max/min calculation,” IEEE Trans. Circuits Systems.,

vol. 36, no. 6, pp. 795-804, Jun. 1989

[13] IEEE Subcommittee, “IEEE recommended

practice speech quality measurements”,” IEEE

Trans. Audio Electroacoustics., vol. 17, no. 3, pp.

225-246, Sep. 1969.

[14] B. G. Shinn-Cunningham, N. Kopco, and T. J.

Martin, “Localizing nearby sound sources in a

classroom: binaural room impulse responses,” J. Acoust. Soc. Amer., vol. 117, pp. 3100-3115, 2005.

[15] Perceptual Evaluation of Speech Quality

(PESQ), an Objective Method for End-To-End

Speech Quality Assessment of Narrow-Band

Telephone Networks and Speech Coders

ITU-T Recommendation, 2001, ITU-T

Recommendation P.862.

[16] Y. Hu and P. C. Loizou, “Evaluation of objective

quality measures for speech enhancement,” IEEE Trans.

Audio, Speech, Lang. Process., vol. 16, no. 1, pp. 229-238, Jan. 2008.

[17] J.Benesty , S.Makino and J.Chen , “Seperation and

dereverberation of speech signals with Multiple

Microphones, Springer series on Signals and

Communication Technology , ISBN 3-540-24039-X

Springer Berlin Heidelberg NewYork.

[18] Kostas Kokkinakis and Philipos C. Loizou , “ Selective Tap Blind signal Processing for Speech

Seperation ,” 31st Annual International Conference of

the IEEE EMBS Minneapolis , Minnesota , USA ,

September 2-6, 2009.

[19] John G Proakis, Dimitris G Manolakis, Digital

Signal Processing, Principles, Algorithms and

Applications, 4th edition, PHI Pvt Ltd. 2007

[20] John G Proakis, Masoud Salehi, Digital

Communications, 5thedition, McGraw Hill International

Edition 2008.

A Cooperative Communication method for noise

Elimination Based on LDPC Code

Abstract:

In telecommunication, intersymbol interference (ISI)

is a form of distortion of a signal in which one symbol interferes

with subsequent symbols. This is an unwanted phenomenon as

the previous symbols have similar effect as noise, thus making

the communication less reliable. ISI is usually caused by

multipath propagation or the inherent non-linear frequency

response of a channel causing successive symbols to "blur"

together. The presence of ISI in the system introduces errors in

the decision device at the receiver output. Therefore, in the

design of the transmitting and receiving filters, the objective is

to minimize the effects of ISI, and thereby deliver the digital

data to its destination with the smallest error rate possible.

Ways to fight intersymbol interference include adaptive

equalization and error correcting codes.

Index Terms—RA-Type LDPC codes, splitting, rate

compatibility, convergence, HARQ.

Introduction:

In the simplest case, transmitted messages consist of

strings of 0's and 1's, and errors introduced by the channel consist of

bit inversions: 0 to 1 and 1 to 0. The essential idea of forward error

control coding is to augment messages to produce codeword

containing deliberately introduced redundancy, or check bits. With

care, these check bits can be added in such a way that codeword are

sufficiently distinct from one another so that the transmitted

message can be correctly inferred at the receiver, even when some

bits in the codeword are corrupted or lost during transmission over

the channel.

In this thesis, we consider code design and decoding for a

family of error correction codes known as low-density parity-check

(LDPC) block codes. In the simplest form of a parity-check code, a

single parity-check equation provides for the detection, but not

correction, of a single bit inversion in a received codeword.

To permit correction of errors induced by channel noise,

additional parity checks can be added at the expense of a decrease in

the rate of transmission. Low-density parity-check codes are a special

case of such codes.

Drawback of Existing Parity-check codes

The simplest possible error detection scheme is the single

parity check, which involves the addition of a single extra bit to a

binary message. Whether this parity bit should be a 0 or a 1 depends

on whether even or odd parity is being used. In even parity, the

additional bit added to each message ensures an even number of 1s in

each transmitted codeword. For example, since the 7-bit ASCII code

for the letter S is 1010011, a parity bit is added as the eighth bit. If

even parity is being used, the value of the parity bit is 0 to form the

codeword 10100110.

More formally, for the 7-bit ASCII plus even parity code we

define a codeword c to have the following structure:

c = c1 c2 c3 c4 c5 c6 c7 c8

Where each ci is either 0 or 1, and every codeword satisfies the constraint

c1 c2 c3 c4 c5 c6 c7 c8 0

Here the symbol represents modulo-2 addition, which is

equal to 1 if the ordinary sum is odd and 0 if the ordinary sum is even.

Proceedings of the Third National Conference on RTICT 2010Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

439

While the inversion of a single bit due to channel noise can

easily be detected with a single parity check as this code is not

sufficiently powerful to indicate which bit, or bits, were inverted.

Moreover, since any even number of bit inversions produces a word

satisfying the constraint (1), any even numbers of errors go undetected

by this simple code.

Detecting more than a single bit error calls for increased

redundancy in the form of additional parity checks. To illustrate,

suppose we define a codeword c to have the following structure:

c = c1 c2 c3 c4 c5 c6 where each ci is either 0 or 1, and c is constrained by three parity-

check equations

In matrix form we have that c = [c1 c2 c3 c4 c5 c6] is a codeword if

and only if it satisfies the constraint

THc

where the parity-check matrix, H, contains the set of parity-check

equations which define the code. To generate the codeword for a given

message, the code constraints can be rewritten in the form

where bits c1, c2, and c3 contain the 3-bit message, and parity-check

bits c4, c5 and c6 are calculated from the message. Thus, for example,

the message 110 produces parity check bits

c4 = 1 1 = 0,

c5 = 1 0 = 1

and c6 = 1 1 0 = 0,

and hence the codeword 110010. The matrix G is the generator matrix

of the code. Substituting each of the 32 = 8 distinct messages c1c2c3

= 000, 001,…..111 into equation yields the following set of code

words:

440

The minimum distance of this code is 3, so a single bit error always

results in a word closer to the codeword which was sent than any other

codeword, and hence can always be corrected. In general, for a code

with minimum distance dmin, e bit errors can always be corrected by

choosing the closest codeword

whenever

min 1/ 2e d

where (x) is the largest integer that is at most x.

In this method Error correction by direct search is feasible only when

the number of distinct code words is small. For codes with thousands

of bits in a codeword, it becomes far too computationally expensive to

directly compare the received word with every codeword in the code

Introduction to LDPC: One of the causes of intersymbol interference is what is known as

multipath propagation in which a wireless signal from a transmitter

reaches the receiver via many different paths. The causes of this

include reflection (for instance, the signal may bounce off buildings),

refraction (such as through the foliage of a tree) and atmospheric

effects such as atmospheric ducting and ionospheric reflection. Since

all of these paths are different lengths - plus some of these effects will

also slow the signal down - this results in the different versions of the

signal arriving at different times.

000000 001011 010111 011100

100101 101110 110010 111001:

The reception of a word which is not in this set of code words can

be detected using the parity-check constraint equation THc .

Suppose for example that the word r = 101011 is received from the

channel. Substitution into equation THc gives

which is nonzero and so the word 101011 is not a codeword of our

code. To go further and correct the error requires that the decoder

determine the codeword most likely to have been sent. Since it is

reasonable to assume that a small number of errors is more likely to

occur than a large number, the required codeword is the one closest

in Hamming distance to the received word. By comparison of the

received word r = 101011 with each of the code words in above

matrix, the closest codeword is c = 001011, which is at Hamming

distance 1 from r.

441

This delay means that part or all of a given symbol will be

spread into the subsequent symbols, thereby interfering with the

correct detection of those symbols. Additionally, the various paths

often distort the amplitude and/or phase of the signal thereby causing

further interference with the received signal.

There are several techniques in telecommunication and data storage

that try to work around the problem of intersymbol interference.

One technique is to design symbols that are more robust

against intersymbol interference. Decreasing the symbol rate

(the "baud rate"), and keeping the data bit rate constant (by

coding more bits per symbol), reduces intersymbol

interference.

Other techniques try to compensate for intersymbol

interference. For example, hard drive manufacturers found

they could pack much more data on a disk platter when they

switched from Modified Frequency Modulation MFM to

Partial Response Maximum Likelihood (PRML). Even

though it's impossible to tell the difference between a "1" bit

and a "0" bit if you only look at the signal during that bit,

you can still tell them apart by looking at a cluster of bits at

once and figuring out which binary sequence,

Implementation of LDPC in error correction:

Low-density parity-check (LDPC) codes are a class of

linear block LDPC codes. The name comes from the characteristic of

their parity-check matrix which contains only a few 1’s in comparison

to the amount of 0’s. Their main advantage is that they provide a

performance which is very close to the capacity for a lot of different

channels and linear time complex algorithms for decoding.

Furthermore are they suited for implementations that make heavy use

of parallelism.

Representations for LDPC codes

Basically there are two different possibilities to represent

LDPC codes. Like all linear block codes they can be described via

matrices. The second possibility is a graphical representation.

(i) Matrix Representation

Lets look at an example for a low-density parity-check

matrix first. The matrix defined in below equation is a parity check

matrix with dimension n ×m for a (8, 4) code. We can now define two

numbers describing these matrixes. wr for the number of 1’s in each

row and wc for the columns. For a matrix to be called low-density the

two conditions wc n and wr m must be satisfied. In order to

do this, the parity check matrix should usually be very large.

442

(ii) Graphical Representation

Bit – flipping (BF) in LDPC

Bit-Flipping algorithm is presented, based on an initial hard

decision (0 or 1) assessment of each received bit. An essential part of

iterative decoding is the passing of messages between the nodes of the

Tanner graph of the code. For the bit-Flipping algorithm the messages

are simple: a bit node sends a message to each of the check nodes to

which it is connected declaring if it is a 1 or a 0, and each check node

sends a message to each of the bit nodes to which it is connected,

declaring whether the parity check is satisfied or not.

The sum-product algorithm for LDPC codes operates

similarly but with more complicated messages.

The bit-Flipping decoding algorithm is as follows:

Step 1.

Initialization: Each bit node is assigned the bit value

received from the channel, and sends messages to the check nodes to

which it is connected indicating this value.

Step 2.

Parity update: Using the messages from the bit nodes, each

check node calculates whether or not its parity-check equation is

satisfied. If all parity-check equations are satisfied the algorithm

terminates, otherwise each check node sends messages to the bit nodes

to which it is connected indicating whether or not the parity-check

equation is satisfied.

Step 3.

Bit update: If the majority of the messages received by each

bit node are not satisfied," the bit node ips its current value, otherwise

the value is retained. If the maximum number of allowed iterations is

reached the algorithm terminates and a failure to converge is reported,

otherwise each bit node sends new messages to the check nodes to

which it is connected, indicating its value and the algorithm returns to

Step 2.

443

To illustrate the operation of the bit-flipping decoder, that the

codeword c = 001011 is sent, and the word r = 101011 is received

from the channel.

The steps required to decode this received word are

In Step 1 the bit values are initialized to be 1, 0,1,0,1, and 1,

respectively, and messages are sent to the check nodes indicating these

values.

In Step 2 each parity-check equation is satisfied only if an even

number of the bits included in the parity-check equation is 1. For the

first and third check nodes this is not the case, and so they send “not

satisfied" messages to the bits to which they are connected.

In Step 3 the first bit has the majority of its messages indicating “not

satisfied" and so flips its value from 1 to 0. Step 2 is repeated and

since now all four parity-check equations are satisfied, the algorithm

halts and returns c = 001011 as the decoded codeword. The received

word has therefore been correctly decoded without requiring an

explicit search over all possible code words.

In this type of algorithm when neither of the parity-check equations

are satisfied, it is not possible to determine which bit is causing the

error.

Proposed Algorithm: Bit-flipping (BF) decoding algorithm is a hard-decision

decoding algorithm which is much simpler than SPA or its

modifications but does not perform as well. To reduce the

performance gap between SPA and BF based decoders, variants of the

latter such as weighted bit-flipping (WBF), modified weighted bit-

flipping (MWBF) and improved modified bit- flipping (IMWBF)

algorithms have been proposed. They provide tradeoffs between

computational complexity and error performance. The reliability ratio

based weighted bit-flipping (RRWBF) decoding algorithm needs not

to find optimal parameters as variants of the WBF algorithm do but

yields better performance.

(I) Weighted bit Flipping (WBF):

The standard belief-propagation (BP) decoding can achieve

excellent performance but often with heavy implementation

complexity. Bit-flipping (BF) decoding is simple, but it often incurs

considerable performance loss compared to the BP decoding. To

bridge the performance gap between BF decoding and BP decoding,

weighted BF (WBF) decoding and its variants were proposed.

For some high-rate low-density parity-check (LDPC) codes of large

row weight, it is shown that the WBF algorithm proposed performs

extraordinarily well.

Let C be a regular ( , ) LDPC code of block length N and dimension K, which has a parity check matrix H of M rows, and N columns with exactly “ones” in each row and “ones” in each column.

444

Let Rc = K/N denote its code rate.

Consider now that the LDPC coded bits are BPSK modulated, and

further transmitted over an additive white Gaussian noise (AWGN)

channel.

Let c = (c1, c2,……, cN) denote a codeword of C.

It is mapped to x = (x1, x2,……, xN) by xn = 2cn – 1 before

transmission.

At the receiver, we get the received vector y = (y1,y2,……; yN),

where yn = xn + vn

n = 1, 2,…..,N, vn is the zero-mean additive Gaussian noise

with variance of 2 = 10(2 / )c bR E N .

Let z = (z1, z2,……, zN) be the binary hard-decision vector

from y

i.e., zn = sgn(yn), where

sgn(y) = 1 if y 0 and sgn(y) = 0 if y < 0.

We denote the set of bits that participate check

m,nm by N(m) = n, h =1

Similarly, we denote the set of checks in which bit n participates as m,n M(n) = m, h =1

Conclusion:

In this project , a rate-control scheme called WBF , MWBF and

IMWBF has been proposed. These splitting methods can be a good

candidate for the next generation communication system which

requires high system throughput.

References:

[1] T. J. Richardson and R. L. Urbanke, “Efficient encoding of low-

density

parity-check codes,” IEEE Trans. Inf. Theory, pp. 638-656, Feb. 2001.

[2] H. Jin, A. Khandekar, and R. J. McEliece, “Irregular repeat-

accumulate

codes,” in Proc. 2nd Int. Symp. Turbo Codes Related Topics, Sep.

2000, pp. 1-8.

[3] IEEE 802.16 Broadband Wireless Access Project, “IEEE

C802.16e- 05/066r3,” Jan. 2005.

[4] IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. 57,

NO. 1, JANUARY 2009 Low-Density Parity-Check Codes with 2-

State Trellis Decoding

445

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

446

DATA TRANSMISSION USING OFDM COGNITIVE RADIO

UNDER INTERFERENCE ENVIRONMENT Vinothkumar J

1,Kalidoss R

2

1Student,

2Assistant Professor, Department of Electronics and Communication, SSN College of

Engineering, Kalavakkam, (TN) [email protected]

Abstract—A cognitive radio-based wireless mesh network is

considered. We consider a cognitive radio (CR) network that

makes opportunistic use of a set of channels licensed to a

primary network. During operation, the CR network is required

to carry out spectrum sensing to detect active primary users,

thereby avoiding interfering with them. Interference

temperature model is used to define the occupancy and

availability of a channel. The interference temperature (IT)

model offers a novel way to perform spectrum allocation and

management. Efficient and reliable subcarrier power allocation

in orthogonal frequency-division multiplexing (OFDM)-based

cognitive radio networks is a challenging problem.

Keywords— Cognitive radio, Interference temperature, Sensing

techniques, OFDM-CR.

I. INTRODUCTION:

THE traditional approach of fixed spectrum allocation to

licensed networks leads to spectrum under-utilization. In

recent studies by the FCC, it is reported that there are vast

temporal and spatial variations in the usage of allocated

spectrum.This motivates the concepts of opportunistic

spectrum access that allows cognitive radio (CR) networks to

opportunistically exploit under-utilized spectrum. On the one

hand, opportunistic spectrum access can improve the overall

spectrum utilization. On the other hand, transmission from CR networks can cause harmful interference to primary users of

the spectrum. To mitigate such a problem, CR networks can

frequently carry out spectrum sensing to detect active primary

users. Upon detecting an active cochannel primary user, a CR

network can either change its operating parameters, e.g.,

reduce its transmit powers, or move to another channel to

avoid interfering with the primary user . In most cases, to

achieve reliable spectrum sensing for a particular channel, a

CR network has to postpone all of its transmissions on that

channel, i.e., quiet sensing periods must be scheduled. Note

that scheduling quiet sensing periods results in negative

impacts to various performance metrics of CR networks, such as throughput and latency. One approach to reduce these

negative impacts is to design efficient spectrum sensing

algorithms that require short sensing time. One of the criteria,

proposed by the Interference Protection Working Group of

Spectrum Policy Task Force (set up by FCC) to

opportunistically share the licensed spectrum bands and to

quantify and manage the interference is interference

temperature. Interference temperature is a measure of the RF

power available at a receiving antenna to be delivered to

receiver – power that is generated by other emitters and noise

sources. FCC in its Notice for Proposed Rule Making

(NPRM) has suggested that unlicensed devices can be allowed to use licensed spectrum bands in a geographical region

provided the interference temperature at each licensed

receiver in the region does not exceed an interference

temperature threshold.

II. GENERAL CR SYSTEM MODEL:

We consider a CR deployment on licensed band, as

depicted in Fig. 1. Each CR network consists of a set of nodes

that are supported by a base station (BS). Each CR network

operates based on making opportunistic use of the channels

Fig.1.General system model

that belong to a primary network. This is one of the CR

network architectures similar to the future deployment of

IEEE802.22 technologies.

447

III. INTERFERENCE TEMPERATURE MODEL:

The concept of interference temperature is identical to that of

noise temperature. It is a measure of the power and bandwidth

occupied by interference. Interference temperature TI is

specified in Kevin and is denoted as

KB

BfPBfT

cI

cI

)(),(

,

where PI (fc;B) is the average interference power in Watts

centered at fc, covering bandwidth B measured in Hertz.

Boltzmann's constant k is 1.38 x 10^23 Joules per Kelvin degree. .

Figure 2: Example PSD for an unlicensed signal partially

overlapping a licensed signal.

The idea is that by taking a single measurement, a cognitive

radio can completely characterize both interference and noise

with a single number. Of course, it has been argued that

interference and noise behave differently. Interference

is typically more deterministic and independent of

bandwidth, whereas noise is not. For a given geographic area, the FCC would establish an interference temperature limit,

TL. This value would be a maximum amount of tolerable

interference for a given frequency band in a particular

location. Any unlicensed transmitter utilizing this band must

guarantee that their transmissions added to the existing

interference must not exceed the interference temperature

limit at a licensed receiver. While this may seem clear cut,

there is ambiguity over which signals are considered

interference, and which fc and B to use. Should they reflect

the unlicensed transceiver or the licensed receiver? For

example, consider figure 2 Should we use B1 or B2 as our

bandwidth for our computations? These ambiguities precipitate the need for our two interpretations.

III .INTERFERENCE TEMPERATURE

MEASUREMENT.

The FCC NPRM suggests three possible methods of

measuring interference temperature at different locations in a

region. In this subsection, we briefly summarize the three

methods.

1) Measurement by Primary Receiver

Since, ideally the interference temperature needs to

be measured and controlled at primary receivers, the most

appropriate approach is that the primary receivers themselves

measure the interference temperature at their antenna, and

send the values back to the unlicensed secondary devices in

the region. Though this approach is most accurate (as primary

receivers know the exact modulation type and waveform

details of transmitted primary signals), it requires major

hardware and software modifications in the primary receivers.

This is clearly infeasible, especially for devices (such as TVs,

Laptops, Mobile phones, PDAs, etc.) which are already developed and deployed. Moreover, it requires a channel for

explicitly transmitting the measured values to the secondary

transmitters.

2) Grid of Monitoring Stations

Another approach to measure interference

temperature is to deploy a grid of monitoring stations in the

target region. These devices are dedicated for measuring the

interference temperature at the location of their deployment,

and send these measurements back usually to a well-known sink node in the grid (from which the secondary devices can

obtain the interference temperature in their nearby region). A

major advantage of this approach is that it does not require

any modification in primary system. Moreover, these devices,

being dedicated for interference temperature measurement,

can be fine tuned for high precision, and are usually not

power-starved. This is in contrast to small devices (either

secondary or primary), such as mobile phones and PDAs,

where incorporating measurement capabilities are costly in

terms of silicon real-estate and battery power consumption.

On the other hand, this approach has several disadvantages too. First, the interference temperature is

measured at locations (i.e. at grid nodes), which are different

from where it need to be controlled (i.e. at primary receivers).

The interference temperature at these different locations may

be different due to differing terrain and path loss conditions.

This approach also suffers from hidden terminal problem.

Grid nodes usually cannot exactly know the path loss between

secondary transmitters and primary receivers, as well as the

shadowing and fading effects experienced at primary

receivers.

3) Measurement by Secondary Transmitters

448

The simplest but somewhat inaccurate method to

measure the interference temperature is to let a secondary

device itself measure the interference temperature locally.

This approach neither requires any modification in primary

system, nor any additional deployment of measuring services.

But the interference temperature that the secondary device

measures locally and the one that is present at a primary

receiver may differ significantly (unless both the devices are

very near to each other), due to location and terrain-dependent

multipath interference and shadowing affects. Moreover, secondary devices need to be equipped with interference

temperature measurement capabilities.

IV SPECTRUM SENSING IN CR NETWORKS:

Spectrum sensing is crucial for CR networks to

detect active primary users and avoid causing interference. Let

us briefly go through important parameters that characterize

spectrum sensing in CR networks.

1) Signal to Noise Ratio (SNR): When a primary user is active, the higher the SNR of the primary user’s signal at

the receiver of a CR device, the easier it is to detect. We

denote this SNR by 𝛾.

2) Probability of Detection 𝑃𝑑: This is the

probability that a CR network accurately detects the presence

of an active primary user. The higher the value of 𝑃𝑑, the

better the protection for primary operation.

3) Probability of False Alarm 𝑃𝑓 : This is the

probability that a CR network falsely detects the presence of

primary users when in fact none of them are active at the

sensing time. From the CR network point of view, the lower

the value of 𝑃𝑓, the higher the spectrum utilization.

4) Detection Time 𝑇𝑑: This is the time taken to detect

a primary user since it first turns on.

V OFDM- COGNITIVE RADIO:

Orthogonal-Frequency-Division Multiplexing

(OFDM) is the best physical layer candidate for a CR system

since it allows easy generation of spectral signal waveforms

that can fit into discontinuous and arbitrary-sized spectrum

segments. Besides, OFDM is optimal from the viewpoint of

capacity as it allows achieving the Shannon channel capacity

in a fragmented spectrum. Owing to these reasons, in this

paper, we consider the problem of Data Transmission in an

OFDM based CR system. When we transmit data in OFDM, BER decreases, when signal-to-noise ratio increases.

Simulation result is shown in the fig.3.

VI SIMULATION RESULT:

FIGURE.3.BER Vs SNR

VII CONCLUSION:

In this paper we have considered how to use both interference

temperature and the regulatory interference temperature limit

to select an optimal radio bandwidth for a particular

interference environment.Also we discussed both spectrum

sensing parameters and the effect of data transmission in

OFDM.

REFERENCES

[1] T. Charles Clancy, William A. Arbaugh, “Measuring

Interference Temperature” Dept of Computer science,

University of Maryland, college park, Labouratory for

telecommunication sciences, Dept of Defence,Vol 43,2009.

[2] Anh Tuan Hoang, , Ying-Chang Liang, and Yonghong

Zeng, “Adaptive Joint Scheduling of Spectrum Sensing and

Data Transmission in Cognitive Radio Networks”IEEE

Transactions on communications, Vol 58,No.1,Jan 2010.

[3] Ziaul Hasan, Gaurav Bansal. Ekram Hossain, Vijay K. Bhargava, “ Energy Efficient Power allocation in OFDM-

Based Cognitive Radio Systems: A Risk Return Model”IEEE

Transactions on Wireless Communications, Vol 8, No 12,

DEC 2009.

449

[4] Manuj Sharma, Anirudha Sahoo, K.D. Nayak, “channel

Selection Under IT Model in Multi-hop Cognitive Radio

Mesh Networks”, IEEE transactions on communication, Vol

6,2009.

Wireless Communication of Multi-User MIMO System with ZFBF and Sum Rate Analysis

k.srinivasan Department of Communication System

Dhanalakshmi Srinivasan Engineering College, Perambalur. [email protected]

Abstract-We present the exact sum-rate analysis of the multi-user multiple-input multiple-output (MIMO) systems with zero-forcing transmit beam forming (ZFBF). ZERO-FORCING transmit beam forming (ZFBF) is a practical multi-user transmission strategy for multi-user MIMO systems. We develop the analytical expressions of the ergodic sum-rate for two low-complexity user selection strategies for the dual-transmit antenna scenario. Based on the analytical results, we examine the parameter optimization problem to properly tradeoff between channel power gain and directional gain in term of maximizing the ergodic sum rate. Proper user selection scheme is essential for ZFBF-based multi-user MIMO system to fully benefit from multi-user diversity gain. The optimum ZFBF user selection strategy involves the exhaustive search of all possible user subsets, which becomes prohibitive to implement when the number of users is large.

Index TermsMIMO broadcast channel, zero-forcing beam forming, multi-user diversity, sum-rate analysis, and wireless communications.

I.INTRODUCTION

Zero-forcing transmit beam forming (ZFBF) is a practical Multi-user transmission strategy [1], [2] for multi-user MIMO systems. By designing one users beam forming vector to be orthogonal to other selected users channel vectors, ZFBF can completely eliminate multi-user interference. It has been shown that ZFBF can achieve the same sum-rate scale law of log log(), where is the number of users in the system, as the optimal DPC scheme [3][5] . Proper user selection scheme is essential for ZFBF-based multi-user MIMO system to fully benefit from multi-user diversity gain [6]. The optimum ZFBF user selection strategy involves the exhaustive search of all possible user subsets, which becomes prohibitive to implement when the number of users is large. Several low-complexity user selection strategies have been proposed and studied in the literature [6], [7], [9], [10]. Among them, two greedy search algorithms, the successive projection ZFBF (SUP-ZFBF)[6] and greedy weight clique ZFBF (GWC-ZFBF) [7], are attractive for implementation simplicity and their ability to achieve the same scale rate of double log as the DPC scheme. Most previous work on multi-user MIMO transmission schemes have been based on either simulation study or asymptotic analysis, which have limited applicability to accurate tradeoff analysis and design parameter optimization. In this paper, we adopt a different approach and

study multi-user ZFBF transmission schemes through statistical performance analysis. II. SYSTEM AND CHANNEL MODELS We consider a multiuser MIMO system with two transmit antennas at the base station. There are totally K≥2 single antenna users in the system. The channel gains from transmit antennas to users are assumed to be independent and identically distributed (i.i.d.) zero mean complex Gaussian random variable with unitary variance. As such, the norm square of the k-th users channel vector (termed as channel power gain in this paper) is a Chi-square distributed random variable with four degrees of freedom, with probability density function (PDF) and cumulative distribution function (CDF). III. SUM-RATE ANALYSIS In this section, we analyze the ergodic sum rate of ZFBF Based multi-user MIMO system with SUP-ZFBF and GWCZFBF schemes. For that purpose, we first derive the statistics of channel power gains of selected users.

analytical sum-rate expression derived in this paper, we can easily find the optimal value by using one-dimensional Optimization algorithms. We can also observe that the optimal

Proceedings of the Third National Conference on RTICT 2010Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

450

Value of increases with the number of users , which Indicates that when more users are in the system we should trade channel power diversity for directional diversity gain. EXISTING SYSTEM

Most previous work on multi-user MIMO

transmission schemes have been based on either simulation study or asymptotic analysis, which have limited applicability to accurate tradeoff analysis and design parameter optimization. Several low-complexity user selection strategies have been proposed. Among them, two greedy search algorithms, the successive projection ZFBF (SUP-ZFBF) and greedy weight clique ZFBF (GWC-ZFBF), are attractive for implementation simplicity and their ability to achieve the same scale rate of double log as the DPC scheme. This is based on additive white Gaussian noise channel (AWGN).

PROPOSED SYSTEM In this paper, we adopt a different approach and study multi-user inter symbol interference (ISI) transmission schemes through statistical performance analysis. Specifically, we derive the exact analytical expressions for the ergodic sum rate of the SUP-ZFBF and GWC-ZFBF schemes for the important dual-transmit-antenna scenario. We then apply these analytical results to the parameter optimization for each scheme. We show that while the sum rate performance of GWC-ZFBF scheme is sensitive to the channel direction constraint that of SUP-ZFBF scheme is affected little as long as the constraint is not very strict. By using fading channel transmission and reception is performed without noise. APPLICATION This system has enormous application in wireless communication. Such as transmission of signal without affecting additive white Gaussian noise. This system has various applications in signal transmission and reception. FUTURE ENCHANCEMENT

Hardware implementation of the system will give better results than simulation.

ADVANTAGES

This system uses a noiseless channel transmission. The transmission time duration is less compared to

existing systems. The system does not use AWGN this is the main

advantage of this system Complexity is less. More than companies are occupied in this processing.

Conclusion:

Opportunistic beam forming with clustered OFDM promises throughput on par with a smart antenna solution, with little feedback and possibly a lower complexity base station.

REFERENCES [1] P. Viswanath, D. N. C. Tse, and R. Laroia, Opportunistic beamforming using dumb antennas," IEEE Trans. Inform. Theory, vol. 48, no. 6, pp.1277-1294, June 2002. [2] W. Rhee, W. Yu, and J. M. Cioffi, The optimality of beamforming in uplink multiuser wireless systems," IEEE Trans. Wireless Commun.,vol. 3, no. 1, pp. 86-96, Jan. 2004. [3] N. Jindal and A. Goldsmith, Dirty paper coding vs. TDMA for MIMO broadcast channels," in Proc. IEEE Int. Conf. Commun. (ICC2004), Paris, France, vol. 2, pp. 682-686, June 2004. [4] M. Sharif and B. Hassibi, A comparison of time-sharing, DPC, and beamforming for MIMO broadcast channels with many users," IEEE Trans. Commun., vol. 55, no. 1, pp. 11-15, Jan. 2007. [5] M. Costa, Writing on dirty paper," IEEE Trans. Inform. Theory, vol. 29, pp. 439-441, May 1983. [6] T. Yoo and A. Goldsmith, On the optimality of multiantenna broadcast

451

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

452

Low cost Embedded based Braille Reader

J.Raja

1, M.Chozhaventhar

2, V.Eswaran

2, K.Vinoth Kumar

2

1Assistant Professor

2Final Year Student

Department of Electronics and Communication Engineering

Adhiparasakthi Engineering College, Melmaruvathur, Tamilnadu-603319, India

Abstract—The aim of our project is to design and

develop a Braille Reader. It is ideal for those

unfortunate people who just turned blind and have not

mastered Braille reading. It can also be used as a

learning instrument that helps the user decipher Braille

without constantly going to the Braille dictionary. The

Braille is read by a combination of six push buttons.

When the buttons are pressed against the Braille, the

buttons corresponding to the bumps on the Braille will

be pushed. The binary data thus obtained from the

push buttons is given to the microcontroller. A memory

is loaded with pre-defined entries. These entries are

selected by the microcontroller that generates the audio output of the corresponding Braille.

Keywords: Braille sensor, Speech synthesizer, phonemes, allophones.

I. INTRODUCTION

The Braille system is a method that is widely used by

blind people to read and write. Braille was devised in 1821 by Louis Braille, a blind Frenchman. Braille is a

printed language system designed to enable people

with visual disability to “read” printed documents

with only a sense of touch. Braille can represent

alphabets, numbers, punctuation marks, or anything

found in ordinary printed materials. A standard

Braille character is represented as a 2 by 3 matrix of

sub cells which may or may not contain dots. This is

known as a Braille Cell. Characters are uniquely

identified based on the number and arrangement of

dots embossed. Within a Braille Cell, the height of each dot is approximately 0.02 inches (0.5mm) and

the vertical and horizontal spacing between the dots’

centre are approximately 0.1 inches (2.5mm). In

addition, the blank space between the right hand

column dots’ centre and the left hand column dots’

centre of an adjacent cell is approximately 0.15

inches (3.75mm). The blank space between each

Braille Cell of each row is 0.2 inches (5.0mm). A

standard Braille page is 11 by 11inches (280.5 by

280.5mm) with 40 to 42 Braille Cells per line and a

maximum of 25 lines. Josephson (1961) studied the leisure

activities of blind adults and found that Braille

reading was infrequently mentioned by these adults

as a pleasurable leisure activity. A related report on

blind readers (Josephson, 1964) revealed that many

blind readers engaged in Braille reading only when

there was no alternative access to the information

they needed. It seems plausible that the reading rates

of many blind individuals are so painfully slow, these

persons may prefer to read by listening to the

recorded discs, tapes, and cassettes that are prepared

for them. Others may resort to a passive acceptance of whatever information and pleasure is available to

them by way of radio or television. We do not know

for certain that individuals who read little or no

Braille are, in fact, slow readers. However, we do

know reading as slowly as 90 words per minute for

long periods of time is, at best, tedious. What

accounts for the typically slow Braille reading rate

has yet to be determined?

The Braille Reader is a tool for visually impaired

people to understand printed Braille with ease. It

requires no previous knowledge of printed Braille systems. If you do not possess an understanding of

Braille but have a need to read a Braille document,

our device would be useful.

II. MOTIVATION

As explained above the Braille reading using

optical recognition based on image processing. This

method is too expensive and involves wastage of the resources. As the Braille languages are standardized

and its cells are of constant size, it is not necessary to

use such computationally complex procedures to

decipher the Braille. Also it involves image

453

processing that obviates need of digital camera which

incur additional cost on the device.

Hence in order to provide cost effective

solution and easy to handle device to such visually

challenged people we proposed our system in that

Instead of image sensing, we decided to use 6 push buttons in a 2x3 matrix. When the buttons are

pressed against the Braille, the buttons corresponding

to the bumps on the Braille will be pushed.

Depending upon this combination corresponding

audio output will be generated with the help of

microcontroller and speech synthesizer.

III. PROPOSED SYSTEM

Our proposed system involves three main

components .they are i) Braille sensor ii)

microcontroller and iii) speech synthesizer.

Fig.1.Braille Sensor

Braille sensor is the physical part that acts as an

interface between the blind people and Braille book.

Braille sensor can be realized with the help of six

push buttons as shown. When this Braille sensor are

pressed against Braille, combination of bumps

present in the Braille cell produce unique binary

pattern as shown above. On obtaining binary input, it

is given as an input to the microcontroller which is

programmed in such a way as to recognize it and obtain its corresponding text output.

Thus overall block diagram of our proposed system

can be represented as shown in fig.1.

Once the text of the Braille was obtained, speech

synthesizers enter into play. It contains the phoneme

sets that are previously stored. In order to retrieve the

phonemes that corresponds to the input

microcontroller outputs the sequence of commands

required via its serial port.

Normally speech synthesizers are self contained, single chip that produces complex sounds

depends on the input commands. Hence when

connected it on speaker it produces audio of the

corresponding input.

Fig.2.Braille Reader

IV. HARDWARE REQUIREMENTS

After analyzing the various requirements we have proposed to use the following

hardware components

Protoboard – This is the skeleton of the Braille

reader. It holds the Atmel 8352 microcontroller and

supports links to other components.

Braille Sensor – A combination of 6 NKK buttons.

This is like user’s eye. It converts Braille to Binary

data. (B2B) So the MegaL32 can process the data.

ATMega8535 – The heart of Braille Reader. This

microcontroller receives data from Braille sensor and

generates the sound output at amazing speed. (16

MHz) It also controls different parameters of the

Braille Reader such as Volume.

SpeakJet –This speech synthesizer can acts as the messenger between the microcontroller and the user.

Whenever a button is pushed or a Braille is read, the

SpeakJet will generate a robotic voice and deliver a

message to the user to make sure the user is updated

with his/her surrounding.

CHARACTERER

VOICE

BRAILLE SENSOR

MICROCONTROLLER

SPEECH SYNTHESIZER PHONEME SET

BINARY

454

Headphone or speaker– In ear design, so clear

sound can be transferred to the user without too much

lost from noise.

V. DESIGN

The SpeakJet chip can be setup up in two ways. One

for Demo/Test mode and one for serial control.

Fig.3.Minimum connections for Demo/Test Mode

Fig.4.Minimum connections for Serial Control

In demo mode, the chip outputs random phonemes,

both biological and robotic in order to test it out.

Serial Data is the main method of communicating

with the SpeakJet. The serial configuration to

communicate with it is 8 bits, No-Parity and the

default factory baud rate is 9600. The Serial Data is

used to send commands to the MSA unit which in

turn communicates with the 5 Channel Synthesizers to produce the voices and sounds.

Fig.5.MSA Architecture

In order to make a quality speech generation system,

a vast library of words and their pronunciation is

needed. The onboard flash memory is only 32kB.

Assuming each word takes 8 bytes and its

corresponding allophone takes 16 bytes, we can fit

only about 1330 word. This is definitely not enough

for speech generation. So we decided to use a 2MB eeprom. With that, we can store about 22182 words,

which should cover most of the word we use in daily

speech and some uncommon words.

Each eeprom is 1mbit. It is made of two blocks of

2^16 bytes. The 24AA1025 eeprom uses Twin Wire

Interface (TWI) for data transmission. As its name

implies, it uses only two bi-directional bus lines, one

for clock (SCL) and one for data (SDA)

455

VI. SOFTWARE

In order to pronounce the words that are not in the

library, the MCU will divide the input word into

existing words. It would break the unknown word

down by finding the longest existing word from our

library that matches with the prefix of the unknown

word.

For example:

Input Word: “appletree”

The Braille reader will pronounce: “apple”

+ “tree”

For data storage, all words can be put in a single

character array with “*” between words to indicate

where words start and stop. This design can be

picked over double character array because double

character array requires same length for each word.

So the shortest word “a” will have to take the same

amount of space as the longest word in the

dictionary. However, the single array made binary

search impossible. So linear search can be used.

In order to save time with searching, the words in the

dictionary are arranged in alphabetical order and

recorded the position for each of the 26 letters in the

alphabet. So instead of traverse through the entire

library for searching, it will only search words that

match the first letter.

The CMU dictionary provided by Carnegie Mellon

University for research purposes can be used. The

dictionary contains over 125,000 North American

English words with their transcriptions. The

dictionary uses 39 phonemes in its transcriptions.

Phoneme Set

Phoneme Example Translation

------- ------- -----------

AA odd AA D

AE at AE T

AH hut HH AH T

AO ought AO T

AW cow K AW

AY hide HH AY D B be B IY

CH cheese CH IY Z

D dee D IY

DH thee DH IY

EH Ed EH D

ER hurt HH ER T

EY ate EY T

F fee F IY

G green G R IY N

HH he HH IY

IH it IH T IY eat IY T

JH gee JH IY

K key K IY

L lee L IY

M me M IY

N knee N IY

NG ping P IH NG

OW oat OW T

456

OY toy T OY

P pee P IY

R read R IY D

S sea S IY

SH she SH IY

T tea T IY

TH theta TH EY T AH

UH hood HH UH D

UW two T UW V vee V IY

W we W IY

Y yield Y IY L D

Z zee Z IY

ZH seizure S IY ZH ER

VII. CONCLUSION

This paper makes an attempt to provide a

low cost solution to the visually impaired people. The

paper is based on microcontroller and can also be

used by dumb blind people to communicate freely

with the external world. This device can also be

designed at much lower price without synthesizer by

storing the required allophones on a memory. . This

would also have saved a lot more money than buying

a retail Braille reader which costs $1,400.Thus this

would definitely assist the visually challenged to decipher the Braille with less burden at affordable

cost.

REFERENCES

[1] y. Matsuda, et al., “analysis of emotional expression of finger

braille,” ifmbe proc. Vol. 19, 7th asian-pacific conference on

medical and biological engineering, china, pp.484-487, april 2008.

[2] y. Matsuda, et al., “finger braille teaching system for people

who communicate with deaf blind people,” proc. Of the 2007 ieee

international conference on mechatronics and automation, china,

pp.3202-3207, august 2007.

[3] i/o terminal for finger braille,” proc. Of the human interface

symposium 2000, japan, pp.37-40, september 2000.

[4] m. Fukumoto, et al., “body coupled fingering: wireless

wearable keyboard,” proc. Of the acm conference on human

factors in computing systems, france, pp.147-154, march 1997.

[5] bouraoui a., extending an existing tool to create non visual

interfaces. Proceedings of csun 2006, centre on disabilities 21st

annual international technology and persons with disabilities

conference, california state university northridge, 2006.

[6] information about braille and other reading codes for the blind

http://www.brailler.com/braillehx.htm

http://www.nyise.org/blind/

http://www.duxburysystems.com/braille.asp

http://en.wikipedia.org/wiki/braille

[7] cmu dictionary

http://www.speech.cs.cmu.edu/cgi-bin/cmudict

[8]speakjet datasheet

http://www.magnevation.com/pdfs/speakjetusermanual.pdf

[9]phrase-a-lator software and phonetic dictionary

http://magnevation.com/software.htm

Unbiased Sampling in Ad-hoc (Peer-to-Peer) Networks

M.Deiveegan,A.Saravanan,S.Vinothkannan

Department of Information Technology,Anna university [email protected]

[email protected] [email protected]

Abstract—This paper presents a detailed examination of how the dynamic and heterogeneous nature of real-world peer-to-peer systems can introduce bias into the selection of representative samples of peer properties (e.g., degree, link bandwidth, number of files shared). We propose the Metropolized Random Walk with Backtracking and Linking (MRWBL) technique for collecting nearly unbiased samples and conduct an extensive simulation study to demonstrate that our technique works well for a wide variety of commonly-encountered peer-to-peer net work conditions. We have implemented the MRWBL algorithm for selecting peer addresses uniformly at random. Using the Gnutella network, we empirically show that ion-sampler yields more accurate samples than tools that rely on commonly-used sampling techniques and results in dramatic improvements in efficiency and scalability compared to performing earlier algorithms.

Index Terms—Peer-to-peer, sampling.

I. INTRODUCTION

HE popularity and wide-spread use of peer-to-peer sys-tems has motivated numerous empirical studies aimed at

providing a better understanding of the properties of deployed peer-to-peer systems. However, due to the large scale and highly dynamic nature of many of these systems, directly measuring the quantities of interest on every peer is prohibitively expen-sive. Sampling is a natural approach for learning about these systems using light-weight data collection, but commonly-used sampling techniques for measuring peer-to-peer systems tend to introduce considerable bias for two reasons. First, the dy-namic nature of peers can bias results towards short-lived peers, much as naively sampling flows in a router can lead to bias to-wards short-lived flows. Second, the heterogeneous nature of the overlay topology can lead to bias towards high-degree peers.

In this paper, we are concerned with the basic objective of devising an unbiased sampling method, i.e., one which selects any of the present peers with equal probability. The addresses of the resulting peers may then be used as input to another mea-surement tool to collect data on particular peer properties (e.g., degree, link bandwidth, number of files shared). The focus of our work is on unstructured P2P systems, where peers select neighbors through a predominantly random process. Most pop-ular P2P systems in use today belong to this unstructured cate-gory. For structured P2P systems such as Chord [1] and CAN [2], knowledge of the structure significantly facilitates unbiased sampling as we discuss in Section VII.

Achieving the basic objective of selecting any of the peers present with equal probability is non-trivial when the structure of the peer-to-peer system changes during the measurements. First-generation measurement studies of P2P systems typically

relied on ad-hoc sampling techniques (e.g., [3], [4]) and pro-vided valuable information concerning basic system behavior. However, lacking any critical assessment of the quality of these sampling techniques, the measurements resulting from these studies may be biased and consequently our understanding of P2P systems may be incorrect or misleading. The main contri-butions of this paper are (i) a detailed examination of the ways that the topological and temporal qualities of peer-to-peer sys-tems (e.g., Churn [5]) can introduce bias, (ii) an in-depth ex-ploration of the applicability of a sampling technique called the Metropolized Random Walk with Backtracking (MRWB), repre-senting a variation of the Metropolis-Hastings method [6]-[8], and (iii) an implementation of the MRWB algorithm into a tool called -. While sampling techniques based on the original Metropolis-Hastings method have been considered ear-lier (e.g., see Awanet al. [9] andBar-YossefandGurevich [10]), we show that in the context of unstructured P2P systems, our modification of the basic Metropolis-Hastings method results in nearly unbiased samples under a wide variety of commonly encountered peer-to-peer network conditions.

The proposed MRWB algorithm assumes that the P2P system provides some mechanism to query a peer for a list of its neighbors—a capability provided by most widely deployed P2P systems. Our evaluat ions of the - tool shows that the MRWB algorithm yields more accurate samples than previously considered sampling techniques. We quantify the observed differences, explore underlying causes, address the tool's efficiency and scalability, and discuss the implications on accurate inference of P2P properties and high-fidelity modeling of P2P systems. While our focus is on P2P networks, many of our results apply to any large, dynamic, undirected graph where nodes may be queried for a list of their neighbors.

After discussing related work and alternative sampling tech-niques in Section II, we build on our earlier formulation in [11] and focus on sampling techniques that select a set of peers uni-formly at random from all the peers present in the overlay and then gather data about the desired properties from those peers. While it is relatively straightforward to choose peers uniformly at random in a static and known environment, it poses consid-erable problems in a highly dynamic setting like P2P systems, which can easily lead to significant measurement bias for two reasons.

The first cause of sampling bias derives from the temporal dynamics of these systems, whereby new peers can arrive and existing peers can depart at any time. Locating a set of peers and measuring their properties takes time, and during that time the peer constituency is likely to change. In Section III, we show

T

Proceedings of the Third National Conference on RTICT 2010Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

457

how this often leads to bias towards short-lived peers and ex-plain how to overcome this difficulty.

The second significant cause of bias relates to the connectivity structure of P2P systems. As a sampling program explores a given topological structure, each traversed link is more likely to lead to a high-degree peer than a low-degree peer, significantly biasing peer selection. We describe and evaluate different tech-niques for traversing static overlays to select peers in Section IV and find that the Metropolized Random Walk (MRW) collects unbiased samples.

In Section V, we adapt MRW for dynamic overlays by adding backtracking and demonstrate its viability and effectiveness when the causes for both temporal and topological bias are present. We show via simulations that the MRWB technique works well and produces nearly unbiased samples under a variety of circumstances commonly encountered in actual P2P systems.

Finally, in Section VI we describe the implementation of the tool based on the proposed MRWBL algorithm and

empirically evaluate its accuracy and efficiency through com-parison with complete snapshots of Gnutella taken with Cruiser [12], as well as with results obtained from previously used, more ad-hoc, sampling techniques. Section VII concludes the paper by summarizing our findings and plans for future work.

II. RELATED WORK

A. Graph Sampling

The phrase "graph sampling" means different things in dif-ferent contexts. For example, sampling from a class of graphs has been well studied in the graph theory literature [13], [14], where the main objective is to prove that for a class of graphs sharing some property (e.g., same node degree distribution), a given random algorithm is capable of generating all graphs in the class. Cooper et al. [15] used this approach to show that their algorithm for overlay construction generates graphs with good properties. Our objective is quite different; instead of sam-pling a graph from a class of graphs our concern is sampling peers (i.e., vertices) from a largely unknown and dynamically changing graph. Others have used sampling to extract informa-tion about graphs (e.g., selecting representative subgraphs from a large, intractable graph) while maintaining properties of the original structure [16]—[18]. Sampling is also frequently used as a component of efficient, randomized algorithms [19]. How-ever, these studies assume complete knowledge of the graphs in question. Our problem is quite different in that we do not know the graphs in advance.

A closely related problem to ours is sampling Internet routers by running traceroute from a few hosts to many destinations for the purpose of discovering the Internet's router-level topology. Using simulation [20] and analysis [21], research has shown that traceroute measurements can result in measurement bias in the sense that the obtained samples support the inference of power law-type degree distributions irrespective of the true na-ture of the underlying degree distribution. A common feature of our work and the study of the traceroute technique [20], [21] is that both efforts require an evaluation of sampling techniques without complete knowledge of the true nature of the underlying connectivity structure. However, exploring the router topology and P2P topologies differ in their basic operations for graph-ex-ploration. In the case of traceroute, the basic operation is "What is the path to this destination?" In P2P networks, the basic oper-ation is "What are the neighbors of this peer?" In addition, the

Internet's router-level topology changes at a much slower rate than the overlay topology of P2P networks.

Another closely related problem is selecting Web pages uni-formly at random from the set of all Web pages [22], [23]. Web pages naturally form a graph, with hyper-links forming edges between pages. Unlike unstructured peer-to-peer networks, the Web graph is directed and only outgoing links are easily dis-covered. Much of the work on sampling Web pages therefore focuses on estimating the number of incoming links, to facili-tate degree correction. Unlike peers in peer-to-peer systems, not much is known about the temporal stability of Web pages, and temporal causes of sampling bias have received little attention in past measurement studies of the Web.

B. Random Walk-Based Sampling of Graphs

A popular technique for exploring connectivity structures consists of performing random walks on graphs. Several properties of random walks on graphs have been extensively studied analytically [24], such as the access time, cover time, and mixing time. While these properties have many useful applications, they are, in general, only well-defined for static graphs. To our knowledge the application of random walks as a method of selecting nodes uniformly at random from a dynamically changing graph has not been studied. A number of papers [25]-[28] have made use of random walks as a basis for searching unstructured P2P networks. However, searching simply requires locating a certain piece of data anywhere along the walk, and is not particularly concerned if some nodes are preferred over others. Some studies [27], [28] additionally use random walks as a component of their overlay-construction algorithm. Two papers that are closely related to our random walk-based sampling approach are by Awan et al. [9] and Bar-Yossef and Gurevich [10]. While the former also address the problem of gathering uniform samples from peer-to-peer networks, the latter are concerned with uniform sampling from a search engine's index. Both works examine several random walk tech-niques, including the Metropolis-Hastings method, but assume an underlying graph structure that is not dynamically changing. In addition to evaluating their techniques empirically for static power-law graphs, the approach proposed by Awan et al. [9] also requires special underlying support from the peer-to-peer application. In contrast, we implement the Metropolis-Hast-ings method in such a way that it relies only on the ability to discover a peer's neighbors, a simple primitive operation commonly found in existing peer-to-peer networks. Moreover, we introduce backtracking to cope with departed peers and con-duct a much more extensive evaluation of the proposed MRWB method. Specifically, we generalize our formulation reported in [11] by evaluating MRWB over dynamically changing graphs with a variety of topological properties. We also perform empirical validations over an actual P2P network.

C. Sampling in Hidden Populations

The problem of obtaining accurate estimates of the number of peers in an unstructured P2P network that have a certain prop-erty can also be viewed as a problem in studying the sizes of hidden populations. Following Salganik [29], a population is called "hidden" if there is no central directory of all population members, such that samples may only be gathered through re-ferrals from existing samples. This situation often arises when

458

public acknowledgment of membership has repercussions (e.g., injection drug users [30]), but also arises if the target population is difficult to distinguish from the population as a whole (e.g., jazz musicians [29]). Peers in P2P networks are hidden because there is no central repository we can query for a list of all peers. Peers must be discovered by querying other peers for a list of neighbors.

Proposed methods in the social and statistical sciences for studying hidden populations include snowball sampling [31], key informant sampling [32], and targeted sampling [33]. While these methods gather an adequate number of samples, they are notoriously biased. More recently, Heckathorn [30] (see also [29], [34]) proposed respondent-driven sampling, a snowball-type method for sampling and estimation in hidden populations. Respondent-driven sampling first uses the sample to make infer-ences about the underlying network structure. In a second step, these network-related estimates are used to derive the propor-tions of the various subpopulations of interest. Salganik et al. [29], [34] show that under quite general assumptions, respon-dent-driven sampling yields estimates for the sizes of subpop-ulations that are asymptotically unbiased, no matter how the seeds were chosen.

Unfortunately, respondent-driven sampling has only been studied in the context where the social network is static and does not change with time. To the best of our knowledge, the accuracy of respondent-driven sampling in situations where the underlying network structure is changing dynamically (e.g., unstructured P2P systems) has not been considered in the existing sampling literature

III. SAMPLING WITH DYNAMICS

We develop a formal and general model of a P2P system as follows. If we take an instantaneous snapshot of the system at time , we can view the overlay as a graph G(V, E) with the peers as vertices and connections between the peers as edges. Extending this notion, we incorporate the dynamic aspect by viewing the system as an infinite set of time-indexed graphs, Gt = . The mo st co mmon approach fo r samp l ing from this set of graphs is to define a measurement window, [to, to + , and select peers uniformly at random from the set of peers who are present at any time during the window: Vto,to+A = UtLt Vt- Thus, it does not distinguish between occurrences of the same peer at different times.

This approach is appropriate if peer session lengths are expo-nentially distributed (i.e., memoryless). However, existing mea-surement studies [3], [5], [37], [38] show session lengths are heavily skewed, with many peers being present for just a short time (a few minutes) while other peers remain in the system for a very long time (i.e., longer than ). As a consequence, as A increases, the set Vtotto+A includes an increasingly large fraction of short-lived peers.

A simple example may be illustrative. Suppose we wish to observe the number of files shared by peers. In this example system, half the peers are up all the time and have many files, while the other peers remain for around 1 minute and are imme-diately replaced by new short-lived peers who have few files. The technique used by most studies would observe the system for a long time (A) and incorrectly conclude that most of the peers in the system have very few files. Moreover, their results will depend on how long they observe the system. The longer the

measurement window, the larger the fraction of observed peers with few files.

One fundamental problem of this approach is that it focuses on sampling peers instead of peer properties. It selects each sampled vertex at most once. However, the property at the vertex may change with time. Our goal should not be to select a vertex Vi e Ut=t , but rather to sample the property at Vi at a par-ticular instant . Thus, we distinguish between occurrences of the same peer at different times: samples vi>t and vitt> gathered at distinct times t ^ t' are viewed as distinct, even when they come from the same peer. The key difference is that it must be possible to sample from the same peer more than once, at different points in time. Using the formulation ui;t e , t £ [to, to + , the sampling technique will not be biased by the dynamics of peer behavior, because the sample set is decoupled from peer session lengths. To our knowledge, no prior P2P measurement studies relying on sampling make this distinction.

Returning to our simple example, our approach will correctly select long-lived peers half the time and short-lived peers half the time. When the samples are examined, they will show that half of the peers in the system at any given moment have many files while half of the peers have few files, which is exactly cor-rect.

If the measurement window (A) is sufficiently small, such that the distribution of the property under consideration does not change significantly during the measurement window, then we may relax the constraint of choosing t uniformly at random from [t0, t0 + A].

We still have the significant problem of selecting a peer uni-formly at random from those present at a particular time. We begin to address this problem in Section IV.

IV. SAMPLING FROM STATIC GRAPHS

We now turn our attention to topological causes of bias. Towards this end, we momentarily set aside the temporal issues by assuming a static, unchanging graph. The selection process begins with knowledge of one peer (vertex) and progressively queries peers for a list of neighbors. The goal is to select peers uniformly at random. In any graph-exploration problem, we have a set of visited peers (vertices) and a front of unexplored neighboring peers. There are two ways in which algorithms differ: (i) how to chose the next peer to explore, and (ii) which subset of the explored peers to select as samples. Prior studies use simple breadth-first or depth-first approaches to explore the graph and select all explored peers. These approaches suffer from several problems:

• The discovered peers are correlated by their neighbor rela tionship.

• Peers with higher degree are more likely to be selected. • Because they never visit the same peer twice, they will

introduce bias when used in a dynamic setting as described in Section III.

1) Random Walks: A better candidate solution is the random walk, which has been extensively studied in the graph theory

literature (for an excellent survey see [24]). We briefly summarize the key terminology and results relevant to sampling. The transition matrix P(x. y) describes the

459

probability of transitioning to peer y if the walk is currently at peer :

If the vector v describes the probability of currently being at each peer, then the vector v' = vP describes the probability

In other words, if we select a peer as a sample every r steps, for sufficiently large , we have the following good properties:

• The information stored in the starting vector, , is lost, through the repeated selection of random neighbors. There fore, there is no correlation between selected peers. Alter nately, we may start many walks in parallel. In either cases, afte r steps, the selection is independent of the origin.

• While the stationary distribution, , is biased towards peers with high degree, the bias is precisely known, al lowing us to correct it.

• Random walks may visit the same peer twice, which lends itself better to a dynamic setting as described in Section III.

In practice, r need not be exceptionally large. For graphs where the edges have a strong random component (e.g., small-world graphs such as peer-to-peer networks), it is sufficient that the number of steps exceed the log of the population size, i.e., r>O(\og\V\). 2) Adjusting for Degree Bias: To correct for the bias towards

high degree peers, we make use of the Metropolis-Hastings method for Markov Chains. Random walks on a graph are a spe cial case of Markov Chains. In a regular random walk, the tran sition matrix P(x,y) leads to the stationary distribution , as described above. We would like to choose a new transition matrix, , to produce a different stationary distribution,

. Specifically, we desire /J,(X) to be the uniform distribution so that all peers are equally likely to be at the end of the walk. Metropolis-Hastings [6]-[8] provides us with the

desired Equivalently, to take a step from peer , select a neighbor y of x as normal (i.e., with probability ) . T h e n , w i t h probability min (^^1^Z'X\, 1), accept the move. Otherwise, return to x (i.e., with probability 1 - YJZ^X Qix-).

To collect uniform samples, we have ijA = , so the move-acceptance probability becomes:

.

Therefore, our algorithm for selecting the next step from some peer x is as follows:

• Select a neighbor y of x uniformly at random. Quer y for a list of its neighbors, to determine its degree. • Generate a random value, , uniformly between 0 and 1.

• If V < degree^) ' V iS me DeXt SteP'

• Otherwise, remain at x as the next step. We call this the Metropolized Random Walk (MRW).

Qualitatively, the effect is to suppress the rate of transition to peers of higher degree, resulting in selecting each peer with equal probability. 3) Evaluation: Although [6] provides a proof of

correctness for the Metropolis-Hastings method, to ensure the correctness of our implementation we conduct evaluations through simulation over static graphs. This additionally provides the opportunity to compare MRW with conventional techniques such as Breadth-First Search (BFS) or naive random walks (RW) with no adjustments for degree bias.

To evaluate a technique, we use it to collect a large number of sample vertices from a graph, then perform a goodness-of-fit test against the uniform distribution. For Breadth-First Search, we simulate typical usage by running it to gather a batch of 1,000 peers. When one batch of samples is collected, the process is reset and begins anew at a different starting point. To ensure robustness with respect to different kinds of connectivity structures, we examine each technique over several types of graphs as follows:

• Erdös-Rényi: The simplest variety of random graphs • Watts-Strogatz: "Small world" graphs with high

clus tering and low path lengths

• Barabási-Albert: Graphs with extreme degree distribu tions, also known as power-law or scale-free graphs

• Gnutella: Snapshots of the Gnutella ultrapeer topology, captured in our earlier work [39]

To make the results more comparable, the number of vertices (\V\ = 161.680) and edges (\E\ = 1.946,596) in each graph are approximately the same.1Table I presents the results of the goodness-of-fit tests after collecting 1000-|F| samples, showing that Metropolis-Hastings appears to generate uniform samples over each type of graph, while the other techniques fail to do so by a wide margin. typical node should be selected k times, with other nodes being selected close to k times approximately following a normal dis tribution with variance .2 We used k = 1,000 samples. We also include an "Oracle" technique, which selects peers uniformly at random using global information. The Metropolis-Hastings results are virtually identical to the Oracle, while the other techniques select many peers much more and much less than k times. In the Gnutella, Watts-Strogatz, and Barabási-Albert graphs, Breadth-First Search exhibits a few vertices that are selected a large number of times (> . T h e ( n o t -a d - justed) Random Walk (RW) method has similarly selected a few vertices an exceptionally large number of times in the Gnutella and Barabási-Albert models. The Oracle and MRW, by contrast, did not select any vertex more than

=

otherwise ,0

x,ofneighbor a isy ,)(deg

1),( xreeyxP

460

around 1,300 times. In summary, the Metropolis-Hastings method selects peers

uniformly at random from a static graph. The Section V exam-ines the additional complexities when selecting from a dynamic graph, introduces appropriate modifications, and evaluates the algorithm's performance.

V. Metropolised Random Walk with Backtracking and Linking algorithm.

In this paper we propose a new algorithm called Metropolised Random Walk with Backtracking and Linking. The major drawback of the Metropolised Random walk With backtracking algorithm is there is a lot of chances for failure of the algorithm because they use the stack for storing the samples .Stack works on the principle LIFO(Last In First Out).If we want to pop or remove a sample from the stack means we have to remove all its neibhours list then only we can pop a single element from the stack. So if a node fails to respond then we have to remove even the nodes that responds. In order to avoid this problem we propose a new method that is MRWBL for collecting the samples.In this algorithm we use the linked list for storing the samples. The nodes will get the information about the neer node that was connected to it . If any node does not responds in the list then the previous node will directly get connected with the new node which is connected to the earlier node. So there is no chance of failure of the whole algorithm .So in this way this algorithm will help the nodes to transfer the datas over the path that is senced or discovered by using the ion-sampler tool and saved in the linked list.

VI. DISCUSSION

A. How Many Samples are Required?

An important consideration when collecting samples is to know how many samples are needed for statistically significant results. This is principally a property of the distribution being sampled. Consider the problem of estimating the underlying fre-

quency / of an event, e.g., that the peer degree takes a particular value. Give n unbiased samples, an unbiased estimate of / is / = M/N where M is the number of samples for which the event occurs. / has root mean square (RMS) relative error

From this expression, we derive the following observations: • Estimation error does not depend on the population size;

in particular the estimation properties of unbiased sampling scale independently of the size of the system under

study. • The above expression can be inverted to derive the

number of samples Nfj(J required to estimate an outcome of frequency / up to an error . A simple bound is Nf „ <

• Unsurprisingly, smaller frequency outcomes have a larger relative error. For example, gathering 1,000 unbiased samples gives us very little useful information about events which only occur one time in 10,000; the associated a value is approximately 3: the likely error dominates the value to be estimated. This motivates using biased sampling .

The presence of sampling bias complicates the picture. If an event with underlying frequency / is actually sampled with fre- quency , then the RMS relative error acquires an additional term (1 - /o//)2 which does not reduce as the number of samples N grows. In other words, when sampling from a biased distribution, increasing the number of samples only increases the accuracy with which we estimate the biased distribution.

B. Unbiased Versus Biased Sampling

At the beginning of this paper, we set the goal of collecting unbiased samples. However, there are circumstances where unbiased samples are inefficient. For example, while unbiased samples provide accurate information about the body of a distribution, they provide very little information about the tails: the pitfall of estimating rare events we discussed in the previous subsection.

In circumstances such as studying infrequent events, it may be desirable to gather samples with a known sampling bias, i.e., with non-uniform sampling probabilities. By deliberately introducing a sampling bias towards the area of interest, more relevant samples can be gathered. During analysis of the data, each sample is weighted inversely to the probability that it is sampled. This yields unbiased estimates of the quantities of interest, even though the selection of the samples is biased. This approach is known as importance sampling [50].

A known bias can be introduced by choosing an appropriate definition of fix) in the Metropolis-Hastings equations presented in Section IV and altering the walk accordingly. Because the desired type of known bias depends on the focus of the research, we cannot exhaustively demonstrate through simulation that Metropolis-Hastings will operate correctly in a dynamic

461

environment for any . Our results show that it works well in the common case where unbiased samples are desired (i.e., fi(x) = fj,(y) for all x and y).

C. Sampling From Structured Systems

Throughout this paper, we have assumed an unstructured peer-to-peer network. Structured systems (also known as Dis-tributed Hash Tables or DHTs) should work just as well with random walks, provided links are still bidirectional. However, the structure of these systems often allows a more efficient technique.

In a typical DHT scheme, each peer has a randomly generated identifier. Peers form an overlay that actively maintains certain properties such that messages are efficiently routed to the peer "closest" to a target identifier. The exact properties and the defi-nition of "closest" vary, but the theme remains the same. In these systems, to select a peer at random, we may simply generate an identifier uniformly at random and find the peer closest to the identifier. Because peer identifiers are generated uniformly at random, we know they are uncorrelated with any other prop-erty. This technique is simple and effective, as long as there is little variation in the amount of identifier space that each peer is responsible for. We made use of this sampling technique in our study of the widely-deployed Kad DHT [51].

VII. CONCLUSIONS AND FUTURE WORK

This paper explores the problem of sampling representative peer properties in large and dynamic unstructured P2P systems. We show that the topological and temporal properties of P2P systems can lead to significant bias in collected samples. To collect unbiased samples, we present the Metropolized Random Walk with Backtracking and linking (MRWBL), a modification of the Metropolis-Hastings technique, which we developed into t h e - tool. Using both simulation and empirical evaluation, we show that MRWB can collect approximately unbiased samples of peer properties over a wide range of realistic peer dynamics and topological structures.

We are pursuing this work in the following directions. First, we are exploring improving sampling efficiency for uncommon events (such as in the tail of distributions) by introducing known bias, Second, we are studying the behavior of MRWBL under flash-crowd scenarios, where not only the properties of individual peers are changing, but the distribution of those properties is also rapidly evolving. Finally, we are developing additional plug-ins for -and using it in conjunction with other measurement tools to accurately characterize several properties of widely-deployed P2P systems.

REFERENCES

[1] I. Stoica, R. Morris, D. Liben-Nowell, D. R. Karger, M. F. Kaashoek, F. Dabek, and H. Balakrishnan, "Chord: A scalable peer-to-peer lookup protocol for Internet applications," IEEE/ACM Trans. Networking, vol. 11, no. 1, pp. 17-32, Feb. 2002.

[2] S. Ratnasamy, P. Francis, M. Handley, R. Karp, and S. Shenker, "A scalable content-addressable network," presented at the ACM SIGCOMM 2001, San Diego, CA.

[3] S. Saroiu, P. K. Gummadi, and S. D. Gribble, "Measuring and ana-lyzing the characteristics of Napster and Gnutella hosts," Multimedia Syst. J., vol. 9, no. 2, pp. 170-184, Aug. 2003.

[4] R. Bhagwan, S. Savage, and G. Voelker, "Understanding availability," presented at the 2003 Int. Workshop on Peer-to-Peer Systems, Berkeley, CA.

[5] D. Stutzbach and R. Rejaie, "Understanding churn in peer-to-peer net-

works," presented at the 2006 Internet Measurement Conf., Rio de Janeiro, Brazil.

[6] S. Chib and E. Greenberg, "Understanding the Metropolis-Hastings algorithm," The Americian Statistician, vol. 49, no. 4, pp. 327-335, Nov. 1995.

[7] W. Hastings, "Monte carlo sampling methods using Markov chains and their applications," Biometrika, vol. 57, pp. 97-109, 1970.

[8] N. Metropolis, A. Rosenbluth, M. Rosenbluth, A. Teller, and E. Teller, "Equations of state calculations by fast computing machines," J. Chem. Phys., vol. 21, pp. 1087-1092, 1953.

[9] A. Awan, R. A. Ferreira, S. Jagannathan, and A. Grama, "Distributed uniform sampling in unstructured peer-to-peer networks," presented at the 2006 Hawaii Int. Conf. System Sciences, Kauai, HI, Jan. 2006.

[10] Z. Bar-Yossef and M. Gurevich, "Random sampling from a search engine's index," presented at the 2006 WWW Conf., Edinburgh, Scotland.

[11] D. Stutzbach, R. Rejaie, N. Duffield, S. Sen, and W. Willinger, "Sampling techniques for large, dynamic graphs," presented at the 2006 Global Internet Symp., Barcelona, Spain, Apr. 2006. [12] D. Stutzbach and R. Rejaie, "Capturing accurate snapshots of the Gnutella network," in Proc. 2005 Global Internet Symp., Miami, FL, Mar. 2005, pp. 127-132. [13] B. Bollobás, "A probabilistic proof of an asymptotic formula for

the number of labelled regular graphs," Eur. J. Combinator., vol. 1, pp. 311-316, 1980.

[14] M. JerrumandA. Sinclair, "Fast uniform generation of regular graphs," Theoret. Comput. Sci., vol. 73, pp. 91-100, 1990.

[15] C. Cooper, M. Dyer, and C. Greenhill, "Sampling regular graphs and a peer-to-peer network," in Proc. Symp. Discrete Algorithms, 2005, pp. 980-988.

[16] V. Krishnamurthy, J. Sun, M. Faloutsos, and S. Tauro, "Sampling In-ternet topologies: How small can we go?," in Proc. 2003 Int. Conf. Internet Computing, Las Vegas, NV, Jun. 2003, pp. 577-580.

[17] V. Krishnamurthy, M. Faloutsos, M. Chrobak, L. Lao, and J.-H. C. G. Percus, "Reducing large Internet topologies for faster simulations," presented at the 2005 IFIP Networking Conf., Waterloo, Ontario, Canada, May 2005.

[18] M. P. H. Stumpf, C. Wiuf, and R. M. May, "Subnets of scale-free networks are not scale-free: Sampling properties of networks," Proc. National Academy of Sciences, vol. 102, no. 12, pp. 4221-4224, Mar. 2005.

[19] A. A. Tsay, W. S. Lovejoy, and D. R. Karger, "Random sampling in cut, flow, and network design problems," Math. Oper. Res., vol. 24, no. 2, pp. 383-413, Feb. 1999.

[20] A. Lakhina, J. W. Byers, M. Crovella, and P. Xie, "Sampling biases in IP topology measurements," presented at the IEEE INFOCOM 2003, San Francisco, CA.

[21] D. Achlioptas, A. Clauset, D. Kempe, and C. Moore, "On the bias of traceroute sampling; or, power-law degree distributions in regular graphs," presented at the 2005 Symp. Theory of Computing, Baltimore, MD, May 2005.

[22] P. Rusmevichientong, D. M. Pennock, S. Lawrence, and C. L. Giles, "Methods for sampling pages uniformly from the World Wide Web," in Proc. AAAI Fall Symp. Using Uncertainty Within Computation, 2001, pp. 121-128.

[23] M. Henzinger, A. Heydon, M. Mitzenmacher, and M. Najork, "On near-uniform URL sampling," in Proc. Int. World Wide Web Conf., May 2001, pp. 295-308.

[24] L. Lovász, "Random walks on graphs: A survey," Combinatorics: Paul Erdös is Eighty, vol. 2, pp. 1^6, 1993.

[25] Q. Lv, P. Cao, E. Cohen, K. Li, and S. Shenker, "Search and replication in unstructured peer-to-peer networks," presented at the 2002 Int. Conf. Supercomputing, New York, NY.

[26] Y. Chawathe, S. Ratnasamy, and L. Breslau, "Making Gnutella-like P2P systems scalable," presented at the ACM SIGCOMM 2003, Karl-sruhe, Germany.

[27] C. Gkantsidis, M. Mihail, and A. Saberi, "Random walks in peer-to-peer networks," presented at the IEEE INFOCOM 2004, Hong Kong.

[28] V. Vishnumurthy and P. Francis, "On heterogeneous overlay construc-tion and random node selection in unstructured P2P networks," pre-sented at the IEEE INFOCOM 2006, Barcelona, Spain, Apr. 2006.

[29] M. Salganik and D. Heckathorn, "Sampling and estimation in hidden populations using respondent-driven sampling," Sociological Method-ology, vol. 34, pp. 193-239, 2004.

[30] D. Heckathorn, "Respondent-driven sampling: a new approach to the study of hidden populations," Social Problems, vol. 44, no. 2, pp. 174-199, 1997.

[31] L. Goodman, "Snowball sampling," Ann. Math. Statist., vol. 32, pp. 148-170, 1961.

[32] E. Deaux and J. Callaghan, "Key informant versus self-report estimates of health behavior," Eval. Rev., vol. 9, pp. 365-368, 1985.

[33] J. Watters and P. Biernack, "Targeted sampling: Options for the study

462

of hidden populations," Annu. Rev. Sociol., vol. 12, pp. 401^129, 1989. [34] M. Salganik, "Variance estimation, design effects, and sample size cal-

culation for respondent-driven sampling,"J. Urban Health, vol. 83, pp. 98-112, 2006, Suppl. 1.

[35] J. Leskovec, J. Kleinberg, and C. Faloutsos, "Graphs over time: Den-sification laws, shrinking diameters and possible explanations," presented at the ACM SIGKDD Int. Conf. Knowledge Discovery and Data Mining, Chicago, IL, Aug. 2005.

[36] A. Bonato, "A survey of models of the web graph," Combinatorial and Algorithmic Aspects of Networking, pp. 159-172, 2004.

[37] J. Pouwelse, P. Garbacki, D. Epema, and H. Sips, "The BitTorrent P2P file-sharing system: Measurements and analysis," presented at the 2005 Int. Workshop on Peer-to-Peer Systems (IPTPS), Ithaca, NY.

[38] M. Izal, G. Urvoy-Keller, E. W. Biersack, P. A. Felber, A. A. Hamra, and L. Garces-Erice, "Dissecting BitTorrent: Five months in atorrent's lifetime," presented at the Passive and Active Measurement Conf. (PAM), Antibes Juan-les-Pins, France, Apr. 2004.

[39] D. Stutzbach, R. Rejaie, and S. Sen, "Characterizing unstructured overlay topologies in modern P2P file-sharing systems," in Proc. Internet Measurement Conf., Berkeley, CA, Oct. 2005, pp. 49-62.

[40] K. P. Gummadi, S. Saroiu, and S. D. Gribble, "King: Estimating la-

tency between arbitrary Internet end hosts," presented at the Internet Measurement Workshop, Marseille, France, Nov. 2002.

[41] D. Liben-Nowell, H. Balakrishnan, and D. Karger, "Analysis of the evolution of peer-to-peer systems," in Principles of Distributed Com-puting, Monterey, CA, Jul. 2002.

[42] S. Rhea, D. Geels, and J. Kubiatowicz, "Handling Churn in a DHT," in Proc. USENIX, 2004, pp. 127-140.

[43] J. Li, J. Stribling, F. Kaashoek, R. Morris, and T. Gil, "A performance vs. cost framework for evaluating DHT design tradeoffs under churn," presented at the IEEE INFOCOM 2005, Miami, FL.

[44] F. E. Bustamante and Y. Qiao, "Friendships that last: Peer lifespan and its role in P2P protocols," in Proc. Int. Workshop on Web Content Caching and Distribution, 2003, p. 2.

[45] S. Sen and J. Wang, "Analyzing peer-to-peer traffic across large networks," IEEE/ACM Trans. Networking, vol. 12, pp. 219-232, Apr. 2004.

[46] K. P. Gummadi, R. J. Dunn, S. Saroiu, S. D. Gribble, H. M. Levy, and J. Zahorjan, "Measurement, modeling, and analysis of a peer-to-peer file-sharing workload," in Proc. 19th ACM Symp. Operating Systems Principles (SOSP 2003), Bolton Landing, NY, 2003, pp. 314-329.

[47] D. Leonard, V. Rai, and D. Loguinov, "On lifetime-based node failure and stochastic resilience of decentralized peer-to-peer networks," pre-sented at the 2005 ACM SIGMETRICS, Banff, Alberta, Canada.

[48] H. Dämpfling, Gnutella Web Caching System: Version 2 Specifications Client Developers' Guide. Jun. 2003 [Online]. Available: http://www. gnucleus.com/gwebcache/newgwc.html

[49] P. Karbhari, M. Ammar, A. Dhamdhere, H. Raj, G. Riley, and E. Ze-gura, "Bootstrapping in Gnutella: A measurement study," presented at the 2004 Passive and Active Measurement Conf. (PAM) Antibes Juan-les-Pins, France, Apr. 2004.

[50] R. Srinivasan, Importance Sampling—Application in Communications and Detection. Berlin, Germany: Springer-Verlag, 2002.

[51] D. Stutzbach and R. Rejaie, "Improving lookup performance over a widely-deployed DHT," presented at the IEEE INFOCOM 2006, Barcelona, Spain, Apr. 2006.

463

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology, Sathyamangalam-638401

9-10 April 2010

464

Global Chaos Synchronization of Identical and

Different Chaotic Systems by Nonlinear Control V. Sundarapandian

#, R. Suresh

*

#R & D Department, Vel Tech Dr. RR & Dr. SR Technical University, Avadi, Chennai-600 062, Tamil Nadu [email protected]

* Department of Mathematics, Vel Tech Dr. RR & Dr. SR Technical University, Avadi, Chennai-600 062, Tamil Nadu [email protected]

Abstract— This paper investigates the global synchronization of

two identical and two different chaotic systems. The proposed

method by nonlinear control and Lyapunov functions is very

effective and convenient to synchronize two identical and two

different chaotic systems. The nonlinear control method has been

successfully applied to synchronize two identical Duffing

attractors and then to synchronize two different chaotic sytems,

namely Duffing attractor and Murali-Lakshmanan-Chua circuit.

Numerical simulations have been provided to illustrate the

validity of theoretical results.

Keywords— Chaos synchronization, Duffing attractor, Murali-

Lakshmanan-Chua circuit, nonlinear control, Lyapunov

function.

I. INTRODUCTION

Chaos synchronization is an important topic in the

nonlinear control systems and this has been developed and

studied extensively over the last few years [1-12]. Chaos

synchronization can be applied in the vast areas of physics,

engineering and biological science. The idea of synchronizing two identical chaotic systems was first introduced by Carroll

and Pecora [1-2]. Murali and Lakshmanan [3] derived

important results in secure communication using chaos

synchronization.

In most of the synchronization approaches, the master-slave

or drive-response formalism is used. If a particular chaotic

system is called the master or drive system and another

chaotic system is called the slave or response system, then the

idea of synchronization is to use the output of the master

system to control the slave system so that the output of the

response system tracks the output of the master system asymptotically. Since the seminal work by Carroll Pecora

[1,2], a variety of impressive approaches have been proposed

for the synchronization for the chaotic systems such as PC

method [1,2], sampled-data feedback synchronization method

[4], OGY method [5], adaptive design method [6,7], time-

delay feedback approach [8], backstepping design method

[9,10], sliding mode control method [11], active control

method [12], etc.

We organize this paper as follows. In Section 2, we give the

methodology of chaotic synchronization by nonlinear control

method. In Section 3, we discuss chaos synchronization of

two identical Duffing chaotic attractors. In Section 4, we

discuss the chaos synchronization of two different chaotic

systems with Murali-Chua circuit system as the master system

and Duffing attractor as the slave system. Finally,

conclusions and references close the paper.

II. THE PROBLEM STATEMENT AND OUR METHODOLOGY

Consider the chaotic system described by the dynamics

( )x Ax f x (1)

where nxR is the state of the system, A is the n n

matrix of the system parameter and : n nf R R is the

nonlinear part of the system. We consider the system (1) as

the master or drive system. As the slave or response system, we consider the following

chaotic system described by the dynamics

( )y By g y u (2)

where nyR is the state vector of the response system, B is

the n n matrix of the system parameter, : n ng R R is

the nonlinear part of the response system and nuR is the

controller of the response system. If A B and ,f g then

x and y are the states of two identical chaotic systems. If

A B and ,f g then x and y are the states of two

different chaotic systems.

The synchronization problem is to design a controller

,u which synchronizes the states of the drive system (1) and

the response system (2). If we define the synchronization error

as

,e y x (3)

then the synchronization error dynamics is obtained as

( ) ( ) .e By Ax g y f x u (4)

Thus, the synchronization problem is essentially to find a

controller u so as to stabilize the error dynamics (4), i.e.

lim ( ) 0.t

e t

(5)

We use Lyapunov function technique as our methodology.

We take as Lyapunov function

465

( ) TV e e Pe (6)

where P is a positive definite matrix. Note that ( )V e is a

positive definite function by construction.

We assume that the parameters of the master and response

systems are known and the states of both systems (1) and (2)

are measurable.

If we find a controller u such that

,TdVV e Qe

dt (7)

where Q is a negative definite matrix, then V is a negative

definite function.

Hence, by Lyapunov stability theory [13], the error

dynamics (4) is globally asymptotically stable and hence the

condition (5) will be satisfied for all initial conditions

(0) ne R .

III. SYNCHRONIZATION OF TWO IDENTICAL DUFFING SYSTEMS

In this section, we apply the nonlinear control technique for

the synchronization of two identical Duffing chaotic systems

described by

1 2

3

2 2 1 1 cos

x x

x x x x a t

(8)

which is the master or drive system and

1 2 1

3

2 2 1 1 2

( )

cos ( )

y y u t

y y y y a t u t

(9)

which is the slave or response system, where 1 2[ , ]Tu u u is

the nonlinear controller to be designed.

Our goal is to design the controller u for the global

synchronization of the two identical Duffing chaotic systems

(8) and (9).

The synchronization error e is defined by

1 1 1 2 2 2, e y x e y x (10)

The error dynamics is obtained as

1 2 1

3 3

2 2 1 1 1 2

e e u

e e e y x u

(11)

We choose the controller u defined by

1 2 1

3 3

2 1 1 1 2( )

u e e

u e y x e

(12)

where 0.

Then the error dynamics (11) simplifies to

1 1

2 2

e e

e e

(13)

Consider the candidate Lyapunov function

2 2

1 2

1( )

2V e e e (14)

which is clearly a positive definite function.

We find that

2 2

1 1 2 2 1 2( )V e e e e e e e

which is clearly a negative definite function.

Thus, by Lyapunov stability theory, the error dynamics (13) is

globally asumptotically stable.

3.1 Numerical Results:

The Matlab-simulink method is used to solve the systems of

differential equations.The initial state of the drive system are x1(0) = 2 and x2(0) = 2 and the initial state of the response

system are y1(0) = 5and y2(0) = 5.The results of the two

identical systems with active control are as shown in figures.

Figure:1 Chaotic Portrait-Duffing System

Figure:2 Time History of Duffing System

466

Figure:3 Controlled Unified Chaotic System

Figure:4 Controlled Unified Chaotic System

IV. SYNCHRONIZATION OF TWO DIFFERENT CHAOTIC SYSTEMS

In this section, we apply the nonlinear control technique for the synchronization of two different chaotic systems described

by Murali-Lakshmanan cha circuit as drive system and

duffing system as response system.

Consider the Murali-Lakshmanan Chua circuit

(14)

Where

x1, x2 are the state variables. From equation (8) and (12), the

following error system equation can be obtained,

1 1 1 2 2 2, e y x e y x

The error dynamics is obtained as

(15)

We choose the controller u defined by

+

(16)

where 0.

Consider the candidate Lyapunov function

2 2

1 2

1( )

2V e e e

which is clearly a positive definite function.

We find that

2 2

1 1 2 2 1 2( )V e e e e e e e

which is clearly a negative definite function.

Thus, by Lyapunov stability theory, the error dynamics (15) is

globally asumptotically stable.

4.1 Numerical Results:

The Matlab-simulink method is used to solve the systems of

differential equations.The initial state of the drive system are

x1(0) = 2 and x2(0) = 2 and the initial state of the response

system are y1(0) = 5and y2(0) = 5.The results of the two identical systems with active control are as shown in figures.

Figure:5 Chaotic Portrait-Murali-Lakshmanan

Circuit

Figure:6 Time History of Murali-Lakshmanan

Circuit

467

Figure:7 Controlled Two different Chaotic

System

Figure:8 Controlled Two different Chaotic

System

V. CONCLUSIONS

In this paper, modification based on Lyapunov stability theory

to design a non linear to design a non linear controller is

proposed to synchronize two identical chaotic system and two

different chaotic systems. Numerical simulations are also

given to validate the proposed synchronization approach.

The simulation results show that the states of two identical

duffing system are globally exponentially synchronized. For two different chaotic systems, the Murali -Lakshmanan chua

circuit system is forced to trace the duffing system and the

states of two systems become ultimately the same. Since the

Lyapunov exponents are not required for the calculation, this

method is effective and convenient to synchronize two

identical systems and two different chaotic systems.

References:

[1].L.M.Pecora, T.L.Carroll, Synchronization in chaotic

system Physics Review Letter, 64 (1990) 821–824.

[2].G.Chen,X.Dong,Form Chaos to Order: Methodologies,

Perspective and Application World Scientifics,

Singapore,(1998)

[3] .J.A.Lu, J.Xie, J.Lu,S.H. Chen, Control chaos in transition

system using sampled-data feedback, Applied Mathe-

matics and Mechanics 24, 11 (2003) 1309–1315.

[4].M.Itoh, T.Yang, L.O.Chua, Conditions for impulsive

synchronization of chaotic and hyper chaotic system,

International journal of Bifurcation and Chaos 11, 2 (2000). [5].K.Y.Lian, P.Liu,T-S.Chiang, C-S.Chiu, Adaptive

synchronization for chaotic system via a scalar signal IEEE

Transaction on circuit and system-I, 49 (2002)17–27.

[6] Z.M.Ge,Y.S.Chen, Adaptive synchronization of

unidirectional and mutual coupled chaotic system Chaos,

Solitons and Fractals, 26 (2005)881–888.

[7] J.Q.Fang,Y.Hong,G.Chen, A switching manifold approach

to chaos synchronization, Physical Review E, 59(1999)2523.

[8] X.Tan,J.Zhang, Y.Yang, Synchronizing chaotic system

using back stepping design Chaos, Solitons and Fractals,16

(2003)37–45. [9] H.T.Yau, Design of adaptive sliding mode controller for

chaos synchronization with uncertainties Chaos, Solitons and

Fractals, 22 (2004) 341–347.

[10] C.C.Wang, J.P.Su, A new adaptive variable structure

control for chaotic synchronization and secure

communication, Chaos, Solitons and Fractals, 20 (2004) 967–

977.

[11] X.Yu,Y.Song, Chaos synchronization via controlling

partial state of chaotic system International journal of

Bifurcation and Chaos , 11 (2001),1737–1741.

[12] X.F.Wang,Z.Q.Wang, G.R.Chen, A new criterion for synchronization of coupled chaotic oscillators with application

to chua’s circuits, International journal of Bifurcation and

Chaos , 9 (1999),1285–1298.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

468

Reducing Interference by Using Synthesized Hopping Techniques By

A.Priya, B.V.Sri Laxmi Priya

[email protected], [email protected]

ELECTRONICS AND COMMUNICATION ENGINEERING

K.L.N.COLLEGE OF ENGINEERING

ABSTRACT:

Synthesized frequency hopping allows the

mobile station to change its operating frequency

from a time slot to another in a GSM network. This

technique cannot be applicable to BCCH

(Broadcast Control Channels) which must transmit

on a fixed frequency in order to improve RX quality

level. Synthesized frequency hopping can be used to

improve the quality of the network. It also is used to

increase the capacity of the network thereby

reducing the number of sites required for capacity.

Both simulation and measurements reveals for the

better results.

Keywords: RX quality, Drop call rate, call

successful rate.

1. INTRODUCTION

GSM introduced antenna diversity and slow

frequency hopping to improve the transmission quality,

as described in GSM .while antenna diversity combats

multipath fading only, frequency hopping additionally

averages the effects of interference. This allows a

tighter frequency reuse, thus, increasing network

capacity through carrier updating while maintaining a

given quality of service (QoS).Alternatively if no

carrier upgrading is performed, frequency hopping

allows for QoS and network performance to be

improved while maintaining capacity.

Frequency hopping was introduced in GSM

specifications in order to make use of the two features

it provides for the transmission quality improvement,

namely frequency diversity and interference diversity.

In urban environments, especially, but not

only, radio signals reach the receiver on different paths

after multiple reflections and diffraction resulting in

fading effects. The received signals levels vary as a

function of time, operation frequency and receiver

location. Slow MSs may stay in a fading notch for a

long period of time and suffer from a severe loss of

information. Frequency hopping combats multipath

fading because different frequencies experience

different fading and, thus, MSs experience different

fadings at each burst and do not stay in a deep

minimum for a long period of time over which the

codeword is spread with interleaving. Therefore SFH

improves the transmission quality because too long

fading periods are avoided. Since fast moving mobiles

do not stay long periods in deep fading holes, they do

not suffer severely from this type of fading.

469

The improvement results in increased average

power density under fading conditions at the receiver

site and therefore in improved quality both in uplink

and downlink direction as compared to the non-

hopping configuration. This is called frequency

diversity.

Without SFH some receivers (be they MS or

BTS) experience strong interference, while receivers

operating on other frequencies, experience light

interference or not interfered at all. Some interference

could last a long period of time when sources are fixed

interferes incorrectly radiating in the GSM band or

could be permanent such as those produced by BCCH

frequencies on the downlink. With random SFH the

interfering scenario changes from a time slot to

another, due to its own hopping and to the uncorrelated

hopping of potential interferers. Thus all receivers

experience an averaged level of interference. This is

called interference averaging or interference diversity.

Frequency diversity and interference diversity

result in an improved network quality. Thus, the

implementation of SFH in a mature network is a means

to improve the network service quality. Network

capacity increase can be performed in a second step,

after the introduction of SFH. Due to the fact that the

same QoS level can be reached with a lower carrier-to-

interference(C/I) ratio, a SFH network can be planned

with a tighter reuse cluster size. Therefore more

carriers per site can be used.

This paper presents principles and practical

results of SFH implantation in the most congested area

of a specific GSM network covering an urban

environment.

This paper is organized as follows. Then the

network topology and its quality parameters before the

SFH introduction. Then analyze the effects of SFH and

the QoS improvement as revealed by the simulation

and checked through measurements. Also, we present

the necessary tuning of some network parameters after

the SFH introduction. Finally we summarize our

results.

2. NETWORK TOPOLOGY

The above mentioned area is located in a flat

region surrounded by hilly terrain. Because of this no

regular frequency pattern could be used. The base

transceiver stations (BTS) are equipped with minimum

of two transceivers (TRX) and two-sector antennas,

while some of them-with three-sector antennas. The

network is assigned 22 GSM radio channels in two

disjoint frequency bands.

The existing QoS in the network is particularly

good and does not require

Immediate action. For instance call drop rate is

1.54%,call setup successful rate is 98.59%,and the

MOS (mean opinion score) for the voice transmissions

is on average 4.09 and 4.13 on the uplink and the

downlink, respectively. But the forecasted offered

traffic is 20% greater than the present one. This value

was obtained from the traffic history by using a long-

term forecasting technique based on linear regression

without seasonal pattern. Repeated simulations

revealed that the increase in network capacity by using

traditional methods of antenna tilting, and radio

470

planning optimization could not cope with the

forecasted offered traffic increase.

The chosen solution was to introduce the

Synthesized Frequency Hopping in the existing GSM

network in order to face the forecasted increase in the

needed capacity as well as to obtain a better QoS. The

new frequency plan had to offer flexibility in the near

future for network reconfigurations and, also, to take

into account the enlargement of the assigned frequency

band with two new radio channels.

Special care has been taken in the BCCH

channels distribution among network cells in order to

advantageously use the existence of the two disjoint

frequency bands and keep as big as possible guard

between their band and the hopping sequences as one

(at least one radio channel).

Another important issue in the design is the selection

of reuse pattern (cluster size) as it dramatically

influences the maximum allowable limit for the RF

load factor (defined as the ratio between the number of

hopping TRXs and the number of frequencies in a

hopping sequence).By maintaining the RF load factor

under the maximum allowable limit one fulfills the

intra-site constraint: no collision among the same site

sectors for synchronized hopping sequences. The

maximum limit is 50% for a reuse pattern of 1/3 and

only 16.6%for a reuse pattern of 1.

By using 12 radio channels for BCCH, the remaining

10 channels as traffic channels (mobile allocation-MA)

yield an RF load factor of 10% for an all 3-TRX-cell

site network and of 20% for an all 2-TRX-cell site

network, respectively. For the network topology under

consideration with only several 3-TRX cell site this

means that the introduction of SHF is only possible by

upgrading all the sites to 3-TRX (an expensive

solution) or by using a reuse pattern of 1/3 (with

penalties on the network capacity).Considering the two

new radio channels the RF load factor becomes 16.6%

even for an all 2-TRX-cell site network. So the reuse

pattern of 1 is possible and this pattern was chosen.

With the reuse pattern of 1 the percentage of

overlapping areas of neighboring cells become a very

important issue, as collisions are possible in these

regions. By extensive simulations and drive tests

measurements the total overlapping area was reduced

to a minimum through appropriate tilting of the

antennas.

3. SIMULATION RESULTS AND

MEASUREMENTS

Table 1 comparatively shows the main QoS

parameters of the network before and after SFH

introduction, respectively. The data were obtained after

extensive simulations with a professional simulation

tool, also used for the network initial planning.

QUALITY INDICATOR NO SFH SFH

CALL DROP RATE (%) 1.54 1.04

CALL SETUP RATE (%) 98.59 98.8

HANDOVER EXTERN (%) 99.53 97.95

RX QUAL FULL

STD

+/-1.25

+/-1.16

MIN 0 0

MAX 7 7

471

Due to frequency diversity, a noticeable gain

in the call drop rate was obtained. Further decrease of

the call drop rate and bad quality Erlang was obtained

by upgrading of 2TRX cells with 1TRX.Voice quality

(average value, standard deviation, minimum and

maximum values) is better in the case of SFH for the

Uplink, and undesired result is proved by the data

measurements, which offered an explanation too. A

solution to counteract this behavior is presented later in

this section.

The simulated result was checked against the measured

data reports obtained by extensive drive tests.

So, it was noted that the percentage of the RX quality

Rate and call success rate significantly increased after

the introduction of SFH.Fig represents the call setup

successful rate for 10 successive days, day no.4 being

the starting date of SFH.Similary; Fig represents the

distribution of Call Drop Rate and reveals a decrease of

this indicator value. Both of these parameters are main

components of the perceived QoS, so it is very clear

that SFH brings a significant improvement in the QoS

value and the network capacity increases, as well.

DLQ

75.0

80.0

85.0

90.0

95.0

100.0

1/3/2

010

1/5/2

010

1/7/2

010

1/9/2

010

1/11/2

010

1/13/2

010

1/15/2

010

1/17/2

010

1/19/2

010

Date

DLQ

Impr

ovem

ent

DLQ

An impact of SFH on the Handovers Reasons

Distribution was also noticed, with the increased

importance of both the Uplink and Downlink Level,

and quite the same importance of the acquisition of the

Better cell.

Drop call

90.091.092.093.094.095.096.097.098.099.0

1/3/2

010

1/5/2

010

1/7/2

010

1/9/2

010

1/11/2

010

1/13/2

010

1/15/2

010

1/17/2

010

1/19/2

010

Date

Dro

p ca

ll Im

prov

emen

t

Drop call

The drive tests confirmed the simulation results of the

Downlink quality decrease after the SFH introduction

and, also, revealed that the downlink quality

significantly decreases just before handover triggering

in the overlapping regions of the neighboring cells

where the probability of collision is very high.

In order to maintain a good average quality of

the link the handover margin between cells was

reduced from 5 dB to 0.This reduction is made only for

a short period of time (15 to 30 seconds) just before the

handover triggering in order to avoid the ping-pong

effect. Thus, the handover is triggered by the

acquisition of a Better Cell process and the decrease of

the link quality is avoided.

Statistics of the measured data reveal, also,

another positive effect of SFH implementation (as a

result of congestion decreasing): the increase of

Handover Outgoing Successful Rate (Outgoing Inter

BSC/Outgoing Intra BSC Handover Rate) and

Handover Incoming Successful Rate (Incoming Inter

BSC/Incoming Intra BSC Handover Rate).

472

After a power control mechanism optimization

the call drop rate further decreased from 1.04% to

0.8%.

SUMMARY

The benefits of implementing the slow frequency

Hopping in an existing GSM network are underlined, they

are proved through simulation, and, finally, checked by

measurements with protocol analyzers and by extensive

drive tests. Practical solutions to combat adverse side effects

are presented; too.It was proved that calls drop rate

decreases and call setup successful rate and the RX quality

level increases and hence overall network capacity increases.

REFERENCES

1. GSM, GPRS and EDGE performance by Timo Halonen.

2. GSM switching, services and protocols by J.Eberspacher.

3. Multiple access protocols for mobile communications by

Alex Brand.

4. Performance enhancements in a frequency hopping GSM

network by Thomas Toftegaard Nielsen and Jeroen Wigard.

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

473

`

`ABSTRACT:

This paper primarily aims at the new

technique of video image processing used

to solve problems associated with the real-

time road traffic control systems. There is

a growing demand for road traffic data of

all kinds. Increasing congestion problems

and problems associated with existing

detectors spawned an interest in such new

vehicle detection technologies. But the

systems have difficulties with congestion,

shadows and lighting transitions.

Problem concerning any practical

image processing application to road

traffic is the fact that real world images are

to be processed in real time. Various

algorithms, mainly based on back ground

techniques, have been developed for this

purposes since back ground based

algorithms are very sensitive to ambient

lighting conditions, they have not yielded

the expected results. So a real-time image

tracking approach using edged detection

techniques was developed for detecting

vehicles under these trouble-posing

conditions.

This paper will give a general

overview of the image processing

technique used in analysis of video

images, problems associated with it,

methods of vehicle detection and tracking,

pre-processing techniques and the paper

also presents the real-time image

processing technique used to measure

traffic queue parameters.

Finally, general comments will be

made regarding the extensibility of the

method and the scope of the new queue

detection algorithm will be discussed.

Real Time Image Processing

Applied To Traffic –Queue Detection Algorithm

r.ramya, p.saranya

[email protected]

[email protected]

474

AIM:

Fig 1

In normal traffic light system, the

traffic signals are changed in the

predetermined time. This method is

suitable when the no of vehicles in all

sides of a junction are same (fig 1). But

when there is an unequal no of vehicles in

the queues (fig 2) this method is

inefficient. So we go for the image

processing techniques using queue

detection algorithm to change the traffic

signals.

Fig 2

1. INTRODUCTION

Increasing demand for road traffic

data of all sorts

Variation of parameters in real-world

traffic

Aimed to measure queue parameters

accurately

Algorithm has two operations :

vehicle detection and motion detection

Operations applied to profiles

consisting sub-profiles to detect queue

Motion detection is based on applying

a differencing technique on the profiles of

the images along the road

The vehicle detection is based on

applying edge detection on these profiles

2. IMAGE PROCESSING APPLIED

TO TRAFFIC

2.1 NEED FOR PROCESSING OF

TRAFFIC DATA

Traffic surveillance and control, traffic

management, road safety and development

of transport policy.

2.2 TRAFFIC PARAMETERS

MEASURABLE

Traffic volumes, Speed, Headways, Inter-

vehicle gaps, Vehicle classification, Origin

and destination of traffic, Junction turning.

3. STAGES OF IMAGE ANALYSIS

475

Image sensors used: Improved

vision cameras: automatic gain control,

low SNR

ADC Conversion: Analog video

signal received from video camera is

converted to digital/binary form for

processing

Pre-processing: High SNR of the

camera output reduces the quantity of

processing enormous data flow.

To cope with this, two methods are

proposed:

1. Analyze data in real time -

uneconomical

2. Stores all data and analyses off-line at

low speed.

Pipeline Preprocessing does this job

3.1 STAGES IN PIPELINE

PREPROCESSING

(1) Spatial Averaging – contiguous pixels

are averaged (convolution)

(2) Subtraction of background scene from

incoming picture.

(3) Threshold – Large diff.s are true „1‟,

small diff.s are false „0‟

(4) Data Compression – reduces

resulting data.

(5) Block buffering – collects data into

blocks.

(6) Tape Interface – blocks are loaded onto

a digital cassette recorder

Preprocessed picture is submitted to

processor as 2-D array of no.s.

4. METHODS OF VEHICLE

DETECTION:

Background frame differencing: -

grey-value intensity reference image

Inter-frame differencing: -

incoming frame itself becomes the

background for the following frame

Segmentation and classification: -

Sub division of an image into its

constituent parts depending on the context

4.1 QUEUE DETECTION

ALGORITHM

Approach described here is a spatial-

domain technique to detect queue -

implemented in real-time using low-cost

system

For this purpose two different

algorithms have been used,

Motion detection operation, Vehicle

detection operation

Motion detection is first – as in this

case vehicle detection mostly gives

positive result, while in reality, there may

not be any queue at all.

Applying this scheme further reduces

computation time.

5. MOTION DETECTION

OPERATION BLOCK DIAGRAM:

476

a) Differencing two consecutive frames.

b) Histogram of the key region parts of the

frames is analyzed by comparing with the

threshold value.

c) Key region should be at least 3-pixel-

wide profile of the image along the road.

d) A median filtering operation is firstly

applied to the key region (profile) of each

frame and one-pixel-wide profile is

extracted.

e) Difference of two profiles is compared

to detect for motion.

f) When there is motion, the differences of

the profiles are larger than the case when

there is no motion. The motion can be

detected by selecting a threshold value.

5.1 MOTION DETECTION

ALGORITHM THEORY BEHIND

The profile along the road is

divided into a number of smaller profiles

(sub-profiles)

The sizes of the sub-profiles are

reduced by the distance from the front of

the camera.

Transformation causes the equal

physical distances to unequal distances

according to the camera parameters.

Knowing coordinates of any 6

reference points of the real-world

condition and the coordinates of their

corresponding images, the camera

parameters (a11, a12…a33) are calculated.

The operations are simplified for flat plane

traffic scene - (Zo=0).

Solving matrix equation for camera parameters

Above equation is used to reduce

the sizes of the sub-profiles - each sub

profile represents an equal physical

distance.

No. of sub profiles depend on the

resolution and accuracy required

6. VEHICLE DETECTION

ALGORITHM

Following the application of the

motion detection operation, a vehicle

detection operation is applied on the

profile of the unprocessed image.

To implement the algorithm in real

time, two strategies are often applied: key

region processing and simple algorithms.

477

Most of the vehicle detection

algorithms developed so far are based on a

background differencing technique, which

is sensitive to variations of ambient

lighting.

The method used here is based on

applying edge detector operators to a

profile of the image – Edges are less

sensitive to the variation of ambient

lighting and are used in full frame

applications (detection).

Edge detectors consisting of separable

medium filtering and morphological

operators, SMED (Separable

Morphological Edge Detector) are applied

to the key regions of the image. (The

SMED approach is applied (f) to each sub-

profile of the image and the histogram of

each sub-profile is processed by selecting

Dynamic left-limit value and a threshold

value to detect vehicles.

SMED has lower computational

requirement while having comparable

performance to other morphological

operators

SMED can detect edges at different

angles, while other morphological

operators are unable to detect all kinds of

edges.

7. LEFT-LIMIT SELECTION

PROGRAM

This program selects a grey value

from the histogram of the window, where

there are approx. zero edge points above

this grey value.

When the window contains an

object, the left-limit of the histogram shifts

towards the maximum grey value,

otherwise it shifts towards the origin.

This process is repeated for a large

no. of frames(100),and the minimum of

the left-limit of these frames is selected as

the left-limit for the next frame

8. THRESHOLD SELECTION

PROGRAM

The no. of edge points greater than

the left limit grey value of each window is

extracted for a large no. of frames (200) to

get enough parameters below and above a

proper threshold value.

This nos. are used to create a

histogram where it‟s horizontal and

vertical axes correspond to the no. of edge

points greater than left limit and the

frequency of repetition of these numbers

for a period of operation of the algorithm

(200 frames).

This histogram is smoothed using a

median filter and we expect to get two

peaks in the resulted diagram, one peak

related to the frames passing a vehicle and

the other related to the frames without

vehicles for that window.

We use statistical approach based

on selecting a point on the horizontal axis,

where the sum of the entropy of the points

478

above and below this point is maximum.

This point is selected as the threshold

value for the next period.

9. TRAFFIC MOVEMENTS AT

JUNCTIONS (TMJ)

Measuring traffic movements of

vehicles at junctions such as number of

vehicles turning in a different direction

(left, right, and straight) is very important

for the analysis of cross-section traffic

conditions and adjusting traffic lights.

Previous research work for the

TMJ parameter is based on a full-frame

approach, which requires more computing

power and, thus, is not suitable for real-

time applications.

We use a method based on counting

vehicles at the key regions of the junctions

by using the vehicle-detection method.

The first step to measure the TMJ

parameters using the key region method is

to cover the boundary of the junction by a

polygon in such a way that all the entry

and exit paths of the junction cross the

polygon. However, the polygon should not

cover the pedestrian marked lines. This

step is shown in the figure given below.

The second step of the algorithm is to

define a minimum numbers of key regions

inside the boundary of the polygon,

covering the junction.

These key regions are used for

detecting vehicles entering and exiting the

junction, based on first vehicle –in first-

vehicle-out logic.

Following the application of the

vehicle detection on each profile, a status

vector is created for each window in each

frame.

If a vehicle is detected in a window, a

“one” is inserted on its corresponding

status vector, otherwise, a “Zero” is

inserted.

Now by analyzing the status vector of each

window, the TMJ parameters are

calculated for each path of the junction

10. RESULTS AND DISCUSSIONS

The main queue parameters we

were interested in identifying were the

length of the queue, the period of

occurrence and the slope of the occurrence

of the queue behind the traffic lights.

To implement the algorithm in

real-time, it was decided that the vehicle

detection operation should only be used in

a sub-profile where we expect the queue

will be extended. The procedure is as

follows:

479

11. OUTPUT:

SOFTWARE USED: MATLAB

TRAFFIC IN A ROAD:

SIDE 1:

SIDE 2:

SIDE 3:

480

SIDE 4:

BACKGROUND:

IMAGE IS CONVERTED INTO RGB

TO GRAY:

The photos of all the four sides are

taken at regular intervals ( depending upon

traffic say 5 sec), and these are stored in

the computers, which is placed in the

signal and these images are processed

using the following code..

MATLAB CODE:

clc;

clear all;

close all;

i1=imread('D:\DINESH\DINESH

WORK\car1.png');

i1=rgb2gray(i);

i1=imresize(i1,[256 256]);

i2=imread('D:\DINESH\DINESH

WORK\back.png');

i2=rgb2gray(i2);

i2=imresize(i2,[256 256]);

i3=imsubtract(i1,i2);i3=rgb2gray(i3);

% these lines are used to read the

background pictures

s1=imread('D:\GI\GI WORK\back1.png');

s2=imread('D:\GI\GI WORK\back2.png');

s3=imread('D:\GI\GI WORK\back3.png');

s4=imread('D:\GI\GI WORK\back4.png');

% these lines are used to read the taken

traffic pictures

t1=imread('D:\GI\GI

WORK\image1.png');

t2=imread('D:\GI\GI

WORK\image2.png');

t3=imread('D:\GI\GI

WORK\image3.png');

t4=imread('D:\GI\GI

WORK\image4.png');

Figuire,imshow(i1);

figure,imshow(i2);

figure,imshow(i3);

side1=imsubtract(t1-s1);

481

side2=imsubtract(t2-s2);

side3=imsubtract(t3-s3);

side4=imsubtract(t4-s4);

w1=imwhite(side1);

w2=imwhite(side1);

w3=imwhite(side1);

w4=imwhite(side1);

whitebox=[w1 w2 w3 w4];

time=[];

ontime=[];

% we assume that a vehicle takes 10 sec to

move

for i=1:4

time(i)=time(i-1)+10*whitebox(i);

ontime(i)=10*whitebox(i)%it counts the

time for four sides and it is stored in an

array.we give it to the traffic light through

output ports.

end

%function for converint it into white

boxes

function k=imwhite(image)

for i=1:256

for j=1:256

if(image(i,j)>0)

image(i,j)=255;

end

end

end

k=count(image,white);

IMAGE AFTER SUBTRACTION

OUTPUT

Whitebox=[5 3 1 0]

Time=[ 50 80 90 90]

Ontime=[ 50 30 10 0]

EXPLANATION

Side1=5 vehicles

Side1=3 vehicles

Side1=1 vehicles

Side1=0 vehicles

Each takes 10 seconds (assume)

12. CONCLUSIONS

Algorithm measuring basic queue

parameters such as period of occurrence

between queues, the length and slope of

occurrence has been discussed. The

algorithm uses a recent technique by

applying simple but effective operations.

In order to reduce computation

time motion detection operation is applied

on all sub-profiles while the vehicle

detection operation is only used when it is

necessary.

482

The vehicle detection operation is a less

sensitive edge-based technique. The

threshold selection is done

dynamically to reduce the effects

of variations of lighting.

The measurement algorithm has

been applied to traffic scenes with

different lighting conditions.

Queue length measurement showed

95% accuracy at maximum. Error is due to

objects located far from camera and can be

reduced to some extent by reducing the

size of the sub-profiles.

13. REFERENCES

Digital Image Processing by Rafael

C.Gonzalez and Richard Elwood‟s.

Hose, N.: „Computer Image

Processing in Traffic Engineering‟.

Bourke, A., and Bell, M.G.H.: „Queue

detection and congestion monitoring using

mage processing‟, Traffic Egg. And

Control.

Traffic Queue Length Measurement

Using an Image Processing Sensor by

Misacts Higashikobo, Toshio Honeoye and

Kowhai Takeout.

A Real-time Computer Vision System for

Vehicle Tracking and Traffic Surveillance

by Benjamin Coffman (corresponding

author).

Proceedings of the Third National Conference on RTICT 2010

Bannari Amman Institute of Technology ,Sathyamangalam- 638401

9-10 April 2010

483

A FUZZY LOGIC CLUSTERING

TECHNIQUE TO ANALYSE HOLY

GROUND WATER

Abstract-- A fuzzy set theoretic approach

has been developed to study the potable nature of

the holy groundwater samples in summer and

winter by clustering method using equivalence

relation. The physico-chemical parameters viz.,

pH, Salinity, TDS, CH, MH, TH, Chloride and

Fluoride are considered as attributes to develop the

clusters. Based on the WHO recommendations for

the water quality parameters, the linguistic

approach has been developed for the 22 holy

groundwater samples in this study. Normalized

eucilidean distance chosen for this study, measures

the deviation of the determined quality parameters

for any two holy groundwater samples. In the

present paper, the seasonal changes in the quality

of the water samples among the clusters at various

rational alpha cuts are derived. Key words: fuzzy set, potable nature,

groundwater, fuzzy cluster, alpha cuts

1.Introduction

The physico-chemical quality of

drinking water becomes as important as its

availability. The water quality parameters with

desirable, acceptable and not acceptable values,

recommended by World Health Organization

(WHO) create awareness among the public,

which enroot them towards the removal

techniques []. From the WHO guideline values,

the groundwater samples are categorized for

quality with respect to each water quality

parameter. This may lead to a decision but with ambiguity, because it is derived out of only one

parameter. So far, numerous research works

have been carried out in the determination of

water quality using fuzzy synthetic evaluation

(Lu et al., 1999; Jiahhua et al., 2008; Singh et al.,

2008), fuzzy process capability (Kahraman and

Kaya, 2009), fuzzy clustering and pattern

recognition method (Chuntao et al., 2008; Yuan

et al., 2006) and fuzzy logic approach

(Muhammetoglu and Yardimici, 2006). In the

present work, a new focus has been attempted

using fuzzy equivalence relation to arrive at non-

overlapping clusters of 22 groundwater samples by considering various agreement levels (alpha

cuts).

2 Study objectives and methods

The objective of the present study is to

obtain non-overlapping clusters of twenty two holy groundwater samples (Table 1) of Rameswaram

temple in the summer and winter seasons based on the water quality parameters viz., pH, Salinity, Total Dissolved Solids (TDS), Calcium Hardness(CH), Magnesium Hardness(MH), Total Hardness(TH), Chloride(Cl) and Fluoride(F).

2.1 Study area

Rameswaram is located around an

intersection of the 9º28’North Latitude and 79º3’East Longitude with an average elevation of 10 meters above the MSL, covering an area of 61.8 sq.kms and bearing a population of about 38,000, as on September 2007. 120 This Indian Island having connection with main land assumes a shape of conch, is a Taluk with 1 Firka, 2 Revenue villages and 31 Hamlets. Climate prevails with a minimum temperature of 25ºC in

winter and a maximum of 36ºC in summer. The average rainfall is 813mm.

S.Dharani, Student, Department of Information Technology, Thiagarajar College of Engineering, Madurai-625015, [email protected]

D.Anugraha, Student, Department of Computer Science and Engineering, Thiagarajar College of Engineering, Madurai-625015, [email protected]

484

Table 1 Name of the holy groundwater samples

Sample No.

Name of the holy groundwater

1 Mahalakshmi

2 Savithri

3 Gayathri

4 Saraswathi

5 Sangu

6 Sarkarai

7 Sethumadhava

8 Nala

9 Neela

10 Kavaya

11 Kavacha

12 Kandhamadhana

13 Bramahathi

14 Ganga

15 Yamuna

16 Gaya

17 Sarwa

18 Siva

19 Sathyamrudham

20 Surya

21 Chandra

22 Kodi

2.2 Clustering of groundwater samples

Let S1, S2 ......S22 were the twenty two groundwater samples of Rameswaram temple are considered for clustering based on the criteria C1, C2..........C8 viz., pH, Salinity, TDS, CH, MH, TH, Cl and F. Linguistic terms such as Excellent, Fairly Excellent, Good, Fairly Good and Poor were assigned

to the chosen water samples with respect to the recommendations of the World Health Organization (WHO).

2.3 Membership functions and fuzzy relations Membership function (µ) is a critical measure which represents numerically the degrees of elements belonging to a set. Distance measure is a

term that describes a difference between fuzzy sets and can be considered as a dual concept of similarity measure. The linguistic terms (Table 2) were converted into fuzzy numbers (membership functions) using probability technique. Using the fuzzy numbers, the Normalized Euclidean distance (eqn.1)

was used to obtain similarity measures, which was found by subtracting the distance measure from 1 using MATLAB (version 7). The obtained similarity measure possesses tolerance relation (R) between the undertaken groundwater samples (Tables 3 and 4). A fuzzy relation (R) is said to be fuzzy tolerance relation if R is reflexive [if µR(x,x)=1, for every x Є X] and

symmetric [if µR[x,y)=µR(y,x) for every x,y Є X]. A fuzzy relation, R is said to be fuzzy equivalence relation RE, if R is fuzzy tolerance relation and transitive closure [if µR satisfies µR(x,z)≥ minµR(x,y), µR(y,z)for every x,y,z Є X. An equivalence relation (RE) in Tables 5 and 6 was

determined from the computed tolerance relation by the following algorithm using Visual C + + on windows platform.

Step 1: R’ =R o (RR)

Step 2: If R’ ≠R, make R=R’ and go to step 1.

Step 3: Stop: R’ = RE

In the above o is the max-min composition and min-

max composition of fuzzy relations and is the

standard fuzzy union. By the consideration of reasonable alpha cuts (x/µ(x) > α, for some x ЄX), the groundwater samples were clustered in the non-overlapping nature. 3. Results and discussion

Clusters starts forming at the AL of.925in sum&.930in win(max-min). Similarly clusters starts forming at the AL of.905in sum&.910in win(min-max).

Samples which groups in sum 10,4 & 20,21 hav similar characteristics in both the algor.

Samples 10,12 &20,21 hav similar characteristics in both the algor.

Sample “7” sethumadha have Unique

characteristics in all the algor in all seasons. Sample “8” have also 80% unique

characteristics when we starts increasing the AL.

From these results, it is observed that, even at 0.05%

difference in AL reveal some changes in the water quality characteristics.

485

The seasonal influence can also be witnessed from the merging of different clusters at 90% AL.

The above cluster separation from summer to winter indicates the change in the water quality parameters with respect to dilution influenced by seasons. Entry of foreign substances in winter seasons have changed the properties of water in season Also it was observed that the samples 10 and 12 from two different clusters in summer grouped into a

particular cluster in winter. This shows that the water quality parameters of both the samples falls within the standards fixed in this study as per WHO even after the seasonal influence. In winter, the number of clusters at 0.9 AL and 0.85 AL were computed to be 6 and 3 respectively and this

highlights the reduction of clusters indicates the role of dilution which causes recharging of groundwater and setting the water quality parameters with respect to the fixed standards as per WHO. This reflects that these two samples possess the quality characteristics within a particular limit, fixed in the study in accordance with WHO recommendations.

The identity of a single membered cluster containing a groundwater sample No. 7 (Sethumadhava) at both the agreement levels in two different seasons, ascertains the distinct quality of the water sample.

Evidently, the present observation was in

agreement with our earlier work (Sivasankar and

Ramachandramoorthy, 2008).

EQUIVALENCE MATRIX IN WINTER

EQUIVALENCE MATRIX IN SUMMER

486

NON OVERLAPPING CLUSTERS FOR SUITABLE ALPHA CUTS IN SUMMER AND WINTER (MAX_MIN)

Max-Min(ror final)

Alph

a

cuts

SUMMER WINTER

samp

les

7 (10,4)

(20,21)

8 7 (10,12)

(20,21)

8

0.900 * * * * * * * *

0.905 * * * * * * * *

0.910 * * * * * * * *

0.915 * * * * * * * *

0.920 * * * * * * * *

0.925 & * * * * * * *

0.930 & * * * & * * *

0.935 & * * * & * * *

0.940 & * * * & * * *

0.945 & * * * & * * *

0.950 & * * * & * * *

0.955 & * * * & & * *

0.960 & * * & & & * *

0.965 & & & & & & & *

0.970 & & & & & & & &

NON OVERLAPPING CLUSTERS FOR SUITABLE ALPHA CUTS IN SUMMER AND WINTER(MIN_MAX)

* represents absence of sample

& represents presence of the sample

4.Membership values assigned for the water quality parameters

5.ADVANTANTAGES

In this methodology we proposed a sys of forming clusters with the samples which

Min-Max(dhar_ror)

Alpha cuts

SUMMER WINTER

samples

7 (10,4)

(20,21)

8 7 (12,10)

(20,21)

8

0.900 * * * * * * * *

0.905 & * * & * * * *

0.910 & *` & & & * * *

0.915 & * & & & * * *

0.920 & * & & & * * *

0.925 & * & & & * * *

0.930 & * & & & * * *

0.935 & * & & & * * *

0.940 & * & & & * * *

0.945 & * & & & * * &

0.950 & * & & & * * &

0.955 & * & & & * * &

0.960 & * & & & * * &

0.965 & & & & & & * &

0.970 & & & & & & & &

487

proves to be more efficient than the chemistry aspect of comparing each and

every parameters . We provide an automated software which

reduces the time consumption and provides

a more clear understanding about the properties of water

From each appreciable cluster, containing a group of water samples, the similarity of ground water quality among the samples

with respect to the chosen parameters, was identified.

6. FUTURE WORK:

Various data mining algorithms(clustering) are analyzing to improve the efficiency and to find out the nature of these samples

A wide range of samples can be analyzed so

that quality of the water can be increased and unique samples may be preserved .

7. References

1. R.-S,Lu,S.-L Lo, J.-Y.Hu,1999. Analysis of reservoir water quality using fuzzy

synthetic evaluation. Stochastic

Environmental Research and Risk Assessment

13

327-336.

2. WANG Jianhua, LU Xianguo, TIAN Jinghan ,

JIANG Ming, 2008. Fuzzy Synthetic

Evaluation of Water Quality of Naoli River

Using Parameter Correlation Analysis.

Chin. Geogra. Sci. 18(4) 361-368. DOI: 10.1007/s1179-008-0361-5.

3. Bhupinder Singh . Sudhir Dahiya . Sandeep

Jain. Use of fuzzy synthetic evaluation for

Assessment of groundwater quality for

drinking usage: a case study of Southern

Haryana, India. Environ Geol 54: 249 –

255.DOI: 10.1007/s00254-007-0812-9

4. Cengiz Kahraman . Ihsan Kaya, 2009. Fuzzy

process capability indices for quality control of irrigation water. Stoch Environ

Res Risk Assess 23: 451-462. DOI:

10.1007/s00477-008-0232-8.

5. MAO Yuan-yuan, AHANg Xue-gang,

WANG Lian-Sheng, 2006. Fuzzy pattern

recognition method for assessing ground

vulnerability to pollution in the Zhangji area.

Journal of Zhejiang University Science A ,

ISSN 1009-3095(Print); ISSN 1862-

1775(Online).

6. REN Chuntao, LI Changyou, JIA Keli,

ZHANG Sheng, LI Weiping, CAO Youling .

Water quality assessment for Ulansuhai Lake

using fuzzy clustering and pattern Recognition. Chinese Journal of Oceanology

and Limnology, Vol. 26 No.3, P.339-

344,2008.

7. AYSE MUHAMMETOGLU and AHMET

YARDIMCI,2006. A fuzzy logic approach

to asses groundwater pollution levels below

agricultural fields. Environmental

Monitoring and Assessment 118: 337-354.

DOI: 10.1007/s10661-006-1497-3.

Proceedings of the third national conference on RTICT 2010

Bannari Amman Institute of Technology,Sathyamangalam-638401

9-10 April 2010

IMAGE SEGMENTATION USING CLUSTERING ALGORITHM IN

SIX COLOR SPACES

S.Venkatesh Kumar, [email protected],9025895717

S.Thiruvenkadam M.E, Sr.Lecturer

Dr.Mahalingam College of Engineering and Technology, pollachi. .

Abstract

Image segmentation is an subdivides of an

image into constituent parts or objects of natural

image. It is a very efficient segmentation

approach method mainly it is based on fusion

and blending process which aims at combining

relevant information from two or more images

into a single image. The resulting image will be

more informative than any of the input images in order to finally get a good, reliable, accuracy

and high performance quality of the image. In

our applications, we use simple k-means

clustering algorithm technique on an input

image is expressed in different color spaces. This

fusion process is to combine overall

segmentation images and also perform fastly to

improve the quality of the image. The fused

image can have complementary spatial and

spectral resolution characteristics. This fusion

framework process remains simple to implement, fast, easily parallelizable general enough to be

applied to various computer vision applications

such as aerial and satellite imaging, motion

detection, robot vision. Finally a significant

improvement of subjective quality of image will

be achieved reliable and accuracy.

Key terms: Image Segmentation, Image

Fusion, Color Spaces, K-means Algorithm.

I INTRODUCTION

In computer vision, segmentation refers to the

process of partitioning a digital image into

multiple segments (sets of pixels) (Also known

as super pixels). The goal of segmentation is to

simplify and/or change the representation of an

image into something that is more meaningful

and easier to analyze. Segmentation of nontrivial

images is one of the most difficult task in image

processing. Segmentation accuracy determines

the eventual success or failure of computerized

analysis procedures. The goal of image

segmentation is to cluster pixels into salient image regions, i.e., regions corresponding to the

individual surfaces, objects, or natural parts of

objects.

A segmentation could be used for object

recognition, occlusion boundary estimation

within motion or stereo systems, image

compression, image editing, or image database

look-up. Because of its simplicity and efficiency,

clustering approaches were one of the first techniques used for the segmentation of

(textured) natural images. The set of connected

pixels belonging to each estimated class thus

defined the different regions of the image. The

most common algorithm uses an iterative

refinement technique. Due to its ubiquity it is

often called as the k-means algorithm; it is also

referred to as Lloyd’s algorithm particularly in

the computer science community. Many other

methods have been proposed previously to solve

the texture of image segmentation problem such as efficient graph based image segmentation,

contrary to clustering algorithm, mean-shift-

based techniques, Markov random field (MRF)-

based statistical models [4].

In recent years research work has been

focused on color image segmentation, since grey

scale images can not satisfy the needs in many

situations. Color is perceived by humans as a

combination of tristimuli Red(R), Green(G), and

Blue(B) which form RGB color space. Other kinds of color representation can be driven from

RGB color space by using either linear or non

linear transformation. Hence, several color

spaces such as HSV, YIQ, XYZ, LAB, LUV are

used in color image segmentation.

Often an object that is not extracted in gray

levels can be extracted while using the color

information. Generally monochromatic

segmentation techniques are extended to color

images however all these techniques have

advantages and inconvenient. Most method of segmentation are combination of classic

488

techniques and/or fuzzy logic notations, neural

network, genetic algorithms etc complete.

Clustering is the search for distinct groups in

the feature space. It is expected that these groups

have different structures and that can be clearly differentiated. The clustering task separates the

data into number of partitions, which are

volumes in the n-dimensional feature space.

These partitions define a hard limit between the

different groups and depend on the functions

used to model the data distribution.

It’s a new strategy for segmentation approach

method and better designed segmentation model

of natural image. In our technique it explores

the possible alternative of fusing several

segmentation images combining to a simple segmentation models in order to get a reliable,

accuracy and high performance quality of the

image. Mainly this work proposes a fusion

framework which aims at fusing several k-means

clustering result applied on input image

expressed by different color spaces. Among them

k-means clustering is often applied as an

essential step in the segmentation process.

As a traditional clustering algorithm, K-Means

is popular or its simplicity for implementation, and it is commonly applied for grouping pixels

in images or video sequences. Therefore, a fast

and efficient algorithm for K-Means image

segmentation is proposed to handle these

problems. K-Means algorithm is an unsupervised

clustering algorithm that classifies the input data

points into multiple classes based on their

inherent distance from each other. The algorithm

assumes that the data features form a vector

space and tries to find natural clustering in them. These different label fields are fused together by

a simple K-means clustering techniques using as input features, the local histogram of the class

labels, previously estimated and associated to

each initial clustering result.

It demonstrates that the proposed fusion

method, while being simple and fast performs

competitively and often better in terms of visual

evaluations and quantitative performance

measures than the best existing state-of-the-art

recent segmentation methods. to be classified.

Most of the works in color image segmentation can be roughly classified into several categories:

Histogram thresholding, clustering methods,

region growing, edge based approaches and

fuzzy methods. A combination of these

approaches is often used for color image

segmentation. Finally used fusion procedure to

combine several segmentation maps in order to

get a high performance quality of the image.

II PRIMARY SEGMENTATIONS

The initial segmentation maps which will then

be fused together by our fusion framework are

simply given, in our application, by a K -means

[2] clustering technique, applied on an input

image is expressed by different color spaces, and

using as simple cues (i.e., as input

multidimensional feature descriptor) the set of

values of the re-quantized color histogram (with

equidistant binning) estimated around the pixel

to be classified. Several algorithms have been

proposed in the literature for clustering: P-CLUSTER, CLARANS, CLARA, focusing

techniques, Ejcluster and GRIDCLUS. The k-

means method has been shown to be effective in

producing good clustering results for many

practical applications.

In k-means clustering we are given a set of n

data points in d-dimensional space and an integer

k, and the problem is tom determine a set of k-

points in d-space, called centers, so as to minimize the mean squared distance from each

data points to its nearest center. A popular

heuristic for k-means clustering is Lloyd’s

algorithm.

Mathematically, let b(x) 0,1,….,Nb-1 denote

the bin index associated with the color vector

y(x) at pixel location (lying on a pixel grid) and

be the set of pixel locations x and N(x) within

the squared neighborhood region (of fixed-size

Nw X Nw ) centered at pixel location x (in which

local color information will be gathered). H(x) = h (n;x)n=0,1,….,Nb-1 is estimate of 125 bins

descriptor. Its characterizes the color distribution

for each pixel to be classified, is given by the

following standard bin counting procedure:

489

( )

( ; ) [ ( ) ]u N x

h n x b u n

(1)

Where is the Kronecker delta function

and k = 1/ (Nw) is a normalization constant

ensuring function

Fig 2 Estimation, for each pixel x, of the

Nb=q3 bins descriptor (q=5) in the RGB color space. The RGB color cube is first divided into

Nb=q3 equal sized smaller boxes (or bins). Each

Rx, Gx, Bx color value associated to each pixel

contained in a (squared) neighborhood region

(of size Nw X Nw) centered at X, increments(+1)

a particular bin. The set of bin values represents

the (non normalized) bin descriptor. We then

divide all values of this Nb bins descriptor by

(Nw X Nw) in order to ensure that the sum of

these values integrates to one.

Algorithm I. Estimation, for each pixel X,

of the bins descriptor.

Estimation of the Nb = q3 bins descriptor

Nx set of pixel locations X within the

Nw X Nw neighborhood region

centered at X.

h[ ] Bins descriptor: Array of

Nb floats(h[0].h[1],…h[Nb-1] )

integer part of. for each pixel X belonging to Nx with color

Value Rx,Gx,Bx.

do

k←q2.q.Rx/256+q.q.Gx/256

+ q.Bx/256

h[k] ← h[k]+1/(Nw )

Here a texton is a repetitive character or

element of a textured image (also called a texture

primitive), is characterized by a mixture of

colors or more precisely by the values of the re-

quantized local color histogram. This model

while being robust to noise and local image transformations is also simple to compute and

allows significant data reduction and has already

demonstrated all its efficiency for tracking

applications [4].

Fig.3 Primary Segmentation (PS) Phase

Finally, these (125-bin) descriptors are

grouped together into different clusters

(corresponding to each class of the image) by the

classical K-means algorithm [2] with the

classical Euclidean distance. This simple

segmentation strategy of the input image into

classes is repeated for different color spaces

which can be viewed as different image channels

provided by various sensors or captors (or as a

multi channel filtering where the channels are

represented by the different color spaces).

III.COLOR SPACE FEATURES

Segmentations (Ns) provided by the 6 color

spaces, C= RGB, HSV, YIQ, XYZ, LAB,

LUV [1], [5]–[7] are used. These initial

segmentations to be fused can result the same

initial and simple model used on an input image

filtered by another filter bank (e.g., a bank of

Gabor filters [3], [8] or any other 2-D

decomposition of the frequential space) or can

also be provided by different segmentation models or different segmentation results

provided by different seeds of the same

stochastic segmentation model.

The final fusion procedure is more reliable as

the interesting property of each color space has

been taken into account. For example, RGB is

the optimal one for tracking applications

Image in

CSi

Local color

histogram

Initial

segmentation

map

texton

490

[18].due to it being an additive color system

based on trichromatic theory and nonlinear with

visual perception. The HSV is more apt to

decouple chromatic information from shading

effect [4]. The YIQ color channels have the

property to code the luminance and chrominance information which are useful in compression

applications (both digital and analogue). Also

this system is intended to take advantage of

human color characteristics.

XYZ although are nonlinear in terms of linear

component color mixing, has the advantage of

being more psychovisually linear. The LAB

color system approximates human vision, and its

component closely matches human perception of

lightness [1]. The LUV components provide a

Euclidean color space yielding a perceptually uniform spacing of color approximating a

Riemannian space [8]. To measure four image

segmentation indices for each color spaces

finally using fusion process is to combine several

segmented images into a single image inorder to

get a high performance quality of the image.

The complete specification of an RGB color

space also requires a white point chromaticity

and a gamma correction curve. HSV - defines a

type of color space. It is similar to the modern

RGB and CMYK models. The HSV color space has three components: hue, saturation and value. The Y component represents the luma

information, and is the only component used by

black-and-white television receivers. I and Q

represent the chrominance information. A Lab

color space is a color opponent space with

dimension L for lightness and a and b for the

color-opponent dimensions, based on

nonlinearly-compressed CIEXYZ color space

coordinates. LUV color space is extensively used

for applications such as computer graphics which

deal with colored lights.

RGB YIQ XYZ

HSV LUV LAB

Fig 4 Characteristics of Six Color Spaces

IV. IMAGE SEGMENTATION TO BE

FUSED

The key idea of the proposed fusion process

using k-means algorithm as simply consists of

considering, for each site (or pixel to be

classified), the local histogram of the class (or

texton) labels of each segmentation to be fused,

computed on a squared fixed size Nw

neighborhood centered around the pixel, as input

feature vector of a final clustering procedure. For

a fusion of Ns segmentation with K1 classes into

a segmentation with K2 classes, the preliminary feature extraction step of this fusion procedure

thus yields to Ns (K1-bin) histograms which are

then gathered together in order to form, K1 X Ns

dimensional feature vector or a final bin

histogram which is then normalized to sum to

one, so that it is also a probability distribution

function.

In computer vision, Multisensor Image fusion

is the process of combining relevant information

from two or more images into a single image. The resulting image will be more informative

than any of the input images. Image fusion

methods can be broadly classified into two -

spatial domain fusion and transform domain

fusion. The fusion methods such as averaging,

Brovey method, principal component analysis

(PCA) and IHS based methods fall under

spatial domain approaches.

The proposed fusion process and k-means

algorithm is then herein simply considered as a

problem of clustering local histograms of (preliminary estimated) class labels computed

around and associated to each site. To this end,

we use, once again, a K-means clustering

procedure exploiting, for this fusion step, a

histogram based similarity measure derived from

the Bhattacharya similarity coefficient. (Fig 5)

Bhattacharya distance between a normalized

histogram and a reference histogram is given by

1

* * 1/2[ , ( )] 1 ( ) ( ; )( )bN

B

n o

D h h x h n h n x

Fig.5 shows an example of the clustering

segmentation model (into classes) of an input

image expressed in the RGB, HSV, YTQ, XYZ,

LAB, and LUV color spaces and the final

segmentation map (into classes) which results of

the fusion of these clustering. We can find that

491

none of them can be considered as reliable

except the final segmentation result (at bottom

right) which visually identifies quite faithfully

the different objects of the scene.

Fig. 5 from top to bottom and left to right:

(top left) input natural image. Six segmentation

results (into K = 6 classes) associated to

clustering model described in on the top left

input image expressed in the RGB, HSV, YIQ,

XYZ, LAB, and LUV color spaces and final

segmentation map (into K = 6 classes) resulting

of the fusion of these six clustering

VII. AVOID OVER SEGMENTATION

A final merging step (refer Fig.6) is necessary

and is used to avoid over segmentation for some

images. It consists of blending each region (i.e.,

set of connected pixels belonging to the same

class) of the resulting segmentation map with

one of its neighboring region R if the distance

DMERGING is below a given threshold (or if its size

is below 50 pixels with the closest region in the

DMERGING distance sense.

min [ ( ), ( ; )] o

MERGING Bx R

c

D D h n h n x

Fig.6 final merging using the Bhattacharya

distance as merging criterion on a fused segmented image.

VIII PROPOSED WORK

The proposed segmentation approach is

conceptually different and explores a new

strategy. Instead of considering an elaborate and

better designed segmentation model of textured natural image, this technique explores the

possible alternatively combining two process

such as fusing and blending procedure (i.e.,

efficiently combining) several segmentation

maps associated to simpler segmentation models

in order to get a final reliable and accurate

segmentation result. More precisely, this work

proposes a fusion framework which aims at

fusing several K-means clustering algorithms

(herein using as simple cues the values of the

requantized color histogram estimated around the

pixel to be classified) applied on an input image expressed by different color spaces. These

different label fields are fused together by a

simple K-means clustering techniques using as

input features, the local histogram of the class

labels, previously estimated and associated to

each initial clustering result. Finally it measures

four image segmentation indices such as PRI,

VOI, GCE AND BDE for each color spaces and

fusion process. It demonstrates that the proposed

fusion and blending method, while being simple,

easily parallelizable and fast performs competitively and often better (in terms of visual

evaluations and quantitative performance

measures) than the best existing state-of-the-art

recent segmentation methods.

*

*

* k-means algorithm

Fig .7 Block diagram- Segmentation- Different

Color Spaces

Input

image

(i/p)

Six color

spaces PS

phase

Final

segmentation

result

Merging

phase

Fusion

and

blending

phase

492

493

IX CONCLUSION

This paper have presented a new, simple, and

efficient segmentation approach, based on a

fusion procedure which aims at combining

several segmentation maps associated to simpler partition models in order to finally get a more

reliable and accurate segmentation result and

also we get good and high performance quality

of the image . The primary segmentations to be

fused can also be provided by different

segmentation models or different segmentation

results provided by different seeds of the same

stochastic segmentation model.

REFERENCES

[1] Bo Zhau, Zhongxiang Zhu, Enrong Mao and

Zhenghe Song, “Image segmentation based on ant colony optimization and K-means

clustering,” Proc. IEEE International Conference

on Automation and Logistics, August 18-21,

2007.

[2]H.Stokman and T.Gevers, “Selection and

fusion of color models for image feature

detection,” IEEE Tans. Pattern Anal. Mach.

Intell., vol.29, no.3, pp.371-381, mar.2007.

[3] Charles kervrann and Fabrice Heitz, “A MRF

model –based approach to unsupervised texture

segmentation using global spatial statistics,” IEEE Trans. on image processing, vol.4, no.6,

june1995.

[4] Ming Zhang and Reda Alhajj, “Improving the

graph based image segmentation method,” Proc.

IEEE International Conference on tools with

Artificial Intelligence pp.0-7695-2728, Nov

2006.

[5] Mustafa Ozden and Ediz Polat, “image

segmentation using color and texture features,”

Proc. IEEE International Conference on tools

with Artificial Intelligence pp.0-2145-2728, Nov

2005. [6] K.Blekas, A.Likas, N.P.Galatsonas, and

I.E.Lagaris, “A spatially constrained mixture

model for image segmentation,” IEEE Trans. on

neural network, vol.16, no.2, march2005.

[7] Priyam chatterjee and Peyman milanfar,

“Clustering based denoising with locally learned

dictionaries,” IEEE Trans. on Image Processing,

vol.18, no.7, July 2006.

[8] J.P. Braquelaire and L. Brun, “Comparison

and optimization of the methods of color

image quantization,” IEEE Trans. image process., vol. 6, no. 7, pp. 1048–1952, Jul. 1997.

[9] H.Essaqote, N.Zahid, I.Haddaoui and

A.Ettouhami, “Color image segmentation based

on new clustering algorithm,” IEEE journal on

applied sciences 2(8):853-858, May 2007.

[10] Tse-Wei Chen, Yi-Ling Chen and Shao-Yi

chien, “Fast image segmentation based on K-

means clustering on HSV color spaces,” IEEE

Int. Conf. on Image Proc. April 2008. [11] D. Comaniciu and P. Meer, “Mean shift: A

robust approach toward feature space analysis,”

IEEE Trans. Pattern Anal. Mach. Intell., vol.24,

no. 5, pp. 603–619, May 2002.

[12] E. Maggio and A. Cavallaro, “Multi-part

target representation for color tracking,” in Proc.

Int. Conf. Image Processing, Italy, Genova, Sep

2005, pp. 729–732.

Proceedings of the third national conference on RTICT 2010

Bannari Amman Institute of Technology,Sathyamangalam-638401

9-10 April 2010

A Novel Approach for Web Service Composition in Dynamic Environment Using

AR -based Techniques

R.Rebecca, Department of CSE, Anna University Tiruchirappalli,[email protected] I.Shahanaz Begum, Lecturer, Anna UniversityTiruchirappalli,[email protected]

Abstract— Web service compositions are receiving significant amount of interest as a important strategy to allow

enterprise collaboration. Dynamic web service selection refers to determining a subset of component web services to be

invoked so as to orchestrate a composite web service. We observe that both the composite and constituent Web services

often constrain the sequences of invoking their operations. We therefore propose to use finite state machine (FSM) to

model the invocation order of web service operations. We define a measure, called aggregated reliability, to measure the

probability that a given state in the composite web service will lead to successful execution in the context where each

component web service may fail with some probability. In orchestrating a composite web service, we propose a strategy

to select component web services that are likely to successfully complete the execution of a given sequence of operations.

A prototype that implements the proposed approach using BPEL for specifying the invocation order of a Web service is

developed

Index Terms: Web services, reliability, service composition, web service selection, BPEL.

1 INTRODUCTION:

The web service today has radically changed how

businesses interact with applications distributed within and

across organizational boundaries. However a WS may run in

highly dynamic environment: business rules may change to fit new requirement and existing services may become

temporarily unavailable, etc. Such a highly dynamic

environment increases the probability to lead to unsuccessful

execution. So simply choosing a WS for execution an

incoming operation may not be the best strategy because it

may lead to the violation of some constraints. Therefore it is

important to effectively integrate several available WSs into

composite WS to lead successful execution. The composition of WSs involves using an

orchestration model to 1) define the possible orders of calling

WSs at design time and 2) dynamically select WSs to be invoked at runtime. To address the former, several theoretical

orchestration models have been proposed in the literature,

including finite state machine (FSM), Petri net, -calculus,

activity hierarchies, and rule-based orchestration [1]. Practical

service composition languages, such as BPEL, WS-

Choreography, WSCL, XPDL, and OWL-S have also been

proposed, and many commercial service platforms or products

that implement the available standards and composition

languages are available in the market, including the Sun ONE

framework based on J2EE, Microsoft.NET, the Oracle BPEL

Process Manager, the HP WSs Management Platform, and the

IBM Web Sphere Application Server.

However, none of the above works provide strategies

or mechanisms to dynamically select WSs to be invoked when

executing a series of operations in a composite WS. Dynamic WS selection refers to choosing the available WSs

to be invoked so as to realize the functionality of a composite

WS constructed using an orchestration model at runtime.

However, a WS may comprise multiple operations such that

their invocation sequence is constrained. Typical constraints

are separation of duties and binding of duties. It is imperative

that the dynamic WS selection Procedure takes all these

constraints into account. When choosing the operations of

WSs to compose a (composite) WS, the atomicity of each WS,

which requires either none of its operations to be invoked or

some final state to be reached, has to hold at the end of WS composition.

The network (intranet, extranet, or even the Internet)

where WSs operate is a failure-prone environment partly due

to the autonomous requirement of each participating service

provider. A WS may become malfunctioned or unavailable at

runtime, causing failure to the execution of a composite WS.

The WS reliability can be measured from various

perspectives, including correctness, fault tolerance, testability,

interoperability, and timeliness [3]. We define the reliability

of a WS operation as the probability that it successful

responds within a reasonable period of time and the proposed approach is not restricted to any WS modeling language and

employs a generic formalism FSM to define the order of

operations in a WS.

2 RELATED WORKS:

Self-Adaptation by late binding service calls: Bianculli et al. [8] propose architecture for self-adaptive Web service

compositions in which each service has an associated

repudiation. During service calls, the engine monitors QoS

and pushes repudiation data to a registry. For each

service call, a specialized service calls mechanism first checks

for a low repudiation of the called partner. In case the

494

repudiation is low another partner is selected from a registry.

In comparison, this work can monitor/adapt only service calls

and not other BPEL activities. Further, new self-adaptation

features cannot be added.

Karastoyanova et al. Propose to extend BPEL by the “find and bind” mechanism. By adding a new construct <find

bind> to BPEL, they enable the dynamic binding of explicitly

component web services at run time. Another similar work [7]

presents an approach for improving the adaptability by

extending the composition engine. A proper repository before

a web service is invoked. Some other utilizes Aspect-Oriented

Programming (AOP) to improve the flexibility of web service

composition. Charfi [6] proposes an extension to the BPEL

language, which is the service works, oriented BPEL

(AO4BPEL).

Several different QoS models have been proposed in

literature (e.g., [8, 12] ). However, most approaches do not

discuss how QoS can be monitored. An overview of QoS

monitoring approaches for web services is presented by Thio

et al. [12]. The authors discuss various techniques such as

low-level suing or proxy-based solutions. QoS monitoring has

been an active research area for years which is not only

focused on web service technology.

3 PROPOSED WORK:

Each (composite or candidate) WS is associated with a

formal model FSM that constrains the order of its operations.

The objective of automatic WS composition is to determine a

composition that satisfies all the given constraints. The reliability of a composite WS can be derived by aggregating the

reliabilities of constituent WSs based on the occurrence rate of each flow pattern.

Definition 1: A WS W is composed of a set of operations Σ and an FSM that prescribes the legal executions of these

operations. W is a tuple (Σ,S ,s0, δ, F) where

Σ is a set of operations,

S is a finite set of states,

s0 is a state in S, representing the initial state of W,

δ: S×Σ—>S is the transition function of an FSM,

F S represents the set of final states, i.e., the states where the interactions with W can be terminated.

The FSM of a WS prescribes legal executions of its

operations so as to realize the control flow, data flow, and/or

object flow supported by the WS.

Definition2: A target WS, denoted Wt, and is a WS whose

operations are interfaces to operations in other WSs also known as the component WSs. In general, when a client

application invokes operations in a target WS, the target WS

does not execute these operations itself, but delegates them to

the component WSs instead. The set of component WSs to be

delegated is called the WS community.It is possible that all

operations on the target WS can be handled by one single

Fig1: Invoking relationships between components WSs, the

composite WS, and the application

component WS, the more general scenario is to involve

multiple component WSs to complete the entire series of

operations on the target WS.

In this work, we consider a target WS Wt and a WS

community W1, W2… ,Wn. As an invocation of an operation in Wt can be in fact executed by more than one

component WSs.

As an example, suppose that one is planning a trip to

attend an event. The trip planning can be regarded as a target

WS (Wt), as shown in Figure 2. Also assume that there are

five WSs available on the Internet, each capable of

accomplishing only a subset of operations required by the

entire trip planning.W1 provides a package for booking both a

flight and a car.W2 offers the users to book a flight and a car

or a limo. W3 provides hotel booking and shuttle booking

services. W4 is used to register an event. W5 allows the user to book either a flight or a cruise, followed by the booking of

a car or a limo.

Fig 2: Target WS for Trip Planning

495

In Figure 3 the five WSs form the WS community.

Notice that the atomicity property of each component WS

should be preserved. i.e., it should be either in its initial state

or in some final state when the target WS is terminated and

reaches its final state.

A sequence of operations accepted by a WS

composition describes a successful execution of this operation

sequence on the target WS using some delegated operations in

the WS community. It is possible that the successful

execution of operation sequence on the target WS has more

than one delegations. For example, considering the

composition shown in Figure 3, the execution of WSt (login,

flight_booking, hotel_booking, car_booking,

shuttle_booking, event_registering) has delegations (W1, W1,

W3, W1, W3, W4) , (W2, W2, W3, W2, W3, W4), and (W5,

W5, W3, W5, W3, W4).

3.1 Dynamic WS selection Problem:

There are multiple possible delegations of the same

operation sequence on the target WS, the choice of which

delegation to use is then the WS Selection Problem.

When it comes to choosing a WS from a given WS

community to delegate for an incoming operation, we would

like to choose the one that will most likely lead to successful

execution of the entire operation sequence at the same time

maintaining a high degree of composability for the subsequent operations on the target WS.

To choose the delegation for the operation o1, there are

two choices: W1 and W2, which lead to configurations (1, 1,

0s, 0s, 0s, 0s) and (1, 0s, 1, 0s, 0s, 0s), respectively. However,

only the configuration (1, 1, 0s, 0s, 0s, 0s) is composable.

Configuration (1, 0s, 1, 0s, 0s, 0s) is not composable.

Fig 3: WS community for trip planning

The proposed solution to the dynamic WS selection

problem consists of aggregated reliability, a novel metric to

measure the reliability of each configuration in a WS

composition, and strategies to use the computed aggregated

reliabilities for dynamically selecting component WSs for the

operations to be performed on the target WS.

Definition3: (Aggregated Reliability)

The aggregated reliability of a configuration c, denoted

Agg_Reliability(c), is the probability that an execution starts with c will terminate at some final configuration.

The AR of a configuration ci, Agg _Reliability (ci), can be

defined recursively as follows:

Agg_Reliability(ci)=

1, if ci=Success

0, if ci=Failure Σ Pij.Agg_reliability(cj), otherwise

Consider the target WS W0 and the WS community

C = W1;W2;W3;W4 shown in Fig.4 shows the

composition of W0 using C, where, for simplicity, we use a

single final configuration G to denote all the final

configurations. Assume that each operation has an equal

chance to be selected and has the same reliability α= 0.8,

except for operations W3 .o2, W3 .o3, and W4 .o3, whose

reliabilities are all 0.75.

3.2 AR-based dynamic WS selection strategies :

In the AR-based selection strategy, we select the component WS to delegate an operation on the target WS that

can give the maximal reliability.

Fig 4: The composition of the target WS using the WS

community.

. At runtime, when an incoming operation o arrives at a

configuration ci, we first sort the candidate WS operations in

descending order of the aggregated reliabilities of their

destined configurations. These WS operations are tried one at

time in the order until one gets successfully executed. This strategy, called aggregated reliability-based (AR-based)

selection, is shown in Figure 5:

Dynamic web service selection procedure also consider

that, when an atomic WS fails at runtime and all its

operations become unavailable, the aggregated reliabilities of

496

some configurations derived in the design time may no longer

be valid. An efficient recomputation of aggregated reliabilities

at runtime is thus required to allow reliable WS selection in a

dynamic environment.

A prototype that implements the proposed approach that,

the invocation order of operation of either composite or atomic WS, if specified using BPEL, has to be converted into

abstract FSM.FSM is a promising model for analyzing service

connectivity, composition correctness, automatic composition

and scalability. The main building blocks of BPEL are

activities.

The selection of the WSs and the invocation of

operations in the selected WSs are performed by an

intermediary agent called the delegator, which is also

implemented as a WS. The WS composition and the

Fig 5: AR- based selection algorithm

namespace of each atomic WS will be the input to the

delegator. When the composite WS needs to find and

invoke a delegated operation, it delegates this job to the

delegator WS by invoking its “delegator” operation. The

“delegator” operation will decide on a suitable (in our work,

the more reliable and/or composable) atomic WS using some

selection strategy and subsequently invokes its corresponding

operation. For a synchronous operation, the “delegator”

operation will wait until receiving the reply from the selected

WS. For an asynchronous operation, a callback <receive>

activity is expected by the delegator WS.

4 CONCLUSIONS AND FUTURE WORK:

We have formulated the dynamic WS selection

problem in a dynamic and failure-prone environment. We

proposed to use FSM to model the invocation order of

operations in each WS and to construct a WS composition

that enumerates all possible delegations. We defined a measure, called aggregated reliability, to determine the

probability that the execution of operations starting from a

configuration will successfully terminate.

The proposed approach can be perfectly applied to real

industrial applications with quick WS selection at runtime in a

dynamic environment. However, when the number of atomic

WSs becomes large, constructing a composition may take too

much time, thereby making this approach impractical.

The current solution approach does not distinguish

the types of operation invocations, e.g., control flow, data

flow, and object flow and assumes each operation has equal

probability of being chosen. By considering other

implementation issues of dynamic WS selection and relaxing

the existing assumptions, we aim to introduce more practical

selection strategies in our future work.

REFERENCES:

1) G. Alonso, H. Kuno, F. Casati, and V. Machiraju,

Web Services: Concepts, Architectures and

Applications. Springer, 2004.

2) R. Bhatti, E. Bertino, and A. Ghafoor, “A Trust-

Based Context-Aware Access Control Model for

Web-Services,” Distributedz and Parallel Databases,

vol. 18, no. 1, pp. 83- 05, 2005.

3) J. Zhang and L.J. Zhang, “Criteria Analysis

and Validation of the Reliability of Web Services-

Oriented Systems,” Proc. IEEE Int’l Conf. Web

Services (ICWS ’05), pp. 621-628, 2005.

4) V. Grassi and S. Patella, “Reliability Prediction for

Service-Oriented Computing Environments,”IEEE

Internet computing, vol. 10, no. 3, pp. 43-49, 2006.

5) L. Zeng, B. Benatallah, A.H.H. Ngu, M. Dumas, J.

Kalagnanam, and H. Chang, “QoS-Aware

Middleware for Web Services Composition,” IEEE

Trans. Software Eng., vol. 30, no. 5, pp. 311-327,

2004.

6) A. Charfi and M. Mezini. Aspect-Oriented Web

Service Composition with AO4BPEL. In Proc. of European Conference on Web Services (ECOWS).

Springer, 2004.

7) J. Cao, S. Zhang, M. Li, and J. Wang. Verification of

dynamic process model change to support the

497

adaptive workflow. In Proc. of 2004 IEEE

International Conference on- Services Computing

(SCC), 2004.

8) Comprehensive QoS Monitoring of Web Services

and Event-Based SLA Violation Detection Anton Michlmayr, Florian Rosenberg, Philipp Leitner,

Schahram DustdarDistributed Systems Group,

Vienna University of Technology Argentinierstrasse

8/184-1, 1040 Wien, Austria

[email protected]

9) P. Chan, M. Lyu, and M. Malek. Reliable web

services: Methodology, experiment and modeling. In

Proc. of IEEE International Conference on Web

Services, Salt Lake City, Utah, USA, 9-13 Jul 2007.

10) P. P. W. Chan. Building Reliable Web Services: Methodology, Composition, Modeling and

Experiment. PhD thesis, The Chinese University of

Hong Kong, Hong Kong, Dec 2007.

11) Arias-Fisteus, J., Fernandez, L.S., Kloos, C.D.:

Formal Veri¯cation of BPEL4WS Business

Collaborations. In Bauknecht, K., Bichler, M.,

PrÄoll, B., eds.: EC-Web. Volume 3182 of LNCS.

Springer (2004).

12) Comprehensive QoS Monitoring of Web Services and Event-Based SLA Violation Detection Anton

Michlmayr, Florian Rosenberg, Philipp Leitner,

Schahram DustdarDistributed Systems Group,

Vienna University of Technology Argentinierstrasse

8/184-1, 1040 Wien, Austria

[email protected]

498

Our Special Thanks to

To ADVERTISERS, DONORS AND SPONSORS

Compiled and Edited by

Convener

Dr. Amitabh Wahi, Professor & Head, Dept. of Information Technology, BIT

Organizing Secretaries

Mr. S. Sundaramurthy and Mr. S. Daniel Madan Raja,

Dept. of Information Technology, BIT