17_efficient_costume_analysis_and_recognition_acceptance_for_optically_challenged_humanity.pdf

10
ISSN(Online) : 2320-9801 ISSN (Print) : 2320-9798 International Journal of Innovative Research in Computer and Communication Engineering An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015 Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015] On 3 rd October 2015, Organized by Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India Copyright @ IJIRCCE www.ijircce.com 80 Efficient Costume Analysis and Recognition Acceptance for Optically Challenged Humanity N.J Divya #1 , X.Anitha Sarafin #2 Dept. of CSE, Misrimal Navajee Munoth Jain Engineering College, Chennai, Tamil Nadu, India 1,2 ABSTRACT : Automatic clothing pattern recognition and matching costumes with appropriate color is a challenging task for optically defective humanity. We developed a camera based prototype system for recognizing clothing patterns such as plaid, striped, pattern-less and irregular. It identifies 11 clothing colors. A camera mounted upon a pair of sunglasses is used to capture the clothing images. We propose a novel Radon Signature descriptor to capture the global directionality of clothing pattern. SIFT represents the local structural features and STA descriptor is used to extract the global statistical features from wavelet sub band and they are combined with local features to recognize the complex clothing patterns. To evaluate the efficiency, we use the CCNY clothing pattern dataset which captures 627 images. To assist blind persons to read the text written on the clothes, we have a conceived camera based assistive text reading framework to track the object of interest within the rectangular arrangement and extract printed text information from the clothes. Our Optical Character Recognition System(OCR) can effectively handle complex background and multiple patterns, and extract text information from that costumes(object). This prototype system supports more independence in their daily basis of the visually challenged people. KEYWORDS: Radon Signature, SIFT(Scale Invariant Feature Transform),STA(Statistical Descriptor), CCNY Dataset(City College Of New York),Assistive Text Reading Framework, OCR(Optical Character Recognition System) ,Local and Global image features. I. INTRODUCTION Due to large intra-class variations in patterns and designs in clothes cause difficulty to recognize the clothing patterns so we employ the automatic camera based clothing pattern recognition system. Existing texture analysis methods mainly focus on textures having large changes in viewpoint, orientation, and scaling image, but it presents less intra-class pattern and intensity variations. The system contains three major components 1)sensors including a camera to capture the clothing images, micro- phone for giving speech command input and speakers (or Bluetooth, earphone) for audio output display 2) data capture and analysis performs command control, clothing pattern recognition, and color identification uses a computer which can be a desktop or a wearable computer (e.g., a mini -computer or a smart phone) and 3) audio outputs provides recognition results of clothing pattern, colors and also the system status.

Upload: sachin

Post on 17-Feb-2016

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: 17_EFFICIENT_COSTUME_ANALYSIS_AND_RECOGNITION_ACCEPTANCE_FOR_OPTICALLY_CHALLENGED_HUMANITY.pdf

ISSN(Online) : 2320-9801

ISSN (Print) : 2320-9798

International Journal of Innovative Research in Computer and Communication Engineering

An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015

Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015]

On 3rd

October 2015, Organized by

Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India

Copyright @ IJIRCCE www.ijircce.com 80

Efficient Costume Analysis and Recognition

Acceptance for Optically Challenged

Humanity N.J Divya

#1, X.Anitha Sarafin

#2

Dept. of CSE, Misrimal Navajee Munoth Jain Engineering College, Chennai, Tamil Nadu, India1,2

ABSTRACT : Automatic clothing pattern recognition and matching costumes with appropriate color is a challenging

task for optically defective humanity. We developed a camera based prototype system for recognizing clothing patterns

such as plaid, striped, pattern-less and irregular. It identifies 11 clothing colors. A camera mounted upon a pair of

sunglasses is used to capture the clothing images. We propose a novel Radon Signature descriptor to capture the global

directionality of clothing pattern. SIFT represents the local structural features and STA descriptor is used to extract the

global statistical features from wavelet sub band and they are combined with local features to recognize the complex

clothing patterns. To evaluate the efficiency, we use the CCNY clothing pattern dataset which captures 627 images. To

assist blind persons to read the text written on the clothes, we have a conceived camera based assistive text reading

framework to track the object of interest within the rectangular arrangement and extract printed text information from

the clothes. Our Optical Character Recognition System(OCR) can effectively handle complex background and multiple

patterns, and extract text information from that costumes(object). This prototype system supports more independence in

their daily basis of the visually challenged people.

KEYWORDS: Radon Signature, SIFT(Scale Invariant Feature Transform),STA(Statistical Descriptor), CCNY

Dataset(City College Of New York),Assistive Text Reading Framework, OCR(Optical Character Recognition System)

,Local and Global image features.

I. INTRODUCTION

Due to large intra-class variations in patterns and designs in clothes cause difficulty to recognize the clothing

patterns so we employ the automatic camera based clothing pattern recognition system. Existing texture analysis

methods mainly focus on textures having large changes in viewpoint, orientation, and scaling image, but it presents less

intra-class pattern and intensity variations. The system contains three major components 1)sensors including a camera

to capture the clothing images, micro- phone for giving speech command input and speakers (or Bluetooth, earphone)

for audio output display 2) data capture and analysis performs command control, clothing pattern recognition, and color

identification uses a computer which can be a desktop or a wearable computer (e.g., a mini-computer or a smart phone)

and 3) audio outputs provides recognition results of clothing pattern, colors and also the system status.

Page 2: 17_EFFICIENT_COSTUME_ANALYSIS_AND_RECOGNITION_ACCEPTANCE_FOR_OPTICALLY_CHALLENGED_HUMANITY.pdf

ISSN(Online) : 2320-9801

ISSN (Print) : 2320-9798

International Journal of Innovative Research in Computer and Communication Engineering

An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015

Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015]

On 3rd

October 2015, Organized by

Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India

Copyright @ IJIRCCE www.ijircce.com 81

Fig 1.overview architecture of camera based clothing pattern recognition system

In an extension, our system can handle clothes with complex patterns and recognize clothing patterns into four

categories as plaid, striped, pattern-less and irregular. It is able to identify 11 colors in the image of the cloth (red,

orange, yellow, green ,cyan ,blue ,purple ,pink, black, grey and white). If clothes containing multiple colors, then the

dominant color is first spoken to users. To handle the large intra-class variations in the patterns, we propose a novel

descriptor called Radon Signature which captures the global directionality of clothing patterns. This paper presents a

prototype system of assistive text reading framework consists of three functional components: scene capture, data

processing and an audio output. The scene capture component collects scenes from the camera attached to a pair of

sunglasses containing objects of interest in the form of images . The data processing component including

1) object- of- interest detection which is able to extract the image of the object from the cluttered background

or other neutral objects in the camera view. 2) text localization used to obtain the image regions containing only the

text.3)text recognition used to transform the image-based text information into a readable codes. We use a mini laptop

as the processing device and the audio output component inform the recognized text codes to the blind user.

Fig 1.1 A Block Diagram of Assistive Text Reading Framework.

II. RELATED WORK

Some clothing patterns presented as a visual patterns are characterized by the repetition of few basic

primitives (e.g., plaids or striped). Local texture features are effective to extract the structural information of repetitive

Page 3: 17_EFFICIENT_COSTUME_ANALYSIS_AND_RECOGNITION_ACCEPTANCE_FOR_OPTICALLY_CHALLENGED_HUMANITY.pdf

ISSN(Online) : 2320-9801

ISSN (Print) : 2320-9798

International Journal of Innovative Research in Computer and Communication Engineering

An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015

Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015]

On 3rd

October 2015, Organized by

Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India

Copyright @ IJIRCCE www.ijircce.com 82

primitives. Global features includes the directionality and statistical properties of the clothing patterns are more stable

within the same category. Radon Signature, statistical descriptor (STA) and scale invariant feature transform (SIFT)

presents extraction of local and global texture feature.

Fig 2. Sample images of four categories of patterns.

2.1 Image Feature Extraction For Clothing Pattern Recognition

A. RADON SIGNATURE

Clothing images present large intra-class variations results in the major challenge of clothing pattern

recognition. In a global perspective view, the directionality of clothing patterns is more consistent across different

categories and it is used as an important property to distinguish a different clothing patterns. The clothing patterns of

plaid and striped are anisotropic. In contrast, the clothing patterns in the categories of pattern-less and irregular are

isotropic. To indicate this difference of directionality, we propose a novel descriptor called the Radon Signature to

characterize the directionality feature of clothing patterns.

Fig 2.1 (a)An Intensity Image of Clothing Pattern (b)Radon Transform performed on a maximum disk area

within the gradient map (c)Result of Radon Transform (d) Feature vector of Radon Signature.

Page 4: 17_EFFICIENT_COSTUME_ANALYSIS_AND_RECOGNITION_ACCEPTANCE_FOR_OPTICALLY_CHALLENGED_HUMANITY.pdf

ISSN(Online) : 2320-9801

ISSN (Print) : 2320-9798

International Journal of Innovative Research in Computer and Communication Engineering

An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015

Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015]

On 3rd

October 2015, Organized by

Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India

Copyright @ IJIRCCE www.ijircce.com 83

B. STATISTICS OF WAVELET SUB BANDS

DWT(Discrete Wavelet Transform) provides a generalization of a multi resolution spectral analysis which

decompose an image into a low frequency channel. Therefore, we extract the statistical features from wavelet sub bands

to capture global statistical information of images at different scales.

Each decomposition level includes four wavelet sub bands of original, horizontal, vertical and diagonal

components arranged from the close to the distant. Four statistical values calculated in each wavelet sub band are

concatenated to form the final descriptor.

C. SCALE INVARIANT FEATURE TRANSFORM

Detectors are used to detect interest of points by searching the local extrema in a scale space . Descriptors are

employed to compute the representations of interest of points based on their associated support regions.

Fig. 2.3. Process of local image feature extraction.

Page 5: 17_EFFICIENT_COSTUME_ANALYSIS_AND_RECOGNITION_ACCEPTANCE_FOR_OPTICALLY_CHALLENGED_HUMANITY.pdf

ISSN(Online) : 2320-9801

ISSN (Print) : 2320-9798

International Journal of Innovative Research in Computer and Communication Engineering

An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015

Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015]

On 3rd

October 2015, Organized by

Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India

Copyright @ IJIRCCE www.ijircce.com 84

In this paper, the evenly sampled uniform grids are used as interest points, are then represented by SIFT

descriptors which performs well in the context of image matching. The bag- of-words (BOW) method is further applied

to aggregate the extracted SIFT descriptors by labeling each SIFT descriptor as a visual word and counting a

frequencies of each visual word. The local feature representation of an image is represented as the histogram of the

quantized SIFT descriptors. We perform L2-norm and inverse document frequency (IDF) normalization for BOW

histograms.

D. TEXT RECOGNITION AND AUDIO OUTPUT

Stroke orientation and Edge distribution used to extract the text feature from the complex background. The

Cascade-Ada boost classifier confirms there is an existence of text information in an image patch but it cannot done for

the whole images. Text information in the image usually appears as the text strings in horizontal manner containing not

less than the three characters.

Text recognition is performed by off-the-shelf OCR prior to output of informative words from the localized

text regions. A text region first draws the minimum rectangular area for the accommodation of characters inside it, so

that border of the text region identifies the edge boundary of the text characters. OCR first assigned proper margin

areas to the text regions and binarizedb to segment the text characters from background view.

The recognized text codes are recorded in a script files. Then,we employ the Microsoft Speech Software

Development Kit to load a script files and presents an audio output of text information to the user. Blind users can

adjust speech rate, volume and tone according to their preferences.

Fig 2.4 Text written on clothes

III. ARCHITECTURE DIAGRAM

• A camera mounted upon a pair of sun glasses is used to capture the clothing image will be maintained in the

CCNY pattern dataset.

• The captured image is processed in gray scale matrix to plot the matrix data in histogram analysis.

• Filter the image using various illumination conditions like scaling, rotation and view point orientations.

• For feature representation, we perform 3-D transformation such as non rigid surface deformation to extract the

local image features.

• A texton dictionary is generated by clustering the extracted local features.

• we have a conceived camera based assistive text reading framework to track the object of interest

within the rectangular arrangement box and extract printed text information from the clothes.

Page 6: 17_EFFICIENT_COSTUME_ANALYSIS_AND_RECOGNITION_ACCEPTANCE_FOR_OPTICALLY_CHALLENGED_HUMANITY.pdf

ISSN(Online) : 2320-9801

ISSN (Print) : 2320-9798

International Journal of Innovative Research in Computer and Communication Engineering

An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015

Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015]

On 3rd

October 2015, Organized by

Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India

Copyright @ IJIRCCE www.ijircce.com 85

• Optical Character Recognition System uses portable label reader technique to handle complex background,

multiple patterns and extract text information from clothes.

• Using Flite library, audio output will be received by processing the text written on the image by portable

camera based label reader.

Fig 3.1 An Architecture diagram of clothing pattern recognition

3.1 PATTERN RECOGNITION:

The extracted global and local features are combined to recognize clothing patterns by using a support vector machines

(SVMs) classifier. The recognition of clothing color is implemented by quantizing clothing color in the HIS (hue,

saturation, and intensity) space. In the end, the recognition results of both clothing patterns and colors mutually provide

a more precise and meaningful description of clothes to users.In our system, we empirically set the size of the visual

vocabulary to 100. We apply three scaling levels to decompose clothing images..

3.2 CLOTHING COLOR IDENTIFICATION

Clothing color identification is based on the normalized color histogram of each clothing image in the HSI color space.

The key idea is to quantize color space based on the relationships between hue, saturation, and intensity. In particular,

for each clothing image, our color identification method quantizes the pixels in the image to the following 11 colors:

red, orange, yellow, green, cyan, blue, purple, pink, black, grey, and white. The detection of colors of white, black, and

gray is based on the saturation value S and intensity value I.We quantize the color of red in the range of 345◦−360◦ and

0◦−9◦, orange as 10◦−37◦, yellow as 38◦−75◦, green as 76◦−160◦, cyan as 161◦−200◦, blue as201◦−280◦, purple as

281◦−315◦, and pink as 316◦−344◦. The weight of each color is the percentage of pixels belonging to this color. If a

clothing image contains multiple colors, the dominant colors(i.e.,pixelslargerthan5%ofthewholeimage)will be output.

Page 7: 17_EFFICIENT_COSTUME_ANALYSIS_AND_RECOGNITION_ACCEPTANCE_FOR_OPTICALLY_CHALLENGED_HUMANITY.pdf

ISSN(Online) : 2320-9801

ISSN (Print) : 2320-9798

International Journal of Innovative Research in Computer and Communication Engineering

An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015

Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015]

On 3rd

October 2015, Organized by

Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India

Copyright @ IJIRCCE www.ijircce.com 86

The clothing patterns and colors mutually provide complementary information.

If the dominant colors present in a pair of clothing image are the same, then the two clothing images are

determined as color matched. The proposed color identification method achieves 99% matching accuracy in the

experiment evaluation.

IV. ALGORITHM AND TECHNIQUES

4.1 Radon Signature

Radon Signature method is employed to characterize the directionality feature of clothing pattern. Radon Signature

(Radon Sig) is based on the Radon transform [2.1(c)]which is commonly used to detect the principle orientation of an

image. The image is then rotated according to the dominant direction of orientation of the image to achieve rotation

invariance. Then the Radon transform of a 2-D function f (x,y) is defined as

Page 8: 17_EFFICIENT_COSTUME_ANALYSIS_AND_RECOGNITION_ACCEPTANCE_FOR_OPTICALLY_CHALLENGED_HUMANITY.pdf

ISSN(Online) : 2320-9801

ISSN (Print) : 2320-9798

International Journal of Innovative Research in Computer and Communication Engineering

An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015

Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015]

On 3rd

October 2015, Organized by

Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India

Copyright @ IJIRCCE www.ijircce.com 87

R(r,𝜃) = 𝑓 𝑥,𝑦 𝜹 𝒓 − 𝒙𝒄𝒐𝒔𝜽 − 𝒚𝒔𝒊𝒏𝜽 𝒅𝒙𝒅𝒚+∞

−∞

+∞

−∞ (1)

where r is the perpendicular distance of a projection line goes to the origin and θ is the angle of the projection line, as

shown in Fig. 2.1(b). To reduce the intensity variations in the projections we use the Sobel operator to compute the

gradient map similar to that f(x,y) calculations. The directionality of an image can be represented by Var(r,θi), the

variances of under a certain projection direction θi:

var(r,𝜃𝑖) =1

𝑁 (𝑅 𝑟𝑗,𝜃𝑖 − 𝜇(𝑟,𝜃𝑖))2𝑁−1

𝑗=0 (2)

𝜇(𝑟,𝜃𝑖)=1

𝑁 (𝑅 𝑟𝑗,𝜃𝑖 𝑁−1

𝑗=0 (3)

where R(rj,θi) is the projection value at perpendicular distance of rj and projection direction of θi; μ(r,θi) is the

expected value of R(r,θi); N is the number of sampling bins in each projection line. The Radon Sig is formed by the

variances of r under all sampling projection directions:

[Var(r,θ0), Var(r,θ1),..., Var(r,θT-1)]

where T, is the number of sampling projection directions.

The plaid patterns have two principle orientations(Two dominant peak value in the Radon Sig) and the striped

ones have one principle orientation similar to that pattern-less and the irregular images have no dominant direction,

but the directionality of the irregular image presents much larger variations than that of the pattern-less image

4.2 Discrete Wavelet Transform

Discrete wavelet transform (DWT) decompose the clothing image and provides a multi resolution spectral

analysis of images at different scale. The discrete wavelet transform (DWT) is a linear transformation that operates on a

data vector in the image whose length is an integer power of two and transforming it into a numerically different vector

of the same length. It acts as a tool to separate the data into a different frequency components, and then studies each

component with a resolution that matches with the image scale. The main feature of DWT is a multi-scale

representation of function. By using the wavelets, multi-scale representation of function can be analyzed at various

levels of resolution.

4.3 Scale Invariant Feature Transform

Scale Invariant Feature Transform (SIFT) used for image matching is robust to variation in illumination. SIFT

collect features extracted from images which help in reliable matching of the same object in different perspective. The

extracted features from image are invariant to scale and orientation, and are highly distinctive. The first step computes

the locations of potential of interest points in the overall image by detecting the maxima and minima location by

applying a set of Difference of Gaussian (DoG) filters at different scales. Then, these locations are refined by

discarding points of low contrast.

V. SYSTEM AND INTERFACE DESIGN

The camera-based clothing pattern recognition system aided for blind people integrates a camera, a

microphone, a computer, and a Bluetooth earpiece for audio description which includes clothing pattern and color

information. A camera mounted upon a pair of sunglasses is used to capture the clothing images

The clothing patterns and colors are described to blind users by a verbal display and it contains minimal

distraction to hearing. The system can be controlled by a speech input from a microphone. In order to interact with a

blind users, speech command input from a microphone are used to enable the function selection and system control.

Page 9: 17_EFFICIENT_COSTUME_ANALYSIS_AND_RECOGNITION_ACCEPTANCE_FOR_OPTICALLY_CHALLENGED_HUMANITY.pdf

ISSN(Online) : 2320-9801

ISSN (Print) : 2320-9798

International Journal of Innovative Research in Computer and Communication Engineering

An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015

Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015]

On 3rd

October 2015, Organized by

Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India

Copyright @ IJIRCCE www.ijircce.com 88

The interface design includes some basic functions and high priority commands are shown

Fig 5.1 System interface design for the proposed camera based clothing pattern recognition system by using speech

commands.

Basic functions: A blind user can perform some basic function to recognize the clothing pattern using speech

command or by clicking the button. The recognition results will be announced to the blind user as an audio output such

as recognized, not recognized, and start a new function. The recognized function includes the next level functions to

announce the recognized clothing pattern and dominant colors present in the cloth will be spoken to the user , Repeat

result functions repeats the recognized clothing pattern result and save result function used to save the clothing image

associated with pattern and color information are stored in the computer.

High priority commands: A blind user can set the system configuration by clicking several high priority speech

commands such as system restart, turn-off system, stop function (i.e., abort current task),adjust speaker volume, control

commands for speech adjustment(e.g., louder, quieter, slower, and faster), and help option. The high priority commands

can be used at any time. If user wants help option enabled through speech command, then the cloth pattern recognition

system will respond to user with options that are associated with the current function. Bone conducted earphones or

small wireless Bluetooth speakers can be employed to protect privacy information of recognition results and minimize

background sounds. The battery level will be checked automatically and an audio warning is provided if the battery

level is low.

Audio output: Operating system speech facility in modern portable computer systems and smart phone is utilized for

audio display. We currently use a Microsoft Speech Software Development Kit which supports various number of

scripts. The number of system configuration options vary according to user preference such as speech rate, volume and

voice gender

.

VI. CONCLUSION

To evaluate the performance of the system, we maintain two dataset

1) CCNY Clothing Pattern dataset which is invariant to large changes in intra-class variations.

2) UIUC Texture dataset to validate the generalization of multi resolution spectral analysis.

CCNY clothing pattern dataset includes 627 images which of four different categories such as plaid, striped, pattern-

less, and irregular with 156, 157, 156, and 158 images in each category.

UIUC texture dataset contains 1000 unc alibrated and unregistered images .We maintain 25 texture classes with 40

images for each class in the dataset. The texture images present rotation, scaling, view-point change, and non- rigid

surface deformation under various conditioning of lightings.

Page 10: 17_EFFICIENT_COSTUME_ANALYSIS_AND_RECOGNITION_ACCEPTANCE_FOR_OPTICALLY_CHALLENGED_HUMANITY.pdf

ISSN(Online) : 2320-9801

ISSN (Print) : 2320-9798

International Journal of Innovative Research in Computer and Communication Engineering

An ISO 3297: 2007 Certified Organization Vol.3, Special Issue 8, October 2015

Second National Conference on Emerging Trends and Intelligence Technologies [ETIT 2015]

On 3rd

October 2015, Organized by

Dept. of CSE, Anand Institute Of Higher Technology, Kazhipathur, Chennai-603103, India

Copyright @ IJIRCCE www.ijircce.com 89

In our implementation, the training set is selected as a fixed size random subset of each class and all the remaining

images are used as the testing set. To eliminate the dependence of the results on the particular training images we used,

the system will report the average of the classification rates obtained from the randomly selected training sets. The

recognition performance is measured by the average classification accuracy. A combination of multiple features may

obtain better results than any individual feature channel. This system provides a new functions such as high priority

commands and performs some basic functions to improve the life quality of blind and visually impaired people.

REFERENCES

[1]D. Dakopoulos and N. G. Bourbakis, “Wearable obstacle avoidance electronic travel aids for the blind: A survey,” IEEE Trans. Syst.,Man,

Cybern.C, vol. 40, no. 1, pp. 25–35,Jan 2010.

[2]F.Hasanuzzaman, X.Yang, and Y. Tian, “Robust and effective component based banknote recognition for the blind,” IEEE Trans. Syst.,Man, Cybern.C, vol. 42, no. 6, pp 1021–1030, Nov. 2012.

[3]K. Khouzani and H. Zaden,“Radon transform orientation estimation for rotation invariant texture analysis,” IEEE Trans. Pattern Anal. Mach.

Intell.,vol.27, no.6, pp. 1004–1008, Jun. 2005. [4]S. Lazebnik, C. Schmid, and J. Ponce,“A sparse texture representation using local affine regions,” IEEE Trans. Pattern Anal. Mach. Intell., vol.

27,no. 8, pp. 1265–1277, Aug. 2005.

[5]Vasanthi.G and Ramesh Babu.Y, ”Vision Based Assistive System for Label Detection With Voice Output”, ”International Journal Innovative Research in Science, Engineering and Technology”,Vol-3,January 2014

[6]Xiaodong Yang, Shuai Yuan and Yingli Tian, ”Recognizing Clothes Pattern For Blind People By Confidence Margin Based Feature

Combination”,”City college of Newyork”,2014