ABSTRACT
Face recognition has become a popular
area of research in computer vision, it is typically used in network security
systems and access control systems but it is also useful in other multimedia
information processing areas. One of its application is criminal face
identification. Criminal record generally contains personal information about
particular person along with the photograph. To identify any criminal we need
some identification regarding particular person or persons, which are given by
eyewitnesses. Based on the details given by the eyewitnesses, the further
investigation would be carried out. In most cases the quality and resolution of
the recorded image segments is poor and hard to identify a face. In this study,
we have classified image processing operations into three categories; low,
medium and high level to process and analyze a given face. This study presents
better results than conventional methods in use relating to the face
recognition process that are used in criminal identification.
Keywords:
Face Identification, image processing, Biometrics, Face clippings.
CHAPTER
ONE
INTRODUCTION
Face Identification is a technique
that is mainly used to identify criminals based on the clues given by the
eyewitnesses. Based on the clues we develop an image by using the image that we
have in our database and then we compare it with the images already we have. To
identify any criminals we must have a record that generally contains name, age,
location, previous crime, gender, photo, etc. The primary task at hand is,
given still or video images require the identification of the one or more
segmented and extracted from the scene, where upon it can be identified and
matched. The word image is defined as an exact or analogous representation of a
being or a thing. The image or monochrome image such as black and white
paragraph is represented as twodimensional light intensity function f(x,y)
where x and y denotes spatial co-ordinates. A digital image is an image of f
(x, y) that has been digitized both in spatial co-ordinate and brightness. The
elements of such a digital array are called image elements, picture elements or
pixels.
Biometric technologies [3] have been evolved
as an enchanting solution to perform secure identification and personal
verification. The need for highly secure identification and personal
verification technologies is becoming apparent as the level of security
breaches and transaction fraud increases. The increasing use of biometric
technologies in high security applications and beyond has created the
requirement for highly dependable face recognition systems. The Face
recognition system is used to verify an identity of a person by matching a given
face against a database of known faces. It has become alternative to
traditional identification and authentication methods such as the use of keys,
ID cards and passwords. Face recognition involves computer recognition of
personal identity based on geometric or statistical features that are derived
from the face images [8]. Even though human can detect and identify faces in a
scene easily, building an automated system is challenging. Face recognition
technology can be applied to a wide variety of application areas including
access control for PCs, airport surveillance, private surveillance, criminal
identification and for security in ATM
transactions. In addition, face recognition system is moving towards the
next-generation smart environment where computers are designed to interact more
like humans. In recent years, considerable progress has been made in the area
of face recognition with the development of many other useful techniques. The
advances in computing technology have facilitated the development of real-time
vision modules that interact with humans in recent years. Examples abound,
particularly in biometrics and human computer interaction as the information
contained in faces needs to be analyzed for systems to react accordingly [7].
For biometric systems that use faces as nonintrusive input modules, it is
imperative to locate faces in a scene before any recognition algorithm can be
applied. An intelligent vision based user interface should be able to tell the
focus of the user (i.e., where the user is looking at) in order to respond
accordingly [1]. To detect facial features accurately for applications such as
digital cosmetics, faces need to be located and registered first to facilitate
further processing. It is evident that face detection plays an important and
critical role for the success of any face processing systems. The face
detection problem is challenging as it needs to account for all the possible
appearance variation caused by change in illumination, facial features, occlusions,
etc. In addition, it has to detect faces that appear at different scale, pose,
with in plane rotations. Often the size of the image is very large, the
processing time has to be very small and usually real-time constraints have to
be met. Therefore, during the last decades there has been an increasing
interest in the development and the use of parallel algorithms in image
processing. Face detection is attached with finding whether or not there are
any faces in a given image (usually in gray scale) and, if present, return the
image location and content of each face. This is the first step of any fully
automatic system that analyzes the information contained in faces (e.g.,
identity, gender, expression, age, race and pose).This work focuses on how to make
parallel computations by partitioning the image into manageable and meaningful
parts for efficient calculations and results.
CLASSIFICATIONOF
IMAGE OPERATIONS
Image processing is referred to
processing of a 2D picture by a computer. An image may be considered to contain
sub-images sometimes referred to as regions-of-interest, ROIs, or simply
regions. This concept reflects the fact that images frequently contain
collections of objects each of which can be the basis for a region. In a
sophisticated image processing system it should be possible to apply specific
image processing operations to selected regions. Thus one part of an image
(region) might be processed to suppress motion blur while another part might be
processed to improve color rendition. Image processing operations can be
classified as low-level, intermediate-level and high-level [2]. Based on this
classification, it is possible to define a skeleton library for image
operations in order to carry out image recognition operations. Ø Low-level image operations. Ø Intermediate-level image operations. Ø High-level image operations.
Skeletons for image operations: It is possible to use the data-parallelism
paradigm with the masterslave approach for low-level, intermediate-level and
high-level image processing operations [5]. A master processor is selected for
splitting and distributing the data to the slaves. The master can also process
a part of the image (data). Each slave processes its
received part of the image (data) and
then, the master gathers and assembles the image (data) back. Before going to
processing an image, it is converted into a digital form. Digitization includes
sampling of image and quantization of sampled values. After converting the
image into bit information, processing is performed. These processing
techniques are image preprocessing, Image enhancement, Image reconstruction,
and Image compression.
Image
Preprocessing:
Preprocessing functions involve those operations that are normally required
prior to the main data analysis and extraction of information, and are
generally grouped as radiometric or geometric corrections.
Image
Enhancement: It
refers to accentuation, or sharpening, of image features such as boundaries, or
contrast to make a graphic display more useful for display & analysis. This
process does not increase the inherent information content in data. It includes
gray level & contrast manipulation, noise reduction, edge crispening and
sharpening, filtering, interpolation and magnification, pseudo coloring, and so
on.
Image
Restoration: It is
concerned with filtering the observed image to minimize the effect of
degradations. Effectiveness of image restoration depends on the extent and
accuracy of the knowledge of degradation process as well as on filter design.
Image restoration differs from image enhancement in that the latter is
concerned with more extraction or accentuation of image features.
Image
Compression: It is
concerned with minimizing the number of bits required to represent an image.
Application of compression are in broadcast TV, remote sensing via satellite,
military communication via aircraft, radar, teleconferencing, facsimile
transmission, for educational & business documents , medical images that
arise in computer tomography, magnetic resonance imaging and digital radiology,
motion , pictures ,satellite images, weather maps, geological surveys and so
on.
Image Classification and Analysis:
These operations are used to digitally identify and classify pixels in the
data. Classification is usually performed on multi-channel data sets and this
process assigns each pixel in an image to a particular class or theme based on
statistical characteristics of the pixel brightness values. There are a variety
of approaches taken to perform digital classification. The two generic
approaches which are used most often are supervised and unsupervised
classification.
FACE
IDENTIFICATION TECHNIQUE
Face identification is a term that
includes several sub problems. The technique for face identification comprises
of three steps: face detection, feature extraction and face recognition. Face
Detection: Face detection is defined as the process of detecting faces from
images and scenes. So, the system positively identifies a certain image region
as a face. This procedure has many applications like face tracking, pose
estimation or compression.
EXISTING
SYSTEM
Criminal record generally contains
personal information about particular person along with photograph. To identify
any criminal we need some identification regarding person, which are given by
eyewitnesses [6]. Based on the details given by eyewitnesses, the criminal who
did the crime will be identified manually.
PROBLEMS
IN EXISTING SYSTEM
a. In most cases the quality and
resolution of the recorded image segments are poor and hard to identify a face.
b. If a eyewitnesses observe a
criminal only from single direction, it may not be possible to recognize him.
c. The photograph, which is a hard
copy, cannot be able to divide or split into different modules. So it is very
difficult to find, unless we get full-fledged details.
d. Some times the eyewitness may not
be able to draw, the face of criminal [2].
e. Sometimes, if we maintain the
criminal details manually and physically. After a time span, the photographs
and other details may tend to tear out.
THE
PROPOSED SYSTEM AND IMPLEMENTATION
Feature extraction domain has plenty
of collection of generalized face features from several images of the same
subject. Then, each face image is processed, features are extracted and the
collection of features are analyzed and combined into a single generalized features
collection, which is written to the database. The face is our primary focus of
attention in social inter course playing a major role in conveying
identification and emotion. Although the ability to infer intelligence or
character from facial appearance is a guess but still the human ability to
recognize faces is remarkable. This analogy would give us enough scope to
envisage a new algorithm. There are mainly three important ways in construction
of the face i.e., by using the eyewitness function, adding details and clipping
image. This offers us a face as finally identification parameter to know who
has committed the crime.
TOPIC: AN AUTOMATED TECHNIQUE FOR CRIMINAL FACE IDENTIFICATION USING BIOMETRIC APPROACH
Format: MS Word
Chapters: 1 - 5
Delivery: Email
Delivery: Email
Number of Pages: 65
Price: 3000 NGN
In Stock
No comments:
Post a Comment
Add Comment