Vous êtes sur la page 1sur 28

CHAPTER 1

INTRODUCTION
Web-scale image search engines mostly use keywords as queries and rely on
surrounding text to search images. It is well known that they suffer from the ambiguity of
query keywords. For example, using apple as query, the retrieved images belong to
different categories, such as red apple, apple logo, and apple laptop. Online image re
ranking has been shown to be an effective way to improve the image search results . Major
internet image search engines have since adopted the re-ranking strategy. Its diagram is
shown in Figure 1. Given a query keyword input by a user, according to a stored word-image
index file, a pool of images relevant to the query keyword are retrieved by the search engine.
By asking a user to select a query image, which reflects the users search intention, from the
pool, the remaining images in the pool are re-ranked based on their visual similarities with
the query image. The visual features of images are pre-computed offline and stored by the
search engine. The main online computational cost of image re-ranking is on comparing
visual features. In order to achieve high efficiency, the visual feature vectors need to be short
and their matching needs to be fast.Another major challenge is that the similarities of low
level visual features may not well correlate with images high-level semantic meanings
which interpret users search intention. To narrow down this semantic gap, for offline image
recognition and retrieval, there have been a number of studies to map visual features to a set
of predefined concepts or attributes as semantic signature However, these approaches are
only applicable to closed image sets of relatively small sizes. They are not suitable for online
web- based image re-ranking. According to our empirical study, images retrieved by 120
query keywords alone include more than 1500 concepts. Therefore, it is difficult and
inefficient to design a huge concept dictionary to characterize highly diverse web images.

CHAPTER 2

LITERATURE SURVEY
2.1 IMAGE RETRIEVAL USING SEMANTICS OF QUERY IMAGE (2013)
/(SIDDANAGOWDA G. R)

The fields of image-based content retrieval and automatic image annotation are becoming
more and more relevant to the ways in which large libraries of digital media are stored and
accessed. As multimedia and imaging technologies improve, so does the wealth of visual data
in libraries, necessitating an automated mechanism with which to index and thus access
content by a variety of means Content based image retrieval for general-purpose image
databases is a highly challenging problem because of the large size of the database, the
difficulty of understanding images, both by people and computers, the difficulty of
formulating a query, and the issue of evaluating results properly. A number of generalpurpose image search engines have been developed The common ground for CBIR systems is
to extract a signature for every image based on its pixel values and to define a rule for
comparing images. The signature serves as an image representation in the view of a CBIR
system. Content Based Image Retrieval (CBIR) is any technology that in principle helps to
organize digital image archives by their visual content. By this definition, anything ranging
from an image similarity function to a robust image annotation engine falls under the purview
of CBIR the most common form of CBIR is an image search based on visual Features
Re-ranking with user interactions or active re-ranking is introduced in this system. Collecting
information from users to obtain the specified semantic space. Localizing the visual
characteristics of the users intention in this specific semantic space.

Limitations in this paper is that most of the image retrieval systems present today are textbased, in which images are manually annotated by text-based keywords and when we query
by a keyword, instead of looking into the contents of the image, this system matches the
query to the keywords present in the database. Re-ranking with user interactions Collecting
information from users to obtain the specified semantic space. Localizing the visual
characteristics of the users intention in this specific semantic space .These disadvantages of
text-based image retrieval techniques call for another relatively new technique known as
Content-Based Image Retrieval (CBIR). CBIR is a technology that in principle helps
2

organize digital image archives according to their visual content. This system distinguishes
the different regions present in an image based on their similarity in color, pattern, texture,
shape, etc. and decides the similarity between two images by reckoning the closeness of these
different regions.

2.2 CONTENT-BASED IMAGE RETRIEVAL. (YEAR 2012) / (RITENDRA DATTA)


A third commercial CBIR system of interest is IMatch5. Rather than being a
standalone.CBIR system, IMatch is a commercial image management system containing
CBIRComprehensive surveys exist on the topic of content-basedimage retrieval (CBIR) both
of which are primarily on publications prior to the year 2000. Surveys also exist onclosely
related topics such as relevance feedback highdimensional indexing of multimedia data
applications of content-based image retrieval to medicine and applications to art and cultural
imaging One of the reasons for writing this survey is that the field has grown tremendously
after 2000, not just in terms of size, but also in the number of new directions explored. To
validate this, we conducted a simple test. We searched for publications containing the phrases
Image Retrieval using Google Scholar [32] and the digital libraries of ACM, IEEEand
Springer, within each year from 1995 to 2004.In this paper, we review recent work (i.e., year
2000 onwards) in automatic image retrieval and annotation, with a focus on real-world usage
and applications of the same. We leave out retrieval from video sequences and text caption
based image search from our discussion. The rest of this paper is arranged as follows: Some
of the key ideas behind the proposed approaches are discussed in Sec. 2. This leads us to a
discussion on some of the most desirable features of real-world image indexing systems, in ,
learned through our own experiences with retrieval system The authors argue that this
compact representation is more efficient than highdimensionalhistograms in terms of search
and retrieval, and it also gets around some of the drawbacks associated with earlier
propositions such as dimension reduction and colormoment descriptors. In [35], a multiresolution histogram capturing spatial image information has been shown to be effective in
retrieving textured images, while retaining the typical advantages of histograms.
Limitations in this paper is that an image database is created, after which one can select to
import jpg-images into the database .Import of 10 images takes approximately 60
secondWhile a text-based search-engine can successfully retrieve documents without
understanding the content, there is usually no easy way for a user to give a low-level
description of what image she is looking for. Even if she provides an example to search for
3

images with similar content, most current algorithms fail to accurately relate its high level
concept, or the semantics of the image, to its lower level content. The problem with these
algorithms is their reliance on visual similarity in judging semantic similarity. Moreover,
semantic similarity is a highly subjective measure.Shape is a key attribute of segmented
image regions, and its efficient and robust representation plays an importantrole in retrieval.
A shape similarity measure using discrete curve evolution to simplify contours, is discussed
in [59].Doing this contour simplification helps to remove noisy and irrelevant shape features
from consideration.
2.3 INTENT SEARCH: CAPTURING USER INTENTION FOR ONE-CLICK
INTERNET IMAGE SEARCH.(YEAR 2014) /(XIAOOU TANG)
Many commercial Internet scale image search engines use only keywords as queries.Users
type querykeywords in the hope of finding a certain type of images.The search engine returns
thousands of images ranked bythe keywords extracted from the surrounding text. It is
wellknown that text-based image search suffers from theambiguity of query keywords. The
keywords provided byusers tend to be short.Adaptive Similarity is proposed, motivated by
the ideathat a user always has specific intention whensubmitting a query image. For example,
if the usersubmits a picture with a big face in the middle, mostprobably he/she wants images
with similar faces andusing face-related features is more appropriate. In ourapproach, the
query image is first categorized intoone of the predefined adaptive weight categories,such as
portrait and scenery.Keyword expansion. Query keywords input by user stand to be short
and some important keywords maybe missed because of users lack of knowledge on the
textual description of target images. In our approach ,query keywords are expanded to
capture users search intention, inferred from the visual content of query images, which are
not considered in traditional keyword expansion approaches.The image pool retrieved by
text-based search accommodates images with a large variety of semantic meanings and the
number of images related to the query image is small. In this case, re-ranking images in the
pool is not very effective. Thus, more accurate query by keywords is needed to narrow the
intention and retrieve more relevant images.Visual query expansion. One query image is not
diverse enough to capture the users intention. In Step 2, a cluster of images all containing the
sameexpanded keywords and visually similar to thequery image are found.All four of these
steps are automatic with only one clickin the first step withoutincreasing users burden. This
makes it possible for Internet scale image search by both textual and visual content with a
very simple user interface.
4

The drawback is that given a query image, similarity retrieval involves searching the database
for similar color distributions as the input query. Since the number of representative colors is
small, one can first search the database for each of the representative colors separately, and
then combine the results. Calculating Euclidian distance in a color space can do searching for
individual colors very efficiently. Each pixel is associated to a specific histogram bin only on
the basis of its own color, and color similarity across different bins or color dissimilarity in
the same bin are not taken into account.Since any pixel in the image can be described by
three components in a certain color space.
2.4 DESCRIPTOR LEARNING FOR EFFICIENT RETRIEVAL (JAMES PHILBIN) /
(2010)
There have been several recent applications of distance learning to classification problems
however these methods assume clean, labelled data indicating pairs of points that belong to
the same class and pairs that belong to different classes. In our task, even when the same
object appears in two images, the images typically have different backgrounds and there is
anon-trivial transformation between the views of a common object, so we cannot simply
classify images as being matching or non-matching. At the same time the number of
individual descriptors per image and the complexity of the correspondence problem between
them means that manually labelling the sets of matching and non-matching descriptors would
be unacceptably burdensome. Therefore, in this work, we introduce a new method for
generating training data from a corpus of unlabelled images using standard techniques from
multi-view geometry. In contrast to Hua et al. [16], who also generated training pairs from
unlabelled image data via patches matched by the Photo Tourism system here we adopt a
much cheaper pairwise image measure which doesnt require us to compute a global bundle
adjustment over many image pairs. Thus, we can train on patches of objects that appear in as
few as two images. Previous works in distance learning use two categories of point pairs for
training: matching and non-matching, typically derived from known class labels. In this
work, we show that we can significantly improve performance by forming two nonmatching categories: random pairs of features; and those which areeasily confused by a
baseline method. We adopt a margin-based cost function todistinguish these three categories
of points, and show that this gives improvedperformance more than using non-margin-based
methods To optimize this cost function, a fast, stochastic, online learning procedure is used
5

that permits the use of millions of training pairs. We will show that non-linear projection
methods, previously used for hand-written digit classification [13], perform better than the
linear projections previously applied to computer vision distance learning .The next section
motivates the distance learning task by showing that retrieval performance is significantly
worse using standard quantized descriptors than when a much slower, exhaustive search
procedure is applied to the raw SIFT descriptors this indicates the potential gain achievable
from better clustering. After describing in Section 3 how we automatically generate our
training data, we set out our learning methods in Section 4 and then conclude with results
andDescriptor Learning for Efficient Retrieval a discussion. Improved performance is
demonstrated over SIFT descriptors on standard datasets with learnt descriptors.In practice, it
is not possible to fully separate these pairwise distances because of noise in the training data
and restricted model complexity, so instead a margin based approach will be used which
encourages the distance between the three classes of point pairs to separate without enforcing
the distance ordering as a hard constraint. The loss function for this situation is The first
margin aims to separate the positive and NN negative point pairs confused by SIFT in the
original space. The second margin applies a force to the random negatives to keep them
distant from the positive pairs ideally the overlap in histograms between the positive and
random negative point pairs should be small.

Limitations in this paper is that image pair is chosen at random from the dataset. A set of
putative matches is computed between the image pair. Each putativematch consists of a pair
of elliptical features, one in each image, that pass Lowes second nearest neighbour ratio test
[18] on their SIFT descriptors; RANSAC is used to estimate an affine transform between the
images together with a number of inliers consistent with that transform. Point pairs are only
taken from image matches with greater than 20 verified inliers. The ratio test ensures that
putative matches are distinctive for that particular pair of images. This procedure generates
three sets of point pairs.These are the point pairs found as inliers by RANSAC. These are
pairs marked as outliers by RANSACthey are generally close in descriptor space as they
were found to be descriptor-space nearest neighbours between the two images, but are
spatially inconsistent with the best-fitting affine transformation found between the images.

2.5 INTERACTIVE ONLINE IMAGE SEARCH RE-RANKING


(D.MANASASPANDHANA) / ( YEAR 2008)
In this paper an automatic annotation method by hybriding decision tree (DT) and support
vector machine (SVM) is proposed and a novel inverted file is used to rank the search result.
Experiments of both word search and image search in a Corel dataset and a Yahoo! dataset
are performed. The preliminary result is satisfied and promising. With the number of digital
images in the WWW increasing explosively, efficient image search in large scale datasets
has attracted great interest from both academia and industry. However, image retrieval is
currently far less efficient than text retrieval because images are unstructured and much more
difficult to process than texts. The approaches of retrieving and ranking images from largescale datasets.Text-based approaches the search engine returns corresponding images by
processing the associated textual information, such as file name, surrounding text, URL, etc.
,according to keywords input by users. Most of popular commercial Web image search
engines like Google and Yahoo! adopt this method. While text-based search techniques have
been verified to perform well in textual documents, they often result in mismatch when
applied to the image search. The reason is that metadata cannot represent the semantic
content of images.Content-based approaches: these arch engine extracts semantic information
from image content features, such as color, shape, texture, spatial location of objects in
images, etc . The extracted visual information is natural and objective, but completely ignores
the role of human knowledge in the interpretation process. As the result, a red flower may be
regarded as the same as a rising sun, and a fish the same as an airplane etc.We tested the idea
of re-ranking on six text queries to a large-scale web image search engine, Google Image
Search which has been on-line since July 2001. With the huge amount of indexed images,
there should be large varieties of images, and testing on the search engine of this scale will be
more realistic than on an in house ,small-scale web image search system. Six queries are
chosen, as listed in Table 1, which are among image categories in Corel Image Database.
Corel Database is often used for evaluating image retrieval [5] and classification. Each text
query is typed into Google Image Search, and top 200 entries are saved for evaluation. The
default browsing setting for Google Image Search is to return 20entries per page, and thus
200 entries takes usersEach entry in the rank list contains a filename, image size, image
resolution, and URL that points to the image. We build a web crawler program to fetch and
save both the image and associated HTML document for each entry.

Drawbacks is that the proposed re-ranking procedure where the dashed box represents the
Relevance Model Re-ranking box in the users input a query consisting of keywords to
describe the pictures they are looking for, and a web image search engine returns a rank list
of images. The same query is also fed into a web text search engine, and retrieved documents
are used to estimate the relevance model for the query Q. We then calculate the KL
divergence between the relevance model and the unigram mode of each document D
associated with the image I in the image rank list, and re-rank the list according to the
divergence. The re-ranking process based on relevance model still can improve the
performance, suggesting that global informationfrom the document can provide additional
cues to judge the relevance of the image. Internet users are usually with limit time and
patience, and high precision at top-ranked documents will save user a lot of efforts and help
them find relevant images not easily and quickly.

CHAPTER 3

SYSTEM ANALSIS
3.1 EXISTING SYSTEM
In this paper, a novel framework is proposed for web imagere-ranking. Instead of
constructing a universal conceptdictionary, it learns different visual semantic spaces for
differentquery keywords individually and automatically. Believe that the semantic space
related to the images to bere-ranked can be significantly narrowed down by the
querykeyword provided by the user. For example, if the querykeyword is apple, the
semantic concepts of mountainsand Paris are unlikely to be relevant and can be
ignored.Instead, the semantic concepts of computers and fruit will be used to learn the
visual semantic space related toapple. The query-specific visual semantic spaces canmore
accurately model the images to be re-ranked, sincethey have removed other potentially
unlimited number of non-relevant concepts, which serve only as noise and deteriorate the
performance of re-ranking in terms of both accuracy and computational cost. The visual
features of images are then projected into their related visual semantic spaces to get semantic
signatures. At the online stage, images are re-ranked by comparing their semantic signatures
obtained from the visual semantic space of the query keyword. Our experiments show that
the semantic space of a query keyword can be described by just 20

30 concepts (also

referred as reference classes in our paper). Therefore the semantic signatures are very short
and online image re-ranking becomes extremely efficient. Because of the large number of
keywords and the dynamic variations of the web, the visual semantic spaces of query
keywords need to be automatically learned. Instead of manually defined, under our
framework this is done through keyword expansions. Another contribution of the paper is to
introduce a large scale benchmark database1 with manually labeled ground truth for the
performance evaluation of image re-ranking. It includes 120; 000 labeled images of around
1500 categories (which are defined by semantic concepts) retrieved by the Bing Image
Search using 120 query keywords. Experiments on this benchmark database show that
20% 35% relative improvement has been achieved on re-ranking precisions with much
faster speed by our approach, compared with the state-of-the-art methods. Our experiments
show that the semantic space of a queryThis is the most common form of text search on the
Web. Most search engines do their text query and retrieval using keywords. The keywords
based searches theyusually provide results from blogs or other discussion boards. The user
9

cannot have a satisfaction with these results due to lack of trusts on blogs etc.low precision
and highrecall rate. In early search engine that offered disambiguation to search terms. User
intention identification plays an important role in the intelligent semantic search engine.

3.2 DISADVANTAGES OF EXISTING SYSTEM


Traditional and

common methods of image retrieval utilize

some method of

adding metadata such as captioning', keywords, or descriptions. Manual image annotation is


time-consuming, laborious and expensive. Recently, for general image recognition and
matching, there have been a number of works on using predefined concepts or attributes as
image signature.Mapped visual features to a universal concept dictionary.considering the
huge collection of images present, it is not feasible to manually annotate them. Secondly, the
rich features present in an image cannot be described by keywords completely. User intention
is not considering in the already existing system. Re ranking methods usually fail to capture
the users intention when the query term is ambiguous. Text-based search techniques are
problematic when applied to the image search because of the mismatching between images
and their associated textual information Textual information is insufficient to represent the
semantic content of the images.

3.3 PROPOSED SYSTEM


The semantic web based search engine which is also called as Intelligent Semantic Web
Search Engines. We use the power of xml meta-tags deployed on the web page to search the
queried information. The xml page will be consisted of built-in and user defined tags. Here
propose the intelligent semantic web based search engine. We use the power of xml metatags deployed on the web page to search the queried information. The xml page will be
consisted of built-in and user defined tags. The metadata information of the pages is extracted
from this xml into rdf. our practical results showing that proposed approach taking veryless
time to answer the queries while providing more accurate informationContent-based image
retrieval uses visual features to calculate image similarity. Relevance feedback was widely
used to learn visual similarity metrics to capture users search intention. However, it required
more users effort to select multiple relevant and irrelevant image examples and often needs
online training. For a web-scale commercial system, users feedback has to be limited to the
minimum with no online training. Cui et al. proposed an image re-ranking approach which
10

limited users effort to just one-click feedback. Such simple image re-ranking approach has
been adopted by popular web-scale image search engines such as Bing and Google recently,
as the find similar images function. The key component of image re-ranking is to compute
the visual similarities between images. Many image features have been developed in recent
years. However, for different query images, low-level visual features that are effective for
one image category may not work well for another. To address this, Cui et al. classified the
query images into eight predefined intention categories and gave different feature weighting
schemes to different types of query images. it was difficult for only eight weighting schemes
to cover the large diversity of all the web images. It was also likely for a query image to be
classified to a wrong category.Mapped visual features to a universal concept
dictionary.Lampert et al. used predefined attributes with semantic meanings to detect novel
object classes. Some approaches transferred knowledge between object classes by measuring
the similarities between novel object classes and known object classes (called reference
classes). All these concepts/attributes/reference-classes were universally applied to all the
images and their training data was manually selected. For each query keyword, a multi-class
classifier on low level visual features is trained from the training sets of its reference classes
and stored offline. If there are K types of visual features, one could combine them to train a
single classifier. It is also possible to train a separate classifier foreach type of features. Our
experiments show that the latter choice can increase the re-ranking accuracy but will also
increase storage and reduce the online matching efficiency because of the increased size of
semantic signatures.

ADVANTAGES
The query-specific visual semantic spaces can more accurately model the images to be reranked automatically.Visual features of images are then projected into their related visual
semantic spaces to get semantic signatures.Images are re-ranked by comparing their semantic
signatures obtained from the visual semantic space of the query keyword. An image may be
relevant to multiple query keywords. Therefore it could have several semantic signatures
obtained in different semantic spaces. According to the wordimage index file, each image in
the database is associated with a few relevant keywords. For each relevant keyword,
asemantic signature of the image is extracted by computing the visual similarities between
the image and the reference classes of the keyword using the classifiers trained in the
11

previous step. The reference classes form the basis of the semantic space of the keyword. If
an image has N relevantkeywords, then it hasN semantic signatures to be computed.

12

CHAPTER 4

SYSTEM DESIGN
4.1 ARCHITECTURE DIAGRAM
Keyword
Associated with images
DiscoveryOf

Query specific

Classifier

Reference

Reference
classes

Of

classes

signature over
reference classes

Reference classes

Offline
Online
Query Text
Text based image list

Re-ranking based on semantic


signature

FIGURE 4.1SYSTEM ARCHITECTURE DIAGRAM

This diagram, Given a query key- word input by a user, according to a stored word-image
index le, a pool of images relevant to the query keyword are retrieved by the search engine.
By asking a user to select a query image, which reects the users search intention, from the
pool, the remaining images in the pool are re-ranked based on their visual similarities with
the query image. The visual features of images are pre-computed ofine and stored by the
search engine. The main online computational cost of image re-ranking is on comparing
visual features. Instead of constructing a universal concept dictionary, it learns different
visual semantic spaces for different query keywords individually and automatically. We
believe that the semantic space related to the images to be re-ranked can be signicantly
narrowed down by the query keyword provided by the user. For example, if the query
keyword is apple, the semantic concepts of mountains and Paris are unlikely to be
13

relevant and can be ignored. Instead, the semantic concepts of computers and fruit will
be used to learn the visual semantic space related to apple. The query-specic visual
semantic spaces can more accurately model the images to be re-ranked, since they have
removed other potentially unlimited number of non-relevant concepts, which serve only as
noise and deteriorate the performance ofre-ranking in terms of both accuracy and
computational cost. The visual features of images are then projected into their related visual
semantic spaces to get semantic signatures. At the online stage, images are re-ranked by
comparing their semantic signatures obtained from the visual semantic space of the query
keyword.
Several subsystems can be identified within the architecture of retrieval systems for images:
Pre-processing of the image to reduce the influence of different acquisition circumstances
(e.g., differences in illumination). Extraction of low-level features of visual data (e.g., Shape
and color).Image processing and pattern recognition is used to measure such features.
Extraction of high-level feature and image semantics (e.g., recognition of a shoe brand and
type based on the shoe profile).In some cases, semantics can be extracted automatically from
the images based on a combination of low-level features and rules. Description in textual
form of the image contents and acquisition characteristics (type of camera, image size, image
resolution, number of images in a sequence, information available about the individual
correspondence with the images, metadata describing the content etc.)Visualization, which
presents a view of the data for inspection, herby improving the effectiveness of the search.
Indexing and pre-selection, this filters out images that are not pertinent to a query and
extracts only those database items that are relevant to the query.

Retrieval: matching

procedures and similarity metrics. Relevancy feedback: a mechanism by which the user can
give feedback by indicating positive or negative relevance of retrieved items.
Functionality: Initially, user interface typically consists of a query formulation part and a
result presentation part. Specification of which images to retrieve from the database can be
done in many ways. One way is to browse through the database one by one. Another way is
to specify the image in terms of keywords, or in terms of image features that are extracted
from the image, such as a color histogram. Yet another way is to provide an image or sketch
from which features of the same type must be extracted as for the database images, in order to
match these features A combination of four feature extraction methods namely color
Histogram, Color moment texture, and edge histogram descriptor. There is a provision to add
new features in future for better retrieval efficiency. Any combination of these methods,
14

which is more appropriate for the application, can be used for retrieval. This is provided
through User Interface (UI) in the form of relevance feedback. Indexing is often used as
identifying features within an image; with indexing data structures we here mean structures to
speed up the retrieval of features within image collections.
4.2 DATAFLOW DIAGRAM

Start

Enter the
query

Query analysing by
image

Query grouping by
images

Save the queries in


database

query relevance by
image

Flittering the
images

Display the flittering


images

Graph

FIGURE 4.2 SYSTEM DATA FLOW DIAGRAM

15

In this diagram, For each query keyword, a multi-class classier on low- level visual features
is trained from the training sets of its reference classes and stored ofine. If there are K types
of visual features, one could combine them to train a single classier. It is also possible to
train a separate classier for each type of features. Our experiments show that the latter
choice can increase the re-ranking accuracy but will also increase storage and reduce the
online matching efciency because of the increased size of semantic signatures. An image
may be relevant to multiple query keywords. Therefore it could have several semantic
signatures obtained in different semantic spaces. According to the word- image index le,
each image in the database is associated with a few relevant keywords. For each relevant
keyword, a semantic signature of the image is extracted by computing the visual similarities
between the image and the reference classes of the keyword using the classiers trained in the
previous step. The reference classes form the basis of the semantic space of the keyword. If
an image has N relevant keywords, then it has N semantic signatures to be computed and
stored ofine. At the online stage, a pool of images are retrieved by the search engine
according to the query keyword input by a user. Since all the images in the pool are relevant
to the query keyword, they all have pre-computed semantic signatures in the semantic space
of the query keyword. Once the user chooses a query image, all the images are re-ranked by
comparing similarities of the semantic signatures.
Indexing and Feature Extraction: Indexing deals with the insertion of feature vectors into the
database and is a fundamental task in every query-by-content database system. In many
applications, due to the diverse nature of queries, the feature vector may need to be
constructed from multiple, mostly unrelated, features. However, when diverse features are
present in a feature vector; it is very important how the feature space is organized. For
example, if two vectors are compared, which parts of the feature vector should be matched,
all or partial? Some feature may be of higher importance than the others. What is the
computational complexity of this matching? When all the features in a feature vector are
assumed to be equally important, the problem reduces to computing Euclidean distance in a
multidimensional space. In some other cases, different weights may be assigned to each of
the features in the feature vector, and the weights and the similarity metric are determined
through simulation based on certain optimization criteria.

16

In parallel to an efficient indexing system, a query system with the fast query response time is
very crucial in content- based image retrieval systems. the complexity of the computations
involved in query processing must be reduced as much as possible .since in content based
image indexing and retrieval systems images are indexed once(preprocessing stage) and
queried many times over and over, we may want to allocate complex operations to the
preprocessing stage.

17

CHAPTER 5

DEVELOPMENT REQUIREMENT
5.1 HARDWARE REQUIREMENT
Processor

Pentium III

Speed

1.1 Ghz

RAM

256 MB (min)

Hard Disk

20 GB

Floppy Drive

1.44 MB

Key Board

Standard Windows Keyboard

Mouse

Two or Three Button Mouse

Monitor

SVGA

5.2 SOFTWARE REQUIREMENT:


Operating System

Windows95/98/2000/XP

Application Server

Glassfish

Front End

HTML, Java, Jsp

Scripts

JavaScript.

Server side Script

Java Server Pages.

Database

MySQL

Database Connectivity -

JDBC.

5.3 LANGUAGE DESCRIPTION


5.3.1 JAVA
Java is a programming language created by James Gosling from Sun Microsystems (Sun) in
1991. The first publicly available version of Java (Java 1.0) was released in 1995. Sun
Microsystems was acquired by the Oracle Corporation in 2010. Oracle has now the
steermanship for Java. Over time new enhanced versions of Java have been released. The
18

current version of Java is Java 1.7 which is also known as Java 7. From the Java
programming language the Java platform evolved. The Java platform allows software
developers to write program code in other languages than the Java programming language
and still runs on the Java virtual machine. The Java platform is usually associated with
the Java virtual machine and the Java core libraries.Java and Open Source: In 2006 Sun
started to make Java available under the GNU General Public License (GPL). Oracle
continues this project called OpenJDK. Java Virtual machine: The Java virtual machine
(JVM) is a software implementation of a computer that executes programs like a real
machine.
JSP technology is used to create web application just like Servlet technology. It can be
thought of as an extension to servlet because it provides more functionality than servlet such
as expression language, jstl etc. A JSP page consists of HTML tags and JSP tags. The jsp
pages are easier to maintain than servlet because we can separate designing and development.
It provides some additional features such as Expression Language, Custom Tag etc.
5.3.2 MySQL
MySQL is the most popular Open Source Relational SQL database management system.
MySQL is one of the best RDBMS being used for developing web-based software
applications. What is Database?A database is a separate application that stores a collection of
data. Each database has one or more distinct APIs for creating, accessing, managing,
searching and replicating the data it holds. Other kinds of data stores can be used, such as
files on the file system or large hash tables in memory but data fetching and writing would
not be so fast and easy with those types of systems. So nowadays, we use relational database
management systems (RDBMS) to store and manage huge volume of data. This is called
relational database because all the data is stored into different tables and relations are
established using primary keys or other keys known as foreign keys.
MySQL is a fast, easy-to-use RDBMS being used for many small and big businesses.
MySQL is developed, marketed, and supported by MySQL AB, which is a Swedish
company. MySQL is becoming so popular because of many good reasons: MySQL is
released under an open-source license. So you have nothing to pay to use it. MySQL is a very
powerful program in its own right. It handles a large subset of the functionality of the most
expensive and powerful database packages. MySQL uses a standard form of the well-known
SQL data language. MySQL works on many operating systems and with many languages
19

including PHP, PERL, C, C++, JAVA, etc. MySQL works very quickly and works well even
with large data sets. MySQL is very friendly to PHP, the most appreciated language for web
development. MySQL supports large databases, up to 50 million rows or more in a table. The
default file size limit for a table is 4GB, but you can increase this (if your operating system
can handle it) to a theoretical limit of 8 million terabytes (TB). MySQL is customizable. The
open-source GPL license allows programmers to modify the MySQL software to fit their own
specific environments.

5.3.3 NetBeans IDE


NetBeans is an integrated development environment (IDE) for developing primarily
with Java, but also with other languages, in particularPHP, C/C++, and HTML5.[3] It is also
an application platform framework for Java desktop applications and others.The NetBeans
IDE is written in Java and can run on Windows, OS X, Linux, Solaris and other platforms
supporting a compatible JVM.The NetBeans Platform allows applications to be developed
from a set of modular software components called modules. Applications based on the
NetBeans Platform (including the NetBeans IDE itself) can be extended by third party
developers.The NetBeans Team actively support the product and seek future suggestions
from the wider community. Every release is preceded by a time for Community testing
NetBeans began in 1996 as Xelfi (word play on Delphi), a Java IDE student project under the
guidance of the Faculty of Mathematics and Physics at Charles University in Prague. In 1997
Roman Stank formed a company around the project and produced commercial versions of
the NetBeans IDE until it was bought bySun Microsystems in 1999. Sun open-sourced the
NetBeans IDE in June of the following year. Since then, the NetBeans community has
continued to grow.[9] In 2010, Sun (and thus NetBeans) was acquired by Oracle.
Framework for simplifying the development of Java Swing desktop applications. The
NetBeans IDE bundle for Java SE contains what is needed to start developing NetBeans
plugins and NetBeans Platform based applications; no additional SDK is required.
Applications can install modules dynamically. Any application can include the Update Center
module to allow users of the application to download digitally signedupgrades and new
features directly into the running application. Reinstalling an upgrade or a new release does
not force users to download the entire application again.

20

The platform offers reusable services common to desktop applications, allowing developers
to focus on the logic specific to their application. Among the features of the platform are:User
interface management (e.g. menus and toolbars)User settings managementStorage
management (saving and loading any kind of data)Window managementWizard framework
(supports step-by-step dialogs)NetBeans Visual LibraryIntegrated development tools
NetBeans IDE is a free, open-source, cross-platform IDE with built-in-support for Java
Programming Language.

NetBeans IDE is an open-source integrated development environment. NetBeans IDE


supports development of all Java application types (Java SE(including JavaFX), Java
ME, web, EJB and mobile applications) out of the box. Among other features are an Antbasedprojectsystem, Maven support,refactorings, versioncontrol (supporting CVS, Subversio
n, Git, Mercurial and Clearcase).All the functions of the IDE are provided by modules. Each
module provides a well defined function, such as support for the Java language, editing, or
support for the CVS versioning system, and SVN. NetBeans contains all the modules needed
for Java development in a single download, allowing the user to start working immediately.
Modules also allow NetBeans to be extended. New features, such as support for other
programming languages, can be added by installing additional modules. For instance, Sun
Studio, Sun Java Studio Enterprise, and Sun Java Studio Creator from Sun Microsystems are
all based on the NetBeans IDE.
License: From July 2006 through 2007, NetBeans IDE was licensed under Sun's Common
Development and Distribution License (CDDL), a license based on the Mozilla Public
License (MPL). In October 2007, Sun announced that NetBeans would henceforth be offered
under a dual license of the CDDL and the GPL version 2 licenses, with the GPL linking
exception for GNU Classpat

21

CHAPTER 6

SYSTEM IMPLEMENTATION
6.1 MODULES:
Login module.
Admin module
Text based image search.
Image re-ranking on one click.
Image retrieval

6.2 MODULE DESCRIPTION:


6.2.1 LOGIN MODULE
In this module, Users are having authentication and security to access the detail which
is presented in the ontology system. Before accessing or searching the details user should
have the account in that otherwise they should register first.
6.2.1.1 Registration:
In this module if a user wants to access the data which is stored in a cloud
server,he/she should register their details first.These details are maintained in a Database. if a
User have to register first,then only he/she has to access the data base.
6.2.1.2 Login:
In this module,any of the above mentioned person have to login,they should login by
giving their username and password .

22

6.2.2 ADMIN MODULE


In this module, admin can view the user details and also view the registered
user in the database. Administration modules interpret the model by way of a special
configuration file. which can be altered to extend all the generated components and the
module look and feel. Such modules benefit from the usual module mechanisms described in
previous chapters (layout, routing, custom configuration, auto loading, and so on). You can
also override the generated action or templates, in order to integrate your own features into
the generated administration, but should take care of the most common requirements and
restrict the use of java code only to the very specific.
6.2.3 TEXT BASED IMAGE SEARCH.
Our search engine first searches the pages and then gets the result searching for the
metadata to get the trusted results search engines require searching for pages that maintain
such information at some place. Here propose the intelligent semantic web based search
engine. we use the power of xml meta-tags deployed on the web page to search the queried
information. The xml page will be consisted of built-in and user defined tagsour practical
results showing that proposed approach taking very less time to answer the queries while
providing more accurate information.
6.2.4 IMAGE RE-RANKING ON ONE CLICK.
The user first submits query keywords q. A pool of images is retrieved by text-based
search. Then the user is asked to select a query image from the image pool. The query image
is classified as one of the predefined adaptive weight categories. Images in the pool are reranked based on their visual similarities to the query image and the similarities are computed
using the weight schema specified by the category to combine visual features. In the keyword
expansion step words are extracted from the textual descriptions (such as image file names
and surrounding texts in the html pages) of the top k images most similar to the query image,
and the tf-idf method [45] is used to rank these words. To save computational cost, only the
top m words are reserved as candidates for further processing. However, because the initial
image reranking result is still ambiguous and noisy, the top k images may have a large
diversity of semantic meanings and cannot be used as visual query expansion.

23

6.2.5 IMAGE RETRIEVAL.


Information retrieval by searching information on the web is not a fresh idea but has
different challenges when it is compared to general information retrieval. Different search
engines return different search results due to the variation in indexing and search process.
The results are shown . It still greatly outperforms the approaches of directlycomparing visual
features. This result can be explained from two aspects. the multiple semantic signatures
obtained from different types of visual features separately have the capability to characterize
the visual content of images outside the reference classes.Many negative examples (images
belonging to differentcategories than the query image) are well modeled by the reference
classes and are therefore pushed backward on the ranking list.

24

CHAPTER 7

CONCLUSION AND FUTURE ENHANCEMENT.


Image Retrieval systems emerged as one of the most active research areas in the fast few
years. Most of the early research effort focused on finding the best image feature
representation. Retrieval was performed as summarization of similarities of individual feature
representation of fixed weights. While this computer centric approach establishes the basis of
Image retrieval, the usefulness of such systems was limited due to the difficulty in
representing high level concepts using low level features and human perception subjectivity.
The proposed system summarizes an immense survey, incorporating all relevant technical
details with a focus present ongoing research works in the related areas. It has been
computationally superior and highly scalable in terms of query search time that depends only
on the number of images similar to the query image and is relatively independent of the
database size. This paper can also contribute many innovative research solutions to Computer
Cognition Technology, on looking back we can safely assume that this project been
successful in implementing most of the objectives that it has sought to achieve. This would
not be farfetched to conclude with a mention that, this is a just a beginning and a positive
approach in the Successful practical implementation. We propose a novel image re-ranking
framework, which learns query-specic semantic spaces to signicantly improve the
effectiveness and efciency of online image re-ranking. The visual features of images are
projected into their related visual semantic spaces automatically learned through keyword
expansions at the ofine stage. The extracted semantic signatures can be 70 times shorter
than the original visual feature on average, while achieve 20%35% relative improvement on
re-ranking precisions over state of theart methods.The features of each image are precomputed and stored in database. When test image is input to the system, the pre-processing,
feature extraction process is applied to compare with images in Database. The matching
degree and ranking are computed, the results are then output to screen according to the
ranking of similarity.

25

REFERENCES

[1] E. Bart and S. Ullman. Single-example learning of novel classes using representation by
similarity. In Proc. BMVC, 2005.

[2] Y. Cao, C. Wang, Z. Li, L. Zhang, and L. Zhang.Spatial-bag-of- features.In Proc. CVPR,
2010.

[3] G. Cauwenberghs and T. Poggio. Incremental and decremental sup- port vector machine
learning. In Proc. NIPS, 2001.

[4] J. Cui, F. Wen, and X. Tang.Intentsearch: Interactive on-line image search re-ranking. In
Proc. ACM Multimedia.ACM, 2008.

[5] J. Cui, F. Wen, and X. Tang. Real time google and live image search re-ranking. In Proc.
ACM Multimedia, 2008.

[6] N. Dalal and B. Triggs. Histograms of oriented gradients for human detection.In Proc.
CVPR, 2005.

[7] C. Lampert, H. Nickisch, and S. Harmeling.Learning to detect unseen object classes by


between-class attribute transfer.In Proc. CVPR, 2005.

[8] D. Lowe. Distinctive image features from scale-invariant keypoints. Intl Journal of
Computer Vision, 2004.

[9] B. Luo, X. Wang, and X. Tang. A world wide web based image search engine using text
and image content features. In Proceedings of the SPIE Electronic Imaging, 2003.

[10] J. Philbin, M. Isard, J. Sivic, and A. Zisserman.Descriptor Learning for Efcient


Retrieval.In Proc. ECCV, 2010.

26

[11] N. Rasiwasia, P. J. Moreno, and N. Vasconcelos. Bridging the gap: Query by semantic
example. IEEE Trans. on Multimedia, 2007.

[12] M. Rohrbach, M. Stark, G. Szarvas, I. Gurevych, and B. Schiele. What helps wherevand
why? semantic relatedness for knowledge transfer. In Proc. CVPR, 2010.

[13] Y. Rui, T. S. Huang, M. Ortega, and S. Mehrotra. Relevance feed- back: a power tool for
interactive content-based image retrieval. IEEE Trans. on Circuits and Systems for Video
Technology, 1998.

[14] D. Tao, X. Tang, X. Li, and X. Wu. Asymmetric bagging and randomsubspaceforsupportvectormachines-basedrelevancefeedback
inimageretrieval.IEEETrans.onPatternAnalysisandMachineIn- telligence, 2006.

[15] Q. Yin, X. Tang, and J. Sun. An associate-predict model for face recognition. In Proc.
CVPR, 2011.

[16] X. S. Zhou and T. S. Huang. Relevance feedback in image retrieval: A comprehensive


review. Multimedia Systems, 2003.

27

28

Vous aimerez peut-être aussi