Regression Models for Texture Image Analysis

Abstract

The article describes universal model for creating algorithms for calculating textural image features. The proposed models are used for images that are realizations of Markov Random Field. Experimental classification results are shown for different images sets.

Keywords: Textutre Image Analysis, Textutre Image Recognition, Regression models, Markov Random Fields.

Introduction

Many natural micro and macro images are combinations of textural primitives, e.g. aerial photographs of the earth surface, images of material nanostructures, biomedical images. That’s why in many applied problems of the image processing (like, segmentation, coding, simulation, etc.) texture processing methods can be used.

Two main approaches to the texture analysis can be noted: structural and stochastic [1]. In structural approach textures are considered as a combination of repeating primitive elements, which relative position is defined by specific rules. In stochastic approach, the texture is defined as a realization of a 2D random process, and specified by a set of random statistics. In our research work we used statistical approach.

It should be noted, that common texture analysis methods, like co-occurrence matrix [2], spectral descriptors, Tamura features [3], etc., give common features for all types of images and ignore problem peculiarities. However, in practice it becomes necessary to calculate features for classification of limited texture sets. Also many methods deal with gray-scale images only i.e. these methods can’t be applied to analysis of some classes of texture images. Methods based on Markov random field are successfully used in many applied problems. This methods proved to be efficient in image segmentation problems, image synthesis, etc. [4,5,6,7]. Thus, the goal of our investigation is to provide general framework for estimation of color-texture features based on domain specific information.


Problem of Texture Image Recognition

Let I is a set of all possible images, that consists of K classestmp1411-1637_thumb[2][2]

tmp1411-1638_thumb[2][2]of objects that are not included in any class. Original class separation is unknown.

Suppose we are given a finite set of imagestmp1411-1641_thumb[2][2]the class label is specified for each image from this set, i.e. following map is defined u : tmp1411-1642_thumb[2][2]wheretmp1411-1643_thumb[2][2]is a set of class labels. A training set is defined as a set of image-class pairs:tmp1411-1644_thumb[2][2]

The goal is to design a method that allows to identify class label based on image I. To use pattern recognition theory [7,8], one should associate features vectortmp1411-1645_thumb[2][2]with each image, thus following map should be defined:

tmp1411-1646_thumb[2][2]where F is a feature space. Hence, here exists an unknown dependence tmp1411-1647_thumb[2][2]values are observed on finite training settmp1411-1648_thumb[2][2]. One should define a maptmp1411-1649_thumb[2][2]that is an approximation of unknown target dependency on both elements of training set and whole set F.

All existing feature calculation methods are defined for a wide range of problems, i.e. it doesn’t fully utilize domain specific information from training set.

To exploit domain specific information we should define parametric feature calculation method Fa, where a is a set of method parameters. Thus, specific method parameters a = P (I) are calculated based on training set I at a training step, and feature vector is estimated for each specific image using calculated parameters vector f = Fa(I). The general classification scheme is shown in figure 1.

Classification scheme

Fig. 1. Classification scheme

On first step the method’s vector parameters a = P(I) are calculated based on training set I. Then resulted parameters used for estimation of features are used on next classifier training step.

Image Model

To design features estimation algorithm in case of texture images analysis, one can use Markov Random Field (MRF) model [7], it’s effectively used in many texture analysis problems.

Let’s define image as map:

tmp1411-1660_thumb[2][2]

wheretmp1411-1661_thumb[2][2]are spatial coordinates, and G is  a set of colors; e.g. for gray-scale image G = [0; 1], and for colored image (RGB color space)tmp1411-1662_thumb[2][2]

Let the image be a realization of any Markov random field [7]. Following notation will be used below:

tmp1411-1663_thumb[2][2]neighborhood of a pixel p; tmp1411-1664_thumb[2][2]neighborhood element;

tmp1411-1665_thumb[2][2]of pixel values in the neighborhood of current pixel.

Features Design Method

Since the value of central pixel Ip depends on values of pixels in the neighborhood, we can consider approximation function — the function where the pixels from the neighborhood are arguments and the value of function is an estimate of central pixel value. This function can be defined as:

tmp1411-1671_thumb[2][2]

or using regression equation:

tmp1411-1672_thumb[2][2]

Let’s consider solving regression problem.tmp1411-1673_thumb[2][2]

tmp1411-1674_thumb[2][2]of regression problem. We use kernel ridge regression method [9]. According to representer theorem [8] we’re going to find solution in form:

tmp1411-1677_thumb[2][2]

where k is positive defined kernel [8].

The cost function, which needs to be minimized for ridge regression is:

tmp1411-1678_thumb[2][2]

The ridge model coefficients are given by:tmp1411-1679_thumb[2][2]wheretmp1411-1680_thumb[2][2]

tmp1411-1681_thumb[2][2]We use Gaussian kernel in our study:

tmp1411-1685_thumb[2][2]

Hence, each image is associated with its own set of parameters a. Overall, the parameters are similar for images from the same class.

To build features separating several classes, we should use regression function based on all images from training set. But it leads to the next issue — increasing of regression problem training set size. To avoid increasing of the set size, only a subset of training set for image can be used, it’s possible because a lot of vectors in the training set are similar. Hence, only linear independent (in RKHS for kernel k) vectors can be used from the training set.

Selecting of Linearly Independent Subset

Consider we need to select M linear independent vectors from settmp1411-1686_thumb[2][2]Let tmp1411-1687_thumb[2][2]be a fixed finite dimensional mapping from domain space to Reproducing Kernel Hilbert Space with kernel k, such thattmp1411-1688_thumb[2][2]

Let’s use following iterative procedure [10]. On each step t, subset of vectors tmp1411-1689_thumb[2][2]have been collected, after t — 1 vectors were observed, where tmp1411-1690_thumb[2][2]are approximately linearly independent vectors by construction. We test iftmp1411-1691_thumb[2][2]is approximately linearly dependent on subset Dt-1, if not we add it to Dt-1, i.e. Dt = Dt-1 U xt. Hence, all vectors x1,.. .,xt can be approximated as linear combination of vectors Dt. xt is approximately linearly dependent on set Dt-1, it means that it satisfies following condition:

tmp1411-1698_thumb[2][2]

where S is accuracy parameter.

To sum it all up, the set of method parameters a for each particular problem is a set of linear independent neighborhood vectors, generated from all training set images. Thus, feature vector is regression coefficients of approximation function.

Experiments with Test Images

Classification experiments on 2 sets of test images were performed, each set of images consists of 2 classes (see Fig. 2 and 3). N indicates the number of linearly independent vectors used, and S indicates the radius of non-causal neigborhood (e.g. S1; N5 means that 3 x 3 non-causal neigborhood used and number of linearly independent vectors was 5).

Non-linear SVM classifier [8,11] is used for experiments. Each class of images is split into training set of 10 images and validation set of 100 images. Training set is used to estimate parameters of feature calculation algorithm, that are used to calculate features for images in test set. Then these features are used for training of the SVM classifier. Accuracy of classification was investigated for images distorted by additive and impulse noise, using the decision function obtained on previous step. The results are shown in Fig. 4 and 5.

Sample images from 1st set

Fig. 2. Sample images from 1st set

Sample images from 2nd set

Fig. 3. Sample images from 2nd set

Effect of white and impulse noise on 1st image set: a — white noise, b — impulse noise

Fig. 4. Effect of white and impulse noise on 1st image set: a — white noise, b — impulse noise

Effect of white and impulse noise on 2nd image set: a — white noise, b — impulse noise

Fig. 5. Effect of white and impulse noise on 2nd image set: a — white noise, b — impulse noise

Conclusion

In our research work the general framework for estimating textural features, which utilizes domain specific information from training set, is described. Proposed method is used for formalization of features estimation algorithm based on approximation of pixels dependency function, for images with markov random field model.

The method has following advantages: domain specific features increase classification quality (in terms of correct classification probability); features are stable to additive and impulse noise.

Next post:

Previous post: