NEURAL NETWORK FOR THE FAST GAUSSIAN DISTRIBUTION TEST

Similar documents
Fisher Information Determines Capacity of ε-secure Steganography

ECE 3800 Probabilistic Methods of Signal and System Analysis

ECE 3800 Probabilistic Methods of Signal and System Analysis

ECE 3800 Probabilistic Methods of Signal and System Analysis

NEW PIXEL SORTING METHOD FOR PALETTE BASED STEGANOGRAPHY AND COLOR MODEL SELECTION

Outline. Random Variables. Examples. Random Variable

2. Definition & Classification

Information Hiding and Covert Communication

A Hybrid Method with Lorenz attractor based Cryptography and LSB Steganography

EM-algorithm for Training of State-space Models with Application to Time Series Prediction

Improved Adaptive LSB Steganography based on Chaos and Genetic Algorithm

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

ECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture)

Complete characterization of perfectly secure stego-systems with mutually independent embedding operation

Neural Networks and the Back-propagation Algorithm

Detection of Content Adaptive LSB Matching (a Game Theory Approach)

Diversity Combining Techniques

Engineering Physics 3W4 "Acquisition and Analysis of Experimental Information" Part II: Fourier Transforms

ROC Curves for Steganalysts

STATISTICAL DETECTION OF INFORMATION HIDING BASED ON ADJACENT PIXELS DIFFERENCE

Known probability distributions

Heterogeneous mixture-of-experts for fusion of locally valid knowledge-based submodels

Old painting digital color restoration

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

Secure Communication Using H Chaotic Synchronization and International Data Encryption Algorithm

A robust and efficient scheme for image encryption-decryption

Deccan Education Society s FERGUSSON COLLEGE, PUNE (AUTONOMOUS) SYLLABUS UNDER AUTONOMY. FIRST YEAR B.Sc.(Computer Science) SEMESTER I

New Steganographic scheme based of Reed- Solomon codes

Imperfect Stegosystems

QISLSQb: A Quantum Image Steganography Scheme Based on Least Significant Qubit

Research Article A Steganographic Method Based on Pixel-Value Differencing and the Perfect Square Number

Steganalysis with a Computational Immune System

Appendix A Wirtinger Calculus

Fundamental Limits of Invisible Flow Fingerprinting

Detection of LSB Steganography via Sample Pair Analysis

Research Article Improved Adaptive LSB Steganography Based on Chaos and Genetic Algorithm

Steganalysis of Spread Spectrum Data Hiding Exploiting Cover Memory

Modeling and Model Predictive Control of Nonlinear Hydraulic System

Quantization Index Modulation using the E 8 lattice

MATHEMATICAL TOOLS FOR DIGITAL TRANSMISSION ANALYSIS

Matrix Embedding with Pseudorandom Coefficient Selection and Error Correction for Robust and Secure Steganography

Calculating the Required Number of Bits in the Function of Confidence Level and Error Probability Estimation

Informatics 2B: Learning and Data Lecture 10 Discriminant functions 2. Minimal misclassifications. Decision Boundaries

SIMON FRASER UNIVERSITY School of Engineering Science

Lecture 6: Gaussian Mixture Models (GMM)

Image Compression using DPCM with LMS Algorithm

Stego key searching for LSB steganography on JPEG decompressed image

Distributed Clustering and Local Regression for Knowledge Discovery in Multiple Spatial Databases

Artificial Neural Network

Week 1 Quantitative Analysis of Financial Markets Distributions A

Blind Equalization Formulated as a Self-organized Learning Process

Random Variables. P(x) = P[X(e)] = P(e). (1)

Application of Associative Matrices to Recognize DNA Sequences in Bioinformatics

NAME... Soc. Sec. #... Remote Location... (if on campus write campus) FINAL EXAM EE568 KUMAR. Sp ' 00

Upper Bounds for the Average Error Probability of a Time-Hopping Wideband System

Other Topics in Quantum Information

Modeling of Hydraulic Control Valves

ECE Homework Set 2

Computer Vision & Digital Image Processing

Reliability and Quality Mathematics

Using Neural Networks for Identification and Control of Systems

Entropy Evaluation for Oscillator-based True Random Number Generators

Rotation, scale and translation invariant digital image watermarking. O'RUANAIDH, Joséph John, PUN, Thierry. Abstract

Coding into a source: an inverse rate-distortion theorem

P 1.5 X 4.5 / X 2 and (iii) The smallest value of n for

Breaking an encryption scheme based on chaotic Baker map

Revision of Lecture 5

Approximate Inference Part 1 of 2

The Method of Obtaining Best Unary Polynomial for the Chaotic Sequence of Image Encryption

Approximate Inference Part 1 of 2

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING

Intensity Transformations and Spatial Filtering: WHICH ONE LOOKS BETTER? Intensity Transformations and Spatial Filtering: WHICH ONE LOOKS BETTER?

2 2 + x =

Determining the Optimal Decision Delay Parameter for a Linear Equalizer

Lecture 7 Predictive Coding & Quantization

While using the input and output data fu(t)g and fy(t)g, by the methods in system identification, we can get a black-box model like (In the case where

NOISE can sometimes improve nonlinear signal processing

Towards Provably Invisible Network Flow Fingerprints

An Invariance Property of the Generalized Likelihood Ratio Test

Convolution and Pooling as an Infinitely Strong Prior

A Few Notes on Fisher Information (WIP)

Cryptography and Security Final Exam

An Introduction to Independent Components Analysis (ICA)

Performance Analysis of Spread Spectrum CDMA systems

Learning outcomes. Palettes and GIF. The colour palette. Using the colour palette The GIF file. CSM25 Secure Information Hiding

SPATIAL-TEMPORAL TECHNIQUES FOR PREDICTION AND COMPRESSION OF SOIL FERTILITY DATA

HST.582J/6.555J/16.456J

Unit 8: Introduction to neural networks. Perceptrons

Witsenhausen s counterexample and its links with multimedia security problems

University of Siena. Multimedia Security. Watermark extraction. Mauro Barni University of Siena. M. Barni, University of Siena

New polynomials for strong algebraic manipulation detection codes 1

ONE DIMENSIONAL TRIANGULAR FIN EXPERIMENT. Current Technical Advisor: Dr. Kelly O. Homan Original Technical Advisor: Dr. D.C. Look

STEGOSYSTEMS BASED ON NOISY CHANNELS

Book announcements. Sukumar Das Adhikari ASPECTS OF COMBINATORICS AND COMBINATORIAL NUMBER THEORY CRC Press, Florida, 2002, 156pp.

Dirty Paper Writing and Watermarking Applications

Spread Spectrum Watermarking. Information Technologies for IPR Protection

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Bilingual Information Hiding System: A Formalized Approach

Quantum Wireless Sensor Networks

A New Hybrid System for Recognition of Handwritten-Script

Transcription:

IGOR BELI^, ALEKSANDER PUR NEURAL NETWORK FOR THE FAST GAUSSIAN DISTRIBUTION TEST There are several problems where it is very important to know whether the tested data are distributed according to the Gaussian law. At the detection of the hidden information within the digitized pictures (steganography), one of the key factors is the analysis of the noise contained in the picture. The incorporated noise should show the typically Gaussian distribution. The departure from the Gaussian distribution might be the first hint that the picture has been changed possibly new information has been inserted. In such cases the fast Gaussian distribution test is a very valuable tool. The article describes the phase of the noise (in the picture) extraction and the distribution formation. The second phase of the noise analysis is performed by the neural network. The neural network is trained to recognize the Gaussian distribution of the noise. The trained neural network successfully performs the fast Gaussian distribution test. INTRODUCTION The cryptographic method of steganography secures an important message by encrypting it to an unrecognized form of the data. Steganographic methods hide the encrypted message in cover carriers (pictures) so that it cannot be seen while it is transmitted on public communication channels such as computer network. Many steganographic methods embed a large amount of secret information in the first k LSBs (Lest Signifficant Bit) of the pixels of the cover images (Lou, Liu, 00). By other words, the construction of the LSB insertion, the noise of the carrier picture is changed. Normally the noise of the original picture is the Gaussian noise. The change in the "noise" content also changes the shape of the noise distribution function, which is no longer normally distributed. Discovery that the image noise is not normally distributed can be the first hint that the original image has been changed. The steganography is the cryptographic method that is very simple to implement and on the other hand, it is extremely hard to detect and even harder to decipher. One of the possible indices that the original image has been changed, is the noise analysis. Even if the image appers to have the Gaussian noise, this is not the proof that the image is "clen". The sophisticated methods of steganography also reshape the embeded information in such a way that the noise of the changed image again takes the Gaussian distribution profile (Lou, Liu, 00). In order to detect the distribution of the image noise, the special neural network method has been developed and tested. Two goals have been followed. The first is the robustness of the test and the other is the simplicity of its use. Since the determination of the image noise is not the 100% proof of the existance of the embeded information, there is no neccessity to build a very reliable distribution test. The coarse test is often good enough. Sometimes it would be enogh to implement the least square fitting to the sampled image noise distribution, with the function of Gaussian type and observe the least square error of the sampled points from the approximated function. Such test is good enough when the number of the sampled points is high enough. 1

METHOD The method of the Gaussian distribution detection by the neural networks uses two approximation techniques to detect the shape of the noise distribution. First the image noise is amplitude frequency analysed. The analysis results in the amplitude frequency distribution. The image noise distribution is typically discrete distribution, which produces the relatively small value of amplitude samples (for LSB there are only 9 amplitude stages). Therefore the noise amplitude distribution is far from the continuous i.e. normal distribution. This is a typical example of the discrete distribution Poisson distribution (Cherkassky, Mulier, 1998). THE DISTRIBUTIONS Let's first review the relations between the two distributions one discrete and another continuous (Spiegel, 1975). Both distributions and their relations are stated here as the point of reference. THE NORMAL DISTRIBUTION One of the most important examples of a continuous probability density distribution is the normal or Gaussian distribution. The density function for this distribution is given by ( ) / f(x) = 1 e x < x < 1 Where and are the mean and standard deviation respectively. The corresponding distribution function is given by x ( ) / F(x) = P(x x) = 1 e dv In such case we say that the random variable X is normally distributed with mean and variance. If we let Z be the standardized variable corresponding to X, i.e. if we let Z = X then the mean or expected value of Z is 0 and the variance is 1. In such case the density function for Z can be obtained from 1 by formally placing = 0 and, yielding f(z) = 1 e z / This is often referred to as the standard normal density function or distribution. The corresponding distribution function is given by F(z) = P(Z z) = 1 z e u / du = 1 + 1 z e u / du 5 We sometimes call the value z of the standardized variable Z the standard score. The function F(z) is related to the extensively tabulated error function, depicted by erf(z). We have erf(z) = z e u du and F(z) = 1 1 erf 6 0

THE POISSON DISTRIBUTION Let X be a descrete random variable which can take on the values 0, 1,,... such that the probability function of X is given by f(x) = P(X = x) = x e x = 0, 1,,... 7 x where is a given positive constant. This distribution is called the Poisson distribution and a random variable having this distribution is called to be Poisson distributed. RELATION BETWEEN THE POISSON AND NORMAL DISTRIBUTION It can be shown that if X is the Poisson random variable of (7) and (X is the corresponding standardized random variable, then X lim P a b = 1 b e u / du 8 a i.e. the Poisson distribution approaches the normal distribution as or (X )/ is asymptotically normal. SOME PROPERTIES OF THE NORMAL AND THE POISSON DISTRIBUTIONS In the following table some important properties of the general normal distribution and the Poisson distribution are listed. Table1: important properties of the normal and Poisson distributions (Spiegel, 1975) Normal distribution Poisson distribution Mean Variance Standard deviation = Coefficient of skewness = 0 = 1/ Coefficient of kurtosis = = + (1/ ) Moment generating function M(t) = e t ) ( M(t) = e e t 1) i ( / ) Characteristic function ( ) = e ( ) = e e i ( 1) THE NEURAL NETWORK AND LEAST SQUARE APPROXIMATORS In order to detect whether the given noise is at least asymptotically normally distributed we perform two approximations of the amplitude-frequency noise distribution. Figure 1 shows the general idea of two approximations comparrison. Aproximations are needed to enlarge the resolution of sampled distribution and thus to perform a better analysis.

NEURAL NETWORK Figure 1: The noise shape detection by two approximations one model-les and another with the pre-given model EXPERIMENTAL RESULTS The experimetal work was dedicated to design the adequate feedforward neural network (Haykin, 1999) to perform the approximation of the sampled noise distribution function. For the preliminary testing the Cheshire Neuraliyst 1. was used. The least square approximation of the sampled noise data was obtained by the MathWorks Matlab 5.. Figures and show the result of two noise data samples. On the figure there are two approximations of the image noise data. The dot marked curve is the result of the neural networ k approximation while the other curve represents the least square approximation of the same noise data distribution. On the figure, the image noise is distorted. The result of the least square approximation always produces the

Gaussian function, while the neural natwork approximation fits the given data without the limitation of the curve shape this is why it is called model-les aproximation. From both examples (figure and ) it is obvious that such a test can produce the criteria function whether the tested noise distribution is Gaussian or not. Here it is the matter of a distance assesment of the two approximations and setting of the criteria how much should both approximated functions depart from one another that we can still say that the analysed noise profile is still of a Gaussian shape.,5,5 Frequency,5 LS NN 1,5 1 0,5 0 - -1,5-1 -0,5 0 0,5 1 1,5 Amplitude Figure : The result of two approximations. The dot marked curve is the neural network approximation (NN), the curve without the markings is the least square approximation to the normal distribution function. While the sampled noise distribution curve was Gaussian, both approximations were quite similar,5,5,5 LS NN 1,5 1 0,5 0 - -1,5-1 -0,5 0 0,5 1 1,5 Figure : The result of two approximations and with distorted noise distributions. The dot marked curve is the neural network approximation (NN), the curve without the 5

markings is the least square approximation to the normal distribution function. While the sampled noise distribution curve was not Gaussian, both approximations are significantly different. DISCUSSION The shape of the noise that is present in every digital picture is assumed to be normally (Gaussian) distributed. The steganographic methods use the noise level of the picture to embed the new information without notable change in the picture quality. However, the shape of the noise amplitude frequency distribution is change. The noise is no longer normally distributed. The new method of noise distribution shape detection was proposed and tested. The new method uses two types of approximation of the sampled noise amplitude distribution. The first method is called "model-les" and uses the feedforward neural network to perform the approximation. The second method is the classical least square approximation that uses the given gaussian function as the template (model) for approximation. At this approximation method, the best fit of the Gaussian function is found to the sampled noise amplitude distribution. The comparison of the two approximations gives the answer whether the noise in the picture is normally distributed. In the case of normal noise distribution, both approximations produce almost the same results. The experimental work proved that the feedforward neural network with two hidden layers, can approximate the sampled noise amplitude frequency distribution with a smooth approximation function. The answer to the question, how much can both approximated distributions differ from one another that it is still possible to confirm the gaussian nature of the noise, remains the subject of the further research. ABOUT THE AUTHORS Igor Beli~ holds the BS degree from Computer Engineering and works as the senior lecturer at Faculty of Criminal Justice, University of Maribor. Aleksander Pur is working in General police directorate, Informatics and telecommunications service, Ljubljana. His research interests include, DSS and data mining. REFERENCES 1. Cherkassky, V., Mulier, F. (1998). Learning From Data, Concepts, Theory and Methods. New York, John Wiley & Sons.. Haykin, S. (1999). Neural Networks, a Comprehensive Foundation. Upper Saddle River, NJ, Prentice Hall.. Lou, D.C., Liu, J.L. (00). Steganographic Method for Secure Communications, Computers & Security, 1(5), pp 9-60.. Spiegel, M.R. (1975). Theory and Problems of Probability and Statistics, Shaum's Outline Series, New York, McGraw Hill. 6