Compressive Sensing: An Introduction and Survey of Applications.

36
Compressive Sensing: An Introduction and Survey of Applications

Transcript of Compressive Sensing: An Introduction and Survey of Applications.

Page 1: Compressive Sensing: An Introduction and Survey of Applications.

Compressive Sensing:An Introduction and Survey of Applications

Page 2: Compressive Sensing: An Introduction and Survey of Applications.

ObjectivesDescription of theoryDiscussion of important resultsStudy of relevant applications

Page 3: Compressive Sensing: An Introduction and Survey of Applications.

Introduction to the ProblemCS is a new paradigm that makes possible

fast acquisition of data using few number of samples

It tries to bring the number of samples acquired as close to the information content as possible

The classical Shannon-Whittaker sampling theorem has monopolized signal acquisition arena

It applies only to band-limited signalsSays number of samples ~ desired

resolution

Page 4: Compressive Sensing: An Introduction and Survey of Applications.

Disadvantages of STBut band-limitedness not a universal

assumptionMost real-life signals have huge frequency

extentAlso it is not always a true measure of

information content (e.g. : train of spikes in frequency)

Also for increased resolution we need to increase the rate of sampling( but speed of most devices is limited)

In some other situations(e.g. : MRI) no of samples that can be acquired is inherently limited

Page 5: Compressive Sensing: An Introduction and Survey of Applications.

Our objectiveWe want to investigate the possibility

of reconstructing a signal from fewer no. of samples than dictated by ST

Signal model assumed: SparsityMuch broader class than BL signalsA nonlinear classSparsity of a signal=no. of non-zero

coefficients = norm of the signal vector

l

0l

Page 6: Compressive Sensing: An Introduction and Survey of Applications.

Sparsity model holds for most known signals

Most naturally-occurring signals are sparse in a certain basis

This property has been used in compression using transform coding:

Bandlimit

and Sample at

Nyquist Rate

H Thresholding

f

~

f

Page 7: Compressive Sensing: An Introduction and Survey of Applications.

Even though samples are acquired at max possible rate we get rid of most of them

This strategy is wasteful in many waysCS combines the first and second stages to

acquire signals in the compressed form

The cost incurred is increased computational requirement

HfSome

algorithm for reconstructio

n

Sparse result

Page 8: Compressive Sensing: An Introduction and Survey of Applications.

Questions:What is the minimum no of

measurements we require?How can reconstruction be done from

the reduced no of samples?The problem can be shown to be

reducible to a problem of solving an underdetermined system of linear equations

Reconstruction is possible because of sparsity assumption

This method is non-adaptive

Page 9: Compressive Sensing: An Introduction and Survey of Applications.

Some Notation sensing basis sparsifying basis This is a set of orthonormal vectors where the signal is

known to have a sparse representation with S coefficients

• Measurements are linear functionals of the form where ∈ s.t. • It can be shown measurements are of the form where x is a sparse vector

}{ i}{ i Ni 1

, fi i T NMT ||

bUx

Page 10: Compressive Sensing: An Introduction and Survey of Applications.

Effectively we are doing a projection of an n-dimensional vector into an M dimensional space

Need to make sure unique recovery is possible

i.e.; if Stability: RIP condition= Sub-matrix is well-

conditioned with a good condition number(necessary and sufficient)

For the sensing matrix and number of measurements these conditions should be satisfied

0)( 21 xxA21 xx

Page 11: Compressive Sensing: An Introduction and Survey of Applications.

ContdRIP condition is related to the uncertainty

principle for the two orthobasesAnother property of interest in incoherenceThe less the coherence the better the resultsIt’s like saying the sensing matrix should be

as different from the sparsifying basis or the measurements should be holographic (non-concentrated).

Should convey the least amt of info abt the signal(20 weights problem)

Possibly random(or complementary bases like time frequency)

Page 12: Compressive Sensing: An Introduction and Survey of Applications.

Method of reconstruction

Various possibilities existWe need to pick the least sparse solutionThe RIP condition ensures it is the unique solutionCould go for a combinatorial optimisation(directly

sift thru all sparse solutions to pick the actual one)

Then only require S+1 samplesBut it is NP hard –highly intractableAnother choice-l2 norm. Very easy to analyse. But

does not give Sparse solutionsSo a compromise is to go for l1 norm minimisation

Page 13: Compressive Sensing: An Introduction and Survey of Applications.

Geometrical ArgumentL1 norm

Page 14: Compressive Sensing: An Introduction and Survey of Applications.

L2 Norm

Page 15: Compressive Sensing: An Introduction and Survey of Applications.

Why RIP is necessary?

Page 16: Compressive Sensing: An Introduction and Survey of Applications.

ResultLet’s consider sets of orthogonal bases 𝝍 and 𝜱 If then with probability exceeding x supported on a fixed set can be recovered using the following optimisation problem: Imp requirements : incoherence ,randomness , RIPThis ensures that the set of measurements for which reconstruction fails occurs with very small prob: to take adv take random samples

)/log()(2 nUSM

1

1||min lx yUx

Page 17: Compressive Sensing: An Introduction and Survey of Applications.

Non-Uniform Sampling Theorem

Signal composed of S discrete frequenciesTake M random measurements in time

domainFrom above theorem M>Slog(n) gives perfect

reconstructionTime and frequency domains are maximally

incoherentThis is also fundamental i.t.s.t. fewer

samples and reconstruction is virtually impossible

Eg Dirac Comb

Page 18: Compressive Sensing: An Introduction and Survey of Applications.

More on RIPUnder the assumptions for which the

above result holds it can be shown that for x belonging to a fixed set T :

• Ensures any such signal is recovered

uniquely recoverable• To include all sets T we need to

strengthen the above condition (UUP). But then the number of samples required increases; becomes 4-5th power of log(n)

22

22

22 ||

2

3||||

2xmUxx

m

Page 19: Compressive Sensing: An Introduction and Survey of Applications.

DefinitionRestricted Isometry Constant : It’s

smallest constant such that for every x belonging to T with size S <1 for condition to holdThis is an approximate orthogonality condition• UUP=should hold for all T’s with the same size

22

22

22 ||)1(||||)1( xAxx SS

S

Page 20: Compressive Sensing: An Introduction and Survey of Applications.

ResultAssuming the existence of a matrix that

satisfies the above property with sufficiently

then

• Here the condition on M depends on the type of matrix we choose

• This is not probabilistic

12 S

11

12

||||

/||||*

*

lsl

lsl

xxCxx

SxxCxx

Page 21: Compressive Sensing: An Introduction and Survey of Applications.

Two concernsSignals are only approximately sparseMeasurements are noisyOur reconstruction procedure should

be robust against these two casesRIP to the rescue!Has been shown if the

solution to sub to

satisfies

122 S

1||min lx

2|| lyAx

./|||| 1*

12CSxxCxx lsl

Page 22: Compressive Sensing: An Introduction and Survey of Applications.

SummaryWe need to find sensing matrices that

satisfy the isometry propertySimple choice: random matricesIf m>CSlog(n/δ) random matrices satisfy RIPFor orthobases we need 4-5th power of log(n)Random matrices present storage difficulty

Page 23: Compressive Sensing: An Introduction and Survey of Applications.

Application: Spectrum SensingSpectrum sensing is the task of detecting the

presence or absence of a carrier in a wideband of freqs

Cognitive Radios should be equipped with such a mechanism to enable efficient utilisation of

channelA major implementation challenge lies in the

very high sampling rates required by conventional spectral estimation methods which have to operate at or above the Nyquist rate.

Because of high rates no of samples is limitedMay not provide sufficient statistic

Page 24: Compressive Sensing: An Introduction and Survey of Applications.

The situation is appropriate for deployment of CS

The spectrum is sparse because only a relatively small no of users are transmitting

Let be the frequency range in useBandwidth=B

No ff

Page 25: Compressive Sensing: An Introduction and Survey of Applications.

Our job is to detect the N frequency bands and classify them as black, grey or white regions based on the PSD level

In the analysis we use a vector of time samples sampled at Nyquist rate of To. In the actual implementation only sub-Nyquist sampling is done

So let where is a vector of M values in the duration [0 MTo]. And

This is a generic model. S can be any matrix of basis vectors

Since sparsity under consideration is in freq, time domain is the best sampling domain. then S =

tT rSx tr

KRx

MI

Page 26: Compressive Sensing: An Introduction and Survey of Applications.

Steps involvedReconstruct from M measurementsFind high resolution fourier transform

Obtain the frequency edges from Estimate the PSD in each bandLater we’ll see it is not necessary to

reconstruct the frequency vector from Only course sensing is done so noise is not

a key factorTo reduce noise effects a wavelet

smoothing operation is done

tr

tMf rFr fr

tr

Page 27: Compressive Sensing: An Introduction and Survey of Applications.

Edge DetectionProblem similar to edge detection in imagesLet be a wavelet smoothing function with

a compact support. The dilation of by a scale factor s is given by:

For dyadic scales s is a power of twoContinuous wavelet transform of R(f) is given

by:

• Then we do a simple differencing to this function

)( f)( f

Page 28: Compressive Sensing: An Introduction and Survey of Applications.

From above:

Replacing by its estimate found below:

We get the estimated wavelet transform:

Then we take the derivative of the above vector and find local minima

tr

Page 29: Compressive Sensing: An Introduction and Survey of Applications.

Below is the vector with derivative values:

Take local peaks and nous sommes doneAlso in each band we can estimate the

average PSDby just averaging the frequency vector in that

band

One major simplification can be done by noting zs is itself a sparse vector. Thus we can eliminate the need to reconstruct

fr

Page 30: Compressive Sensing: An Introduction and Survey of Applications.

To do this we define:

which is the differentiation matrix• Then we rewrite the above equation:

And finally:

Page 31: Compressive Sensing: An Introduction and Survey of Applications.

Recovered frequency response

Page 32: Compressive Sensing: An Introduction and Survey of Applications.

Channel EstimationCS also has applications in sparse channel

estimation Every delay-dominant channel can be

represented as a superposition of the pilot signal sent

It can be shown we can divide the total delay into bins each 1/W in width

Not all these bins are always occupiedEach bin represents a dominant scattererNo of bins that are non-zero<<p(total

bins)=floor(Tm*W)

Page 33: Compressive Sensing: An Introduction and Survey of Applications.

Two ways to make use of CSUse proper pilot signalsCan be either random bits 1 or -1Or could be a sum of exponentials with

random frequencies.The first corresponds to using a random

sensing basisThe second is random sampling in frequency

using a subset of FT which is maximally incoherent

In both cases we obtain a better MSE less than LS

Page 34: Compressive Sensing: An Introduction and Survey of Applications.

Another app is channel coding

Page 35: Compressive Sensing: An Introduction and Survey of Applications.

References Channel Estimation:

.Bajwa, U.W., Haupt, J., Raz, G. and Nowak, R. (2008). “Compressed Channel Sensing”, Proceedings of the Annual Conference on Information Sciences and Systems, March 2008, Pages 1-6

• Spectrum Sensing: Z. Tian and G. B. Giannakis, “Compressed sensing for wideband

cognitive radios,” in Acoustics, Speech and Signal Processing,2007. ICASSP 2007. IEEE International Conference on, vol.4,Honolulu, HI, Apr. 2007, pp. 1357–1360.

CS E. Candès and J. Romberg, “Sparsity and incoherence in

compressive sampling,” Inverse Prob., vol. 23, no. 3, pp. 969–985, 2007.

An Introduction To Compressive Sampling: Emmanuel Candes Michael B Wakin

A Lecture on CS Richard Baraniuk

Page 36: Compressive Sensing: An Introduction and Survey of Applications.