Date post: | 02-Jan-2016 |
Category: |
Documents |
Upload: | britney-lynch |
View: | 216 times |
Download: | 3 times |
LTSI
(1)Faculty of Mech. & Elec. Engineering, University AL-Baath, Syria
Ahmad Karfoul (1), Julie Coloigner (2,3), Laurent Albera (2,3), Pierre Comon (4,5)
(2)Laboratory LTSI - INSERM U642, France
(3)University of Rennes 1, France
(5)University of Nice Sophia - Antipolis, France
(4)Laboratory I3S - CNRS, France
Semi-nonnegative INDSCAL analysisSemi-nonnegative INDSCAL analysis
OutlinesOutlines
2
• Preliminaries and problem formulation
•Optimization methods• A compact matrix form of derivatives
• Numerical results
• Conclusion
• Global line search
Outer product
Ex. Order 3
Ex. Order q
Outer product of q-vectors rank-one q-th order tensor
Preliminaries and problem formulationPreliminaries and problem formulation
3
4
: Tensor – to – rectangular matrix transformation (unfolding according to the i-th mode)
: Tensor – to – vector transformation
Preliminaries and problem formulationPreliminaries and problem formulation
CANonical Decomposition (CAND) [Hitchcock 1927], [Carroll & Chang 1970], [Harshman 1970]
λPλ1
CAND : Linear combinantion of minimal number of rank -1 terms
Preliminaries and problem formulationPreliminaries and problem formulation
5
INDSCAL decomposition[Carroll & Chang 1970]
λPλ1
Preliminaries and problem formulationPreliminaries and problem formulation
6
CANonical Decomposition (CAND)
λPλ1
INDSCAL decomposition
λ1 λP
INDSCAL = CAND of 3-order tensor symmetric in two of three modes
Preliminaries and problem formulationPreliminaries and problem formulation
7
Example :
(Semi-) nonnegative INDSCAL decomposition for (semi-) nonnegative BSS
Diagonalizing a set of covariance matrices
( ) ( )t tx Z s
[ ] [ ( ) ( ) ] [ ]n n nE t t T Tx sR x x Z R Z
1[ ]xR2[ ]xR
[ ]KxRZ
s : zero-mean random vector of P statistically independent components
Case 1 : Nonnegative INDSCAL decomposition
Case 2 : Semi-nonnegative INDSCAL decomposition
where :
Covariance matrix :
Preliminaries and problem formulationPreliminaries and problem formulation
8
: the (N P) mixing matrix
Problem at hand
Problem 2 : Given , find its INDSCAL decomposition
with
Problem 1 : Given , find its INDSCAL decomposition
subject to
Preliminaries and problem formulationPreliminaries and problem formulation
9
Constrained problem
Unconstrained problem
:Hadamard product (element-wise product)• Parametrizing the nonnegativity constraint: [Chu et al. 04]
• Solution : minimizing the following cost function :
: Khatri-Rao productwith :
Some iterative algorithms
• Steepest Descent
• Newton
• Levenberg Marquardt
First & second order derivatives of ψ
Preliminaries and problem formulation
10
Global line search (1/2)
Update rules :
• Looking for the global optimum in a given direction
Optimization methodsOptimization methods
,A C : learning steps .
11
: Directions given by the iterative algorithm with respect to A and C, respectively.
• 3-th order symmetric (in two modes) tensor Global optimum in the considered direction for
Optimization methodsOptimization methods
• Minimization with respect to
ACand
: Stationary point of a quadratic polynomial
: Stationary point of a 24-th degree polynomialoptA
optC
12
Global line search (2/2)
: Stationary point of a 10-th degree polynomial
• Global optimum in the considered direction for
Steepest Descent (SD)
Update rules :
• Optimization by searching for stationary points of Ψ based on first-order approximation (i.e. the gradient)
Optimization methodsOptimization methods
,A C : learning steps .
13
: Gradient of ψ with respect to A and C, respectively.
In this work• Learning steps are optimal (optimal line search) Global optimum in the
considered direction.
• Gradients are given in a compact matrix form .
14
Steepest Descent (SD)
Optimization methodsOptimization methods
• Computing the differential of ψ are immediat .
Then :
A compact matrix form of
where:
Gradient computation of Ψ(A,C)
Then :
Compact matrix form of derivativesCompact matrix form of derivatives
where IPU : a commutation matrix of size (IP×IP)
N1 : N-dimensional vector of ones
NI : Identity matrix of size (N×N)
15
Update rules :
Newton
• Optimization by including the second-order approximation to accelerate the convergence
Optimization methodsOptimization methods
16
: Hessian of ψ with respect to A and C, respectively.
In this work• Learning steps are also computed optimally (Global line search) .
• Hessians are given in a compact matrix form .
d ( , ) ( ( , ) )dvec( ) ( ( , ) )dvec( )D D D D D A A A C AA C A C A A C CT T T
, ,dvec( ) ( , ) dvec( ( ), ) HH A AA CA C A A C C
d ( , ) ( ( , ) )dvec( ) ( ( , ) )dvec( )D D D D D C C A C AA C A C A A C CT T T
, ,( , ) dvec( ) dvec(, )( )HH C A C CA C A A CC
17
• Convergence requirement : Hessians are positive definite matrices
• Problem : Lack of positive definiteness Lack of convergence & slowness
• Solution : Necessity to regularization (i.e. Eigen-Value Decomposition (EVD) - based technique )
Newton
Optimization methodsOptimization methods
,( ( , ))EVD H A A A C U Σ UT
U : Matrix of eigen - vectors
Σ = diag{λ1,…,λNP} : diagonal matrix of eigen-values
EVD-based regularization
• Replace all negative eigen - values by one .
mNewton1
max/i i • Compute the ratio
• If 0.01 1i i
mNewton2
• Based on a linear approximation to the components of, in the neighborhood of A / C.
Levenberg-Marquardt (LM)
Update rules :
where is the Jacobian of in A .
Jacobians are computed from :and
• : damped parameter influencing both the direction and the size of the step [Madsen et al. 2004]
with :
Optimization methodsOptimization methods
18
Convergence speed VS SNR
• Noise-free random 3-order tensor
• Noisy 3-way array :
: Zero-mean normally distributed noise
: Scalar controling the noise level
• Results averaged over 200 Monte Carlo’s realizations.
Numerical resultsNumerical results
19
Convergence speed VS SNR
SNR = 0 dB
Numerical resultsNumerical results
20
Convergence speed VS SNR
SNR = 15 dB
Numerical resultsNumerical results
21
Convergence speed VS SNR
SNR = 30 dB
Numerical resultsNumerical results
22
• Differential concept Powerful tool for compact matrix derivations forms
• Global line search for symmetric case global optimum in the considered direction
• Iterative algorithms with global line search suitable step to reach the global optimum
ConclusionConclusion
23
Algebraic method + iterative method with global line search global optimum
• Solving an unconstrained semi-nonnegative INDSCAL problem .