cholesky classes are now in separate files
This commit is contained in:
parent
406f2abbc8
commit
c4d2f8eaf9
1 changed files with 0 additions and 310 deletions
|
|
@ -48,316 +48,6 @@ public:
|
|||
typedef typename EOT::AtomType AtomType;
|
||||
|
||||
|
||||
/** Cholesky decomposition, given a matrix V, return a matrix L
|
||||
* such as V = L L^T (L^T being the transposed of L).
|
||||
*
|
||||
* Need a symmetric and positive definite matrix as an input, which
|
||||
* should be the case of a non-ill-conditionned covariance matrix.
|
||||
* Thus, expect a (lower) triangular matrix.
|
||||
*/
|
||||
class Cholesky
|
||||
{
|
||||
public:
|
||||
//! The covariance-matrix is symetric
|
||||
typedef ublas::symmetric_matrix< AtomType, ublas::lower > CovarMat;
|
||||
|
||||
//! The factorization matrix is triangular
|
||||
// FIXME check if triangular types behaviour is like having 0
|
||||
typedef ublas::matrix< AtomType > FactorMat;
|
||||
|
||||
enum Method {
|
||||
//! use the standard algorithm, with square root @see factorize_LLT
|
||||
standard,
|
||||
//! use the algorithm using absolute value within the square root @see factorize_LLT_abs
|
||||
absolute,
|
||||
//! use the method that set negative square roots to zero @see factorize_LLT_zero
|
||||
zeroing,
|
||||
//! use the robust algorithm, without square root @see factorize_LDLT
|
||||
robust
|
||||
};
|
||||
Method _use;
|
||||
|
||||
|
||||
/** Instanciate without computing anything, you are responsible of
|
||||
* calling the algorithm and getting the result with operator()
|
||||
* */
|
||||
Cholesky( Cholesky::Method use = standard ) : _use(use) {}
|
||||
|
||||
|
||||
/** Computation is made at instanciation and then cached in a member variable,
|
||||
* use decomposition() to get the result.
|
||||
*
|
||||
* Use the standard unstable method by default.
|
||||
*/
|
||||
Cholesky(const CovarMat& V, Cholesky::Method use = standard ) :
|
||||
_use(use), _L(ublas::zero_matrix<AtomType>(V.size1(),V.size2()))
|
||||
{
|
||||
factorize( V );
|
||||
}
|
||||
|
||||
|
||||
/** Compute the factorization and return the result
|
||||
*/
|
||||
const FactorMat& operator()( const CovarMat& V )
|
||||
{
|
||||
factorize( V );
|
||||
return decomposition();
|
||||
}
|
||||
|
||||
//! The decomposition of the covariance matrix
|
||||
const FactorMat & decomposition() const
|
||||
{
|
||||
return _L;
|
||||
}
|
||||
|
||||
protected:
|
||||
|
||||
//! The decomposition is a (lower) symetric matrix, just like the covariance matrix
|
||||
FactorMat _L;
|
||||
|
||||
/** Assert that the covariance matrix have the required properties and returns its dimension.
|
||||
*
|
||||
* Note: if compiled with NDEBUG, will not assert anything and just return the dimension.
|
||||
*/
|
||||
unsigned assert_properties( const CovarMat& V )
|
||||
{
|
||||
unsigned int Vl = V.size1(); // number of lines
|
||||
|
||||
// the result goes in _L
|
||||
_L = ublas::zero_matrix<AtomType>(Vl,Vl);
|
||||
|
||||
#ifndef NDEBUG
|
||||
assert(Vl > 0);
|
||||
|
||||
unsigned int Vc = V.size2(); // number of columns
|
||||
assert(Vc > 0);
|
||||
assert( Vl == Vc );
|
||||
|
||||
// partial assert that V is semi-positive definite
|
||||
// assert that all diagonal elements are positives
|
||||
for( unsigned int i=0; i < Vl; ++i ) {
|
||||
assert( V(i,i) > 0 );
|
||||
}
|
||||
|
||||
/* FIXME what is the more efficient way to check semi-positive definite? Candidates are:
|
||||
* perform the cholesky factorization
|
||||
* check if all eigenvalues are positives
|
||||
* check if all of the leading principal minors are positive
|
||||
*/
|
||||
#endif
|
||||
|
||||
return Vl;
|
||||
}
|
||||
|
||||
|
||||
/** Actually performs the factorization with the method given at
|
||||
* instanciation. Results are cached in _L.
|
||||
*/
|
||||
void factorize( const CovarMat& V )
|
||||
{
|
||||
if( _use == standard ) {
|
||||
factorize_LLT( V );
|
||||
} else if( _use == absolute ) {
|
||||
factorize_LLT_abs( V );
|
||||
} else if( _use == zeroing ) {
|
||||
factorize_LLT_zero( V );
|
||||
} else if( _use == robust ) {
|
||||
factorize_LDLT( V );
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
/** This standard algorithm makes use of square root and is thus subject
|
||||
* to round-off errors if the covariance matrix is very ill-conditioned.
|
||||
*
|
||||
* Compute L such that V = L L^T
|
||||
*
|
||||
* When compiled in debug mode and called on ill-conditionned matrix,
|
||||
* will raise an assert before calling the square root on a negative number.
|
||||
*/
|
||||
void factorize_LLT( const CovarMat& V)
|
||||
{
|
||||
unsigned int N = assert_properties( V );
|
||||
|
||||
unsigned int i=0, j=0, k;
|
||||
_L(0, 0) = sqrt( V(0, 0) );
|
||||
|
||||
// end of the column
|
||||
for ( j = 1; j < N; ++j ) {
|
||||
_L(j, 0) = V(0, j) / _L(0, 0);
|
||||
}
|
||||
|
||||
// end of the matrix
|
||||
for ( i = 1; i < N; ++i ) { // each column
|
||||
// diagonal
|
||||
double sum = 0.0;
|
||||
for ( k = 0; k < i; ++k) {
|
||||
sum += _L(i, k) * _L(i, k);
|
||||
}
|
||||
|
||||
// round-off errors may appear here
|
||||
assert( V(i,i) - sum >= 0 );
|
||||
_L(i,i) = sqrt( V(i,i) - sum );
|
||||
|
||||
for ( j = i + 1; j < N; ++j ) { // rows
|
||||
// one element
|
||||
sum = 0.0;
|
||||
for ( k = 0; k < i; ++k ) {
|
||||
sum += _L(j, k) * _L(i, k);
|
||||
}
|
||||
|
||||
_L(j, i) = (V(j, i) - sum) / _L(i, i);
|
||||
|
||||
} // for j in ]i,N[
|
||||
} // for i in [1,N[
|
||||
}
|
||||
|
||||
|
||||
/** This standard algorithm makes use of square root but do not fail
|
||||
* if the covariance matrix is very ill-conditioned.
|
||||
* Here, we propagate the error by using the absolute value before
|
||||
* computing the square root.
|
||||
*
|
||||
* Be aware that this increase round-off errors, this is just a ugly
|
||||
* hack to avoid crash.
|
||||
*/
|
||||
void factorize_LLT_abs( const CovarMat & V)
|
||||
{
|
||||
unsigned int N = assert_properties( V );
|
||||
|
||||
unsigned int i=0, j=0, k;
|
||||
|
||||
_L(0, 0) = sqrt( V(0, 0) );
|
||||
|
||||
// end of the column
|
||||
for ( j = 1; j < N; ++j ) {
|
||||
_L(j, 0) = V(0, j) / _L(0, 0);
|
||||
}
|
||||
|
||||
// end of the matrix
|
||||
for ( i = 1; i < N; ++i ) { // each column
|
||||
// diagonal
|
||||
double sum = 0.0;
|
||||
for ( k = 0; k < i; ++k) {
|
||||
sum += _L(i, k) * _L(i, k);
|
||||
}
|
||||
|
||||
_L(i,i) = sqrt( fabs( V(i,i) - sum) );
|
||||
|
||||
for ( j = i + 1; j < N; ++j ) { // rows
|
||||
// one element
|
||||
sum = 0.0;
|
||||
for ( k = 0; k < i; ++k ) {
|
||||
sum += _L(j, k) * _L(i, k);
|
||||
}
|
||||
|
||||
_L(j, i) = (V(j, i) - sum) / _L(i, i);
|
||||
|
||||
} // for j in ]i,N[
|
||||
} // for i in [1,N[
|
||||
}
|
||||
|
||||
|
||||
/** This standard algorithm makes use of square root but do not fail
|
||||
* if the covariance matrix is very ill-conditioned.
|
||||
* Here, if the diagonal difference ir negative, we set it to zero.
|
||||
*
|
||||
* Be aware that this increase round-off errors, this is just a ugly
|
||||
* hack to avoid crash.
|
||||
*/
|
||||
void factorize_LLT_zero( const CovarMat & V)
|
||||
{
|
||||
unsigned int N = assert_properties( V );
|
||||
|
||||
unsigned int i=0, j=0, k;
|
||||
|
||||
_L(0, 0) = sqrt( V(0, 0) );
|
||||
|
||||
// end of the column
|
||||
for ( j = 1; j < N; ++j ) {
|
||||
_L(j, 0) = V(0, j) / _L(0, 0);
|
||||
}
|
||||
|
||||
// end of the matrix
|
||||
for ( i = 1; i < N; ++i ) { // each column
|
||||
// diagonal
|
||||
double sum = 0.0;
|
||||
for ( k = 0; k < i; ++k) {
|
||||
sum += _L(i, k) * _L(i, k);
|
||||
}
|
||||
|
||||
if( V(i,i) - sum >= 0 ) {
|
||||
_L(i,i) = sqrt( V(i,i) - sum);
|
||||
} else {
|
||||
_L(i,i) = 0;
|
||||
}
|
||||
|
||||
for ( j = i + 1; j < N; ++j ) { // rows
|
||||
// one element
|
||||
sum = 0.0;
|
||||
for ( k = 0; k < i; ++k ) {
|
||||
sum += _L(j, k) * _L(i, k);
|
||||
}
|
||||
|
||||
_L(j, i) = (V(j, i) - sum) / _L(i, i);
|
||||
|
||||
} // for j in ]i,N[
|
||||
} // for i in [1,N[
|
||||
}
|
||||
|
||||
|
||||
/** This alternative algorithm do not use square root in an inner loop,
|
||||
* but only for some diagonal elements of the matrix D.
|
||||
*
|
||||
* Computes L and D such as V = L D L^T.
|
||||
* The factorized matrix is (L D^1/2), because V = (L D^1/2) (L D^1/2)^T
|
||||
*/
|
||||
void factorize_LDLT( const CovarMat& V)
|
||||
{
|
||||
// use "int" everywhere, because of the "j-1" operation
|
||||
int N = assert_properties( V );
|
||||
// example of an invertible matrix whose decomposition is undefined
|
||||
assert( V(0,0) != 0 );
|
||||
|
||||
FactorMat L = ublas::zero_matrix<AtomType>(N,N);
|
||||
FactorMat D = ublas::zero_matrix<AtomType>(N,N);
|
||||
D(0,0) = V(0,0);
|
||||
|
||||
for( int j=0; j<N; ++j ) { // each columns
|
||||
L(j, j) = 1;
|
||||
|
||||
D(j,j) = V(j,j);
|
||||
for( int k=0; k<=j-1; ++k) { // sum
|
||||
D(j,j) -= L(j,k) * L(j,k) * D(k,k);
|
||||
}
|
||||
|
||||
for( int i=j+1; i<N; ++i ) { // remaining rows
|
||||
|
||||
L(i,j) = V(i,j);
|
||||
for( int k=0; k<=j-1; ++k) { // sum
|
||||
L(i,j) -= L(i,k)*L(j,k) * D(k,k);
|
||||
}
|
||||
L(i,j) /= D(j,j);
|
||||
|
||||
} // for i in rows
|
||||
} // for j in columns
|
||||
|
||||
// now compute the final symetric matrix: _L = L D^1/2
|
||||
// remember that V = ( L D^1/2) ( L D^1/2)^T
|
||||
|
||||
// fortunately, the square root of a diagonal matrix is the square
|
||||
// root of all its elements
|
||||
FactorMat D12 = D;
|
||||
for(int i=0; i<N; ++i) {
|
||||
D12(i,i) = sqrt(D(i,i));
|
||||
}
|
||||
|
||||
// the factorization is thus _L*D^1/2
|
||||
_L = ublas::prod( L, D12);
|
||||
}
|
||||
|
||||
}; // class Cholesky
|
||||
|
||||
|
||||
edoSamplerNormalMulti( edoRepairer<EOT> & repairer, typename Cholesky::Method use = Cholesky::absolute )
|
||||
: edoSampler< EOD >( repairer), _cholesky(use)
|
||||
|
|
|
|||
Reference in a new issue