multilinear-0.5.0.0: Comprehensive and efficient (multi)linear algebra implementation.

Copyright(c) Artur M. Brodzki 2018
LicenseBSD3
Maintainerartur@brodzki.org
Stabilityexperimental
PortabilityWindows/POSIX
Safe HaskellNone
LanguageHaskell2010

Multilinear.Class

Description

Multilinear library provides efficient and terse way to deal with linear algebra in Haskell. It is based on concept of tensor - a multidimensional indexed array of numbers. Vectors, matrices and linear functionals are examples of low-order tensors.

Quick tutorial

Tensors are indexed, each having one or several indices numbering its components. A scalar has no indices, as being simply a value. A vector has one index, because it is a one-dimensional list of numbers. A matrix has two indices, being a bidimensional table of numbers. Tensors are arbitrarily - dimensional, so you can have a vector of matrices (third-order tensor) and so on. Such tensors are especially useful when dealing with more complex task, such as programming neural networks e.g. for deep learning.

Index may be either lower (contravariant) or upper (covariant), depending on its function.

A vector has one upper or contravariant index, because it represents a value, point in a space. A linear functional (called in this library as a form) is represented as a list of numbers, similar to vector, but it does not indicate a value, but rather a coefficients of linear map or transformation that takes a vector and returns a number, simply by calculating a linear combination of vector components with weights being a linear functional coefficients. So elements of linear functional are indexed by lower or covariant index.

A matrix also represents a linear transformation, similar to linear functional, but it take a vector and return a vector, not a scalar. So matrix is simply a bunch of linear functionals - each one is a matrix row and returns a one element of resulting vector. These linear functionals are also indexed - of course by an contravariant index, as vector, which is returned by a matrix. So matrix is a tensor with two indices - one uppper (contravariant) and one lower (covariant).

A dot product is - what a surprise! - a linear transformation. But it is different from matrix - it takes two vectors to multiplicate and returns one number. So it needs two lower (covariant) indices. In fact, a dot product is represented as a bidimensional table of numbers - as matrix - but with two lower indices, instead of one upper and one lower. Tensors with several lower indices are called n-forms. Actually, many linear maps that takes two vectors and returns a number are represented in such notation. A dot product is a table filled with zeros but having a 1-s in diagonal of the table - it look visually identical like unit matrix, but it isn't a matrix - it's actually a 2-form. In this library dot product is called (as in abstract maths) a Kronecker delta (in Multilinear.NForm").

A cross product takes two vectors and returns a vector of numbers. So, as you can guess, it has three indices - two covariant (lower) and one upper (contravariant). A tensor that corresponds to cross product is called Levi-Civita symbol and is implemented under Multilinear.Tensor.

As you see, multidimensional indexed arrays - tensors - provide a unified way to deal with linear algebra - such with data (vectors) and functions operating on them (matrices, n-forms). Using tensors instead of limited number of pre-defined functions, you gain a power of universal formalism, where you can express any linear function you want to deal with.

When you apply a linear functional to a vector, you compute a linear combination - or a weighted sum - of its components. It is so common operation in linear algebra, that it requires a convenient way to note. Einstein (1905) introduced a summation rule:

If you multiply a tensor with lower index i by a tensor with upper index with same name /i then the linear combination of its components is automatically computed. The only condition is that an index of vector and index of corresponding linear transformation must have the same name.

Examples

>>> v = Vector.fromIndices "i" 5 $ \i -> i + 2 (vector v has 5 elements and v[i] = i + 2, indexing from 0)
>>> v
<i:5>
   | 2
   | 3
   | 4
   | 5
   | 6
>>> f = Form.const "i" 5 1 (linear functional has 5 elements and f[i] = 1)
>>> f
[i:5] [1,1,1,1,1]
>>> f * v
20

If you want to apply a vector to matrix, you must simply multiply them. The only condition is - as you can guess - that lower (indicating that matrix is a linear transformation) index of matrix must have the same name as the upper index of vector (indicating that this vector is a argument of matrix linear transformation; an upper index of matrix here indicates that this linear transformation returns a vector)

>>> m = Matrix.fromIndices "ij" 5 5 $ \i j -> i + j
>>> m
<i:5>
   | [j:5] [0,1,2,3,4]
   | [j:5] [1,2,3,4,5]
   | [j:5] [2,3,4,5,6]
   | [j:5] [3,4,5,6,7]
   | [j:5] [4,5,6,7,8]
>>> v = Vector.fromIndices "j" 5 $ \j -> j (vector v has 5 elements and v[j] = j, indexing from 0)
>>> v
<j:5>
   | 0
   | 1
   | 2
   | 3
   | 4
>>> m * v
<i:5>
   | 30
   | 40
   | 50
   | 60
   | 70

Note, that result vector is indexed with "i" - the same index as uppper index of our matrix.

If you want to do a matrix multiplication, a lower index of first matrix must have the same name as upper index of second matrix.

>>> m1 = Matrix.fromIndices "ij" 3 5 $ \i j -> i + j
>>> m1
<i:3>
   | [j:5] [0,1,2,3,4]
   | [j:5] [1,2,3,4,5]
   | [j:5] [2,3,4,5,6]
>>> m2 = Matrix.fromIndices "jk" 5 4 $ \j -> i + j (vector v has 5 elements and v[j] = j, indexing from 0)
>>> m2
<j:5>
   | [k:4] [0,1,2,3]
   | [k:4] [1,2,3,4]
   | [k:4] [2,3,4,5]
   | [k:4] [3,4,5,6]
   | [k:4] [4,5,6,7]
>>> m1 * m2
<i:3>
   | [k:4] [30,40,50,60]
   | [k:4] [40,50,60,70]
   | [k:4] [50,60,70,80]

Note, that rule that first matrix lower index must be the same as the second matrix upper index corresponeds to the fact, that to multiply two matrices, the widht of first matrix must be the same as width of the second matrix. The rule of matrix multiplication guarantees, that this operation is equivalent to linear functions composition.

The dot product of vectors may be done by simply making one of vectors with lower index, or - more correctly - by applying two vectors to Kronecker delta:

>>> v1 = Vector.fromIndices "i" 5 $ \i -> i + 2 (vector v has 5 elements and v[i] = i + 2, indexing from 0)
>>> v1
<i:5>
   | 2
   | 3
   | 4
   | 5
   | 6
>>> v2 = Vector.fromIndices "j" 5 $ \j -> j (vector v has 5 elements and v[j] = j, indexing from 0)
>>> v2
<j:5>
   | 0
   | 1
   | 2
   | 3
   | 4
>>> v1 * (lower "j" v2)
50
>>> NForm.dot 5 "ij" * v1 * v2 (A Kronecker delta - representing a dot product - of size 5 with indices "i" and "j" is multiplied by v1 and v2)
50

If you want to know more about linear algebra and Einstein convention, read Wikipedia:

Synopsis

Documentation

class (Generic (t a), NFData a, NFData (t a), Unbox a) => Multilinear t a where Source #

Multidimensional array treated as multilinear map - tensor Superclasses: - Generic (t a) - enforce Generic instance, allowing to derive e.g. serialization instances for any tensor - NFData a, NFData (t a) - for performance reasons, every tensor should be able to be fully evaluated - Unboxed.Unbox a - for performance reasons, every tensor should be implemented using Vector.Unboxed

Minimal complete definition

fromIndices, el, indices, order, size, ($|), transpose, standardize

Methods

fromIndices Source #

Arguments

:: String

Upper indices names (one character per index)

-> String

Lower indices names (one character per index)

-> [Int]

Upper indices sizes

-> [Int]

Lower indices sizes

-> ([Int] -> [Int] -> a)

Combinator function (f [u1,u2,...] [d1,d2,...] returns a tensor element at t [u1,u2,...] [d1,d2,...])

-> t a

Generated tensor

Generic tensor constructor, using combinator function on its indices

const Source #

Arguments

:: String

Upper indices names (one character per index)

-> String

Lower indices names (one character per index)

-> [Int]

Upper indices sizes

-> [Int]

Lower indices sizes

-> a 
-> t a

Generated tensor

Tensor constructor, that returns tensor with all elments equal to v.

el :: t a -> (String, [Int]) -> t a Source #

Accessing tensor elements

el ["i","j"] t [4,5] returns all tensor elements which index i is equal to 4 and index j is equal to 5. Values of other indices are insignificant

If given index value is out of range, then modulo operation is performed: el ["i","j"] t [40 50] = t[40 mod size i, 50 mod size j]

($$|) :: t a -> (String, [Int]) -> t a infixl 7 Source #

Infix equivalent for el

indices :: t a -> [TIndex] Source #

List of all tensor indices

indicesNames :: t a -> [String] Source #

List of tensor indices names

order :: t a -> (Int, Int) Source #

Tensor order - number of covariant and contravariant indices

order t = (cv, cov) where cv is number of upper and cov is number of lower indices

size :: t a -> String -> Int Source #

Return size of index with given name

equiv :: t a -> t a -> Bool Source #

Check if tensors are equivalent (have same indices but in different order)

(|==|) :: t a -> t a -> Bool infixl 1 Source #

Infix equivalent of equiv. Has low priority equal to 1. |

($|) :: t a -> (String, String) -> t a infix 9 Source #

t $| "ij" "kl" renames upper indices of tensor t to ij and lower indices to kl

raise :: t a -> String -> t a Source #

raise t "i" raises an index i of tensor t

(/\) :: t a -> String -> t a infixl 8 Source #

Infix equivalent of raise

lower :: t a -> String -> t a Source #

lower t "i" lowers an index i of tensor t

(\/) :: t a -> String -> t a infixl 8 Source #

Infix equivalent of lower

transpose :: t a -> t a Source #

Switch all indices of tensor t - upper indices becomes lower and vice versa

shiftRight :: t a -> String -> t a Source #

Shift tensor index right

shiftRight t "i" moves index i of tensor t one level depeer in recursion. Elements of tensor as indexed with indices names becomes unchanged.

Right shift of an index is equivalent to left shift of its successor in recursion s, if only s exists, so: Given a tensor t[i1,i2,i3,...]: shiftRight t "i2" == t[i1,i3,i2,...] == shiftLeft t "i3"

(|>>) :: t a -> String -> t a infixl 6 Source #

Infix equivalent of shiftRight

t |>> "i" moves index i of tensor t one level depeer in recursion

shiftRightmost :: t a -> String -> t a Source #

Shift tensor index rightmost

shiftRightmost t "i" moves index i of tensor t to the deepest level in recursion. Elements of tensor as indexed with indices names becomes unchanged.

(|>>>) :: t a -> String -> t a infixl 6 Source #

Infix equivalent of shiftRightmost

t |>>> "i" moves index i of tensor t to the deepest level in recursion

shiftLeft :: t a -> String -> t a Source #

Shift tensor index left. Elements of tensor as indexed with indices names becomes unchanged.

shiftLeft t "i" moves index i of tensor t one level up in recursion

Left shift of an index is equivalent to right shift of its predecessor in recursion p, if only p exists, so: Given a tensor t[i1,i2,i3,...]: shiftLeft t "i3" == t[i1,i3,i2,...] == shiftRight t "i2"

(<<|) :: t a -> String -> t a infixl 6 Source #

Infix equivalent to shiftLeft

t <<| "i" moves index i of tensor t one level up in recursion

shiftLeftmost :: t a -> String -> t a Source #

Shift tensor index leftmost. Elements of tensor as indexed with indices names becomes unchanged.

shiftLeftmost t "i" moves index i of tensor t to the first level in recursion

(<<<|) :: t a -> String -> t a infixl 6 Source #

Infix equivalent of shiftLeftmost

t <<<| "i" moves index i of tensor t to the first level in recursion

standardize :: t a -> t a Source #

Move covariant indices to deeper recursion level

Instances
(NFData a, Unbox a) => Multilinear Tensor a Source # 
Instance details

Defined in Multilinear.Generic.Sequential

(NFData a, Unbox a) => Multilinear Tensor a Source # 
Instance details

Defined in Multilinear.Generic.MultiCore