Zum Inhalt springen

„Blockmatrix“ – Versionsunterschied

aus Wikipedia, der freien Enzyklopädie
[ungesichtete Version][ungesichtete Version]
Inhalt gelöscht Inhalt hinzugefügt
MathMartin (Diskussion | Beiträge)
MathMartin (Diskussion | Beiträge)
rewrote example
Zeile 1: Zeile 1:
In the [[mathematics|mathematical]] subfield of [[matrix theory]], a '''block matrix''' or a '''partitioned matrix''' is a partition of a [[Matrix (mathematics)|matrix]] into rectangular smaller matrices called '''blocks'''. Looking at it another way, the matrix is written in terms of smaller matrices written side-by-side. A block matrix must conform to a consistent way of splitting up the rows, and the columns: we group the rows into some adjacent 'bunches', and the columns likewise. The partition is into the rectangles described by one bunch of adjacent rows crossing one bunch of adjacent columns. In other words, the matrix is split up by some horizontal and vertical lines that go all the way across.
In the [[mathematics|mathematical]] subfield of [[matrix theory]], a '''block matrix''' or a '''partitioned matrix''' is a partition of a [[Matrix (mathematics)|matrix]] into rectangular smaller matrices called '''blocks'''. Looking at it another way, the matrix is written in terms of smaller matrices written side-by-side. A block matrix must conform to a consistent way of splitting up the rows, and the columns: we group the rows into some adjacent 'bunches', and the columns likewise. The partition is into the rectangles described by one bunch of adjacent rows crossing one bunch of adjacent columns. In other words, the matrix is split up by some horizontal and vertical lines that go all the way across.


== Example ===
== Example ==


The matrix
A '''partitioned matrix''' or ''[[block matrix]]'' is a matrix of matrices. For example, take a matrix ''P'':


:<math>P = \begin{bmatrix}
:<math>P = \begin{bmatrix}
Zeile 11: Zeile 11:
6 & 1 & 5 & 8\end{bmatrix}</math>
6 & 1 & 5 & 8\end{bmatrix}</math>


We could partition it into a 2&times;2 partitioned matrix like this:
can be partitioned into 4 2&times;2 blocks


:<math>P_{11} = \begin{bmatrix}
:<math>P_{11} = \begin{bmatrix}
Zeile 22: Zeile 22:
2 & 6\\
2 & 6\\
5 & 8\end{bmatrix}</math>
5 & 8\end{bmatrix}</math>

The partitioned matrix can then be written as


:<math>P_{\mathrm{partitioned}} = \begin{bmatrix}
:<math>P_{\mathrm{partitioned}} = \begin{bmatrix}

Version vom 26. September 2004, 16:33 Uhr

In the mathematical subfield of matrix theory, a block matrix or a partitioned matrix is a partition of a matrix into rectangular smaller matrices called blocks. Looking at it another way, the matrix is written in terms of smaller matrices written side-by-side. A block matrix must conform to a consistent way of splitting up the rows, and the columns: we group the rows into some adjacent 'bunches', and the columns likewise. The partition is into the rectangles described by one bunch of adjacent rows crossing one bunch of adjacent columns. In other words, the matrix is split up by some horizontal and vertical lines that go all the way across.

Example

The matrix

can be partitioned into 4 2×2 blocks

The partitioned matrix can then be written as

Application

In linear algebra terms, the use of a block matrix corresponds to having a linear mapping thought of in terms of corresponding 'bunches' of basis vectors. That again matches the idea of having distinguished direct sum decompositions of the domain and range. It is always particularly significant if a block is the zero matrix; that carries the information that a summand maps into a sub-sum.

Given the interpretation via linear mappings and direct sums, there is a special type of block matrix that occurs for square matrices (the case m = n). For those we can assume an interpretation as an endomorphism of an n-dimensional space V; the block structure in which the bunching of rows and columns is the same is of importance because it corresponds to having a single direct sum decomposition on V (rather than two). In that case, for example, the diagonal blocks in the obvious sense are all square. This type of structure is required to describe the Jordan normal form.

This technique is used to cut down calculations of matrices, column-row expansions, and many computer science applications, including VLSI chip design. An example is the Strassen algorithm for fast matrix multiplication.