Intel® Fortran Compiler 17.0 Developer Guide and Reference
Loops can be formed with the usual DO-END DO and DO WHILE, or by using an IF/GOTO and a label. Loops must have a single entry and a single exit to be vectorized. The following examples illustrate loop constructs that can and cannot be vectorized.
Example: Vectorizable structure |
---|
|
The following example shows a loop that cannot be vectorized because of the inherent potential for an early exit from the loop.
Example: Non-vectorizable structure |
---|
|
Loop exit conditions determine the number of iterations a loop executes. For example, fixed indexes for loops determine the iterations. The loop iterations must be countable; in other words, the number of iterations must be expressed as one of the following:
A constant.
A loop invariant term.
A linear function of outermost loop indices.
In the case where a loops exit depends on computation, the loops are not countable. The examples below show loop constructs that are countable and non-countable.
Example: Countable Loop |
---|
|
The following example demonstrates a different countable loop construct.
Example: Countable Loop |
---|
|
The following examples demonstrates a loop construct that is non-countable due to dependency loop variant count value.
Example: Non-Countable Loop |
---|
|
Strip-mining, also known as loop sectioning, is a loop transformation technique for enabling SIMD-encodings of loops, as well as a means of improving memory performance. By fragmenting a large loop into smaller segments or strips, this technique transforms the loop structure in two ways:
By increasing the temporal and spatial locality in the data cache if the data are reusable in different passes of an algorithm.
By reducing the number of iterations of the loop by a factor of the length of each vector, or number of operations being performed per SIMD operation. In the case of Intel® Streaming SIMD Extensions, this vector or strip-length is reduced by four times: four floating-point data items per single Intel® SSE single-precision floating-point SIMD operation are processed.
First introduced for vectorizers, this technique consists of the generation of code when each vector operation is done for a size less than or equal to the maximum vector length on a given vector machine.
The compiler automatically strip-mines your loop and generates a cleanup loop. For example, assume the compiler attempts to strip-mine the following loop:
Example: Before Vectorization |
---|
|
The compiler might handle the strip mining and loop cleaning by restructuring the loop in the following manner:
Example: After Vectorization |
---|
|
Loop Blocking
It is possible to treat loop blocking as strip-mining in two or more dimensions. Loop blocking is a useful technique for memory performance optimization. The main purpose of loop blocking is to eliminate as many cache misses as possible. This technique transforms the memory domain into smaller chunks rather than sequentially traversing through the entire memory domain. Each chunk should be small enough to fit all the data for a given computation into the cache, thereby maximizing data reuse.
Consider the following example. The two-dimensional array A is referenced in the j (column) direction and then in the i (row) direction (column-major order); array B is referenced in the opposite manner (row-major order). Assume the memory layout is in column-major order; therefore, the access strides of array A and B for the code would be 1 and MAX, respectively. BS = block_size;MAX must be evenly divisible by BS.
Consider the following loop example code:
Example: Original loop |
---|
REAL A(MAX,MAX), B(MAX,MAX) DO I =1, MAX DO J = 1, MAX A(I,J) = A(I,J) + B(J,I) ENDDO ENDDO |
The arrays could be blocked into smaller chunks so that the total combined size of the two blocked chunks is smaller than the cache size, which can improve data reuse. One possible way of doing this is demonstrated below:
Example: Transformed Loop after blocking |
---|
|
Loop interchange is often used for improving memory access patterns. Matrix multiplication is commonly written as shown in the following example:
Example: Typical Matrix Multiplication |
---|
|
The use of B(K,J) is not a stride-1 reference and therefore will not be vectorized efficiently.
If the loops are interchanged, however, all the references will become stride-1 as shown in the following example.
Example: Matrix Multiplication with Stride-1 |
---|
|
Interchanging is not always possible because of dependencies, which can lead to different results.