Skip to content

hngenc/systolic-array

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

A Systolic(ish) Hardware Description Language

This (as-yet-unnamed) project provides a Scala-embedded DSL which can be used to describe systolic arrays, a type of highly-parallel processor. The language helps users to separate the functional correctness of their design from the dataflow which it implements, and enables users to quickly iterate over many different dataflows to find the one which best suits their purposes.

Introduction

With the end of Moore's Law, CPU performance is no longer increasing at the speed it once did. To run compute-intensive workloads today, computer architects are increasingly turning towards specialized hardware accelerators, especially in cases where there is massive parallelism or locality that traditional CPUs are ill-equipped to exploit. Systolic arrays are one such type of hardware accelerator; they have successfully been used to accelerate important algorithms such as convolution1, matrix multiplication2, sorting3 and many more, both in industry and in academia.

Systolic arrays are two-dimensional arrays of small processing units (called PEs) carrying out the same function, which, in most cases, communicate only with the PEs directly adjacent to themselves. Thus, they help to lower communication costs, and they shine with workloads that exhibit high locality where data can be shared across many adjacent PEs within the same array.

How exactly that data flows through the array is determined by a mapping known as a spacetime transform, which this language helps users to explore. In the following sections, we illustrate by focusing on the example of a systolic array used to compute matrix multiplications.

Matrix Multiplication

Matrix multiplications can be expressed with the following simple 3-level nested for loop:

for i ∈ [0, N1]:
    for j ∈ [0, N2]:
        C(i, j) = 0
        for k ∈ [0, N3]:
            C(i, j) += A(i, k) * B(k, j)

In this example, an N1 by N3 matrix A is multiplied an N3 by N2 matrix B to produce an N1 by N2 matrix C.

Matrix multiplications are an embarassingly parallelizable problem. The multiply-accumulates (MACs) in the innermost loop can be executed in parallel, and the results merged together afterwards, rather than executing them sequentially as the imperative notation used above implies.

In addition to the high-degree of parallelism, matrix multiplications exhibit high degrees of temporal and spatial locality. Every element within A is re-used for N2 MACs, while every element of B is re-used N1 times. Additionally, to build up a single element of C, N3 partial sums must be accumulated into a single result.

The promising potential for both parallelism and locality suggests that systolic arrays would be well suited for calculating matrix multiplications. However, the imperative notation used above obscures the parallelism and data movement patterns, which makes it difficult to map the function to a 2D systolic array. A functional notation, described below, is more appropriate.

Functional Notation

The imperative loop written above can be re-expressed using an assignment-free, functional notation. For example, we can condense the three loop indices given above into a single iteration vector, (i, j, k), and use a different variable to denote each intermediate value of C at each iteration:

// Input
c(i, j, 0) = 0

// Calculation
c(i, j, k) = c(i, j, k-1) + A(i, k) * B (k, j)

// Output
C(i, j) = c(i, j, N3)

This is a good start, as it expresses how c values computed in different (perhaps concurrent) iterations can be merged into a final C output. However, it fails to express how elements of A and B can also be shared and re-used. Fortunately, we can add new variables, a and b, to do so:

// Input
a(i, 0, k) = A(i, k)
b(0, j, k) = B(k, j)
c(i, j, 0) = 0

// Calculation
a(i, j, k) = a(i, j-1, k)
b(i, j, k) = b(i-1, j, k)
c(i, j, k) = c(i, j, k-1) + A(i, k) * B (k, j)

// Output
C(i, j) = c(i, j, N3)

And now, it becomes more apparent that a travels along the j index of the iteration vector, while b is shared along the i index.

This project allows users to describe the functionality of their systolic arrays using nearly the same syntax (but with more boilerplate):

class MatMul extends Systolic {
  // Our loop-index bounds
  val N1, N2, N3 = 1

  // The iterators that make up our iteration index, with their bounds
  val i = Iterator(0, N1)
  val j = Iterator(0, N2)
  val k = Iterator(0, N3)

  // The external inputs and outputs to our systolic array
  val A = Input(i, k)
  val B = Input(k, j)
  val C = Output(i, j)

  // The "local" variables that travel through the systolic array
  val a, b, c = Local(16) // 16 bits wide

  // Inputs
  a(i, 0, k) := A(i, k)
  b(0, j, k) := B(k, j)
  c(i, j, 0) := 0
  
  // Calculations
  a(i, j, k) := a(i, j-1, k)
  b(i, j, k) := b(i-1, j, k)
  c(i, j, k) := c(i, j, k-1) + (a(i, j-1, k) * b(i-1, j, k))

  // Outputs
  C(i, j) := c(i, j, N3)
}

Once our functional algorithm has been described, we need a way to map it onto a 2D systolic array. The next section explains how.

Spacetime Transforms

Using our functional notation, we have now specified which iteration each MAC operation should occur in. We must now map these iteration vectors to spacetime vectors, (x, y, t), which describe the x and y coordinates on the 2D systolic array each MAC should occur in, as well as the time step, t, in which that operation should occur.

Generalized spacetime transform

It turns out that to simplify matters, we can concentrate on only the set of linear transforms:

Linear spacetime transform

where T is a 3 by 3 matrix.

The power of these spacetime transforms quickly becomes apparent. Simply by modifying T, even if the matrix multiply's functional algorithm remains completely unchanged, very different systolic arrays can be produced without sacrificing the functional correctness of our algorithm.

  • Output-stationary matrix multiply:

    Output-stationary spacetime transformation

    Output-stationary systolic array
  • Weight-stationary matrix multiply:

    Weight-stationary spacetime transformation

    Weight-stationary systolic array
  • Hexagonal matrix multiply:

    Hexagonal-stationary spacetime transformation

    Hexagonal systolic array

Although all three examples above may ultimately compute the same outputs, they do so in very different ways. Not only do they differ in size or shape, but they also differ in how partial results "flow" throughout the array. In the first example, partial sums of C are fixed to specific PEs, and accumulate in place, while in the others, the partial sums travel across the array, before exiting at the edges. Based on your particular objectives or limitations, one strategy may be more attractive than another.

The language provided requires users to specify the particular spacetime transformation they wish to apply to their functional algorithm:

class MatMul extends Systolic {
  // Functional algorithm, as shown above
  // ...
  
  // Output-stationary spacetime transform
  spaceTimeTransform(Seq(
    Seq(1,0,0),
    Seq(0,1,0),
    Seq(1,1,1)))
}

Constraints

As your spacetime transforms become more complex, it can become increasingly difficult to come up with new ones in an ad-hoc manner, as it becomes harder and harder to visualize how each value in your matrix is going to affect the final shape of your array. Thus, this language allows users to set high-level constraints upon the dataflow they wish to achieve. The compiler will then attempt to generate a list of spacetime transforms that meet those constraints, and which are guaranteed to result in correctly functioning hardware (although there may be other valid spacetime transforms that the compiler misses).

To set a constraint for the direction in which a variable should flow, use the flow(v, (dx, dy)) method, which constrains a variable v to travel dx columns to the right, and dy rows down across the 2D array every cycle. There are also some more wrapper functions provided:

fix(v)   // flow(v, (0, 0))
flowR(v) // flow(v, (0, 1))
flowD(v) // flow(v, (1, 0))

For example, to create an output-stationary systolic array, without manually entering the transform, we can write:

class MatMul extends Systolic {
  // Functional algorithm, as shown above
  // ...
  
  // Output-stationary spacetime transform
  fix(c)
  flowR(a)
  flowD(b)
  spaceTimeTransform()
}

If there are multiple transforms which meet your constraints, the compiler will require you to choose one explicitly.

Compiler Output

After running the spaceTimeTransform() function, the compiler will build a mesh of PEs each implementing the function you specified, and then wire them together to match the dataflow determined by the spacetime transformation. The compiler will then generate a Verilog representation of the systolic array. It will also print out the exact input/output pattern of your systolic array, so you know at which time-steps to pass in your elements of A and B, and when you will be able to read out C (as well as the coordinates of the cells where the input/output operations take place).

The input/output pattern generated by the compiler for a 2 by 2 output-stationary systolic array is shown below:

Your input pattern for a is:
	At time 0, (0, 0) is input to cell (0, 0)
	At time 1, (0, 1) is input to cell (0, 0)
	At time 1, (1, 0) is input to cell (1, 0)
	At time 2, (1, 1) is input to cell (1, 0)

Your input pattern for b is:
	At time 0, (0, 0) is input to cell (0, 0)
	At time 1, (1, 0) is input to cell (0, 0)
	At time 1, (0, 1) is input to cell (0, 1)
	At time 2, (1, 1) is input to cell (0, 1)

Your output pattern for c is:
	At time 1, (0, 0) is output from cell (0, 0)
	At time 2, (0, 1) is output from cell (0, 1)
	At time 2, (1, 0) is output from cell (1, 0)
	At time 3, (1, 1) is output from cell (1, 1)

Mapping Larger Matrices

The output-stationary matrix we built above can only generate output matrices of size N1 by N2. In general, any of the matrices we build may be bounded by any (or all) of the initial loop bound indices. However, it is unlikely that we will know, while designing our matrix multiply unit, what the maximum size will be of the matrices that we wish to compute later. Thus, it becomes necessary to find ways to map bigger matrices into smaller matrix-multiply arrays.

For output- and weight-stationary arrays, this is easy to do in an ad-hoc way. However, without a formal way of expressing the different possible mapping strategies, it becomes difficult to perform any kind of design-space-exploration, which is especially important since mapping strategies may influence architectural characteristics like memory hierarchies.

Fortunately, it turns out that we can re-use the concepts we have learned so far to express different possible mapping strategies for bigger matrices. For example, we can copy the earlier matrix multiplication equation we had, but treat a, b, and c as submatrices instead of scalar values:

// Input
a(i, 0, k) = A(i, k)
b(0, j, k) = B(k, j)
c(i, j, 0) = 0

// Calculation
a(i, j, k) = a(i, j-1, k)
b(i, j, k) = b(i-1, j, k)
c(i, j, k) = c(i, j, k-1) + A(i, k) * B (k, j)

// Output
C(i, j) = c(i, j, N3)

Then, we can once again come up with spacetime transforms that describe how matrix multiplications can be performed on smaller matrices to construct a larger matrix. For example, for the following spacetime transform:

Spacetime transform

we can use the language provided in this project to calculate the following mapping:

Mapping

This figure is particularly interesting because the black dots represent instances where the matrix multiply unit (MM) re-uses a cached submatrix which had been input before to compute the c submatrices. Thus, we find that we can express both the mapping and the storage requirements imposed by that mapping using our functional equations and spacetime transforms.

Suppose that rather than allocating extra area to store cached submatrices, we would rather re-input them externally, essentially trading storage burdens for extra communication costs. In this case, rather than changing the spacetime transformation, we can modify our functional algorithm to eliminate all re-use of a and b, and to instead require a and b to be fed in externally every time step:

// Input
a(i, j, k) = A(i, k)
b(i, j, k) = B(k, j)
c(i, j, 0) = 0

// Calculation
c(i, j, k) = c(i, j, k-1) + A(i, k) * B (k, j)

// Output
C(i, j) = c(i, j, N3)

Then, using the same spacetime transform as earlier, we obtain the following mapping:

Mapping

As we can see, in this case, there is no local re-use of a and b.

We can also experiment with more and more complex mapping strategies. For example, if we have two matrix multiply units, and we wish to divide the computations equally between them, we can use the following transform:

Transform

to generate the following mapping:

Mapping

The main takeaway is that the language provided makes it easy to experiment with different mapping strategies, since the input/output submatrix patterns are automatically generated.

Future Work

The language works well currently as a proof of concept, but it lacks certain features that would give it practical value.

More importantly, the language needs automated design space exploration capabilities, to help users find the best dataflows based on their objectives (which may simply be to maximize performance, minimize power, or reduce area).

Additionally, the Mapping Larger Matrices section showed a case where it was necessary to slightly change the "functional algorithm" part of one's program to select a different mapping, even though the actual functional correctness of their algorithm was unchanged. This suggests that the decoupling between functional correctness and dataflow mapping was not extensive enough. In the future, the exact iterations in which variables are initialized from external inputs, versus when they are propagated locally, should also be based on parameters that can be tuned by an optimizer.

Finally, although the compiler will currently print out the input/output patterns, the task of designing the software/hardware interface between the accelerator and the host computer still lies with the user. In the future, it would be ideal if the compiler also generated the software interface, just as it generates the hardware now.

Further Reading

This chapter from Algorithms of Informatics discusses, in a format suitable for beginners, how systolic arrays can be designed with a formal methodology.

References

[1] Yang, Xuan, et al. "DNN Dataflow Choice Is Overrated." arXiv preprint arXiv:1809.04070 (2018).

[2] Jouppi, Norman P., et al. "In-datacenter performance analysis of a tensor processing unit." Computer Architecture (ISCA), 2017 ACM/IEEE 44th Annual International Symposium on. IEEE, 2017.

[3] Schwiegelshohn, Uwe. "A shortperiodic two-dimensional systolic sorting algorithm." Systolic Arrays, 1988., Proceedings of the International Conference on. IEEE, 1988.

About

A DSL for Systolic Arrays

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages