Recitation 7 Memory Access Patterns Andrew Faulring 15213 Section A 21 October 2002 Andrew Faulring faulring cs cmu edu Office hours NSH 2504 lab 2507 conference room Thursday 5 6 Lab 4 due Thursday 24 Oct 11 59pm Submission is online http www2 cs cmu edu afs cs academic class 15213 f02 www L4 html Today s Plan Loop Unrolling Blocking Loop Unrolling void combine5 vec ptr v int dest int length vec length v int limit length 2 int data get vec start v int sum 0 int i Combine 3 elements at a time for i 0 i limit i 3 sum data i data i 1 data i 2 Finish any remaining elements for i length i sum data i dest sum Optimization Combine multiple iterations into single loop body Amortizes loop overhead across multiple iterations Finish extras at end Practice Problem Problem 5 12 and 5 13 Solution 5 12 void inner5 vec ptr u vec ptr v data t dest int i int length vec length u int limit length 3 data t udata get vec start u data t vdata get vec start v data t sum data t 0 Do four elements at a time for i 0 i limit i 4 sum udata i vdata i udata i 1 vdata i 1 udata i 2 vdata i 2 udata i 3 vdata i 3 Finish off any remaining elements for i length i sum udata i vdata i dest sum Solution 5 12 A We must perform two loads per element to read values for udata and vdata There is only one unit to perform these loads and it requires one cycle B The performance for floating point is still limited by the 3 cycle latency of the floating point adder Solution 5 13 void inner6 vec ptr u vec ptr v data t dest int i int length vec length u int limit length 3 data t udata get vec start u data t vdata get vec start v data t sum0 data t 0 data t sum1 data t 0 Do four elements at a time for i 0 i limit i 4 sum0 udata i vdata i sum1 udata i 1 vdata i 1 sum0 udata i 2 vdata i 2 sum1 udata i 3 vdata i 3 Finish off any remaining elements for i length i sum0 sum0 udata i vdata i dest sum0 sum1 Solution 5 13 For each element we must perform two loads with a unit that can only load one value per clock cycle We must also perform one floating point multiplication with a unit that can only perform one multiplication every two clock cycles Both of these factors limit the CPE to 2 Summary of Matrix Multiplication ijk jik kij ikj 2 loads 0 stores 2 loads 1 store 2 loads 1 store misses iter 1 25 misses iter 0 5 misses iter 2 0 for i 0 i n i for k 0 k n k for j 0 j n j for k 0 k n k sum 0 0 r a i k r b k j for k 0 k n k for j 0 j n j for i 0 i n i sum a i k c i j r c i j b k j c i j sum for j 0 j n j for i 0 i n i b k j jki kji a i k r Improving Temporal Locality by Blocking Example Blocked matrix multiplication block in this context does not mean cache block Instead it mean a sub block within the matrix Example N 8 sub block size 4 A11 A12 A21 A22 B11 B12 X B21 B22 C11 C12 C21 C22 Key idea Sub blocks i e Axy can be treated just like scalars C11 A11B11 A12B21 C12 A11B12 A12B22 C21 A21B11 A22B21 C22 A21B12 A22B22 Blocked Matrix Multiply bijk for jj 0 jj n jj bsize for i 0 i n i for j jj j min jj bsize n j c i j 0 0 for kk 0 kk n kk bsize for i 0 i n i for j jj j min jj bsize n j sum 0 0 for k kk k min kk bsize n k sum a i k b k j c i j sum Provides temporal locality as block is reused multiple times Constant cache performance Blocked Matrix Multiply Analysis Innermost loop pair multiplies a 1 X bsize sliver of A by a bsize X bsize block of B and accumulates into 1 X bsize sliver of C Loop over i steps through n row slivers of A C using same B for i 0 i n i for j jj j min jj bsize n j sum 0 0 for k kk k min kk bsize n k sum a i k b k j Innermost Loop Pair c i j sum kk jj jj i A kk i B C Update successive row sliver accessed elements of sliver bsize times block reused n times in succession Pentium Blocked Matrix Multiply Performance Blocking bijk and bikj improves performance by a factor of two over unblocked versions ijk and jik relatively insensitive to array size 60 kji jki kij ikj jik ijk bijk bsize 25 bikj bsize 25 40 30 20 10 75 10 0 12 5 15 0 17 5 20 0 22 5 25 0 27 5 30 0 32 5 35 0 37 5 40 0 50 0 25 Cycles iteration 50 Array size n Summary All systems favor cache friendly code Getting absolute optimum performance is very platform specific Cache sizes line sizes associativities etc Can get most of the advantage with generic code Keep working set reasonably small temporal locality Use small strides spatial locality
View Full Document