Actual source code: baijfact11.c
petsc-3.10.5 2019-03-28
2: /*
3: Factorization code for BAIJ format.
4: */
5: #include <../src/mat/impls/baij/seq/baij.h>
6: #include <petsc/private/kernels/blockinvert.h>
8: /* ------------------------------------------------------------*/
9: /*
10: Version for when blocks are 4 by 4
11: */
12: PetscErrorCode MatLUFactorNumeric_SeqBAIJ_4_inplace(Mat C,Mat A,const MatFactorInfo *info)
13: {
14: Mat_SeqBAIJ *a = (Mat_SeqBAIJ*)A->data,*b = (Mat_SeqBAIJ*)C->data;
15: IS isrow = b->row,isicol = b->icol;
17: const PetscInt *r,*ic;
18: PetscInt i,j,n = a->mbs,*bi = b->i,*bj = b->j;
19: PetscInt *ajtmpold,*ajtmp,nz,row;
20: PetscInt *diag_offset = b->diag,idx,*ai=a->i,*aj=a->j,*pj;
21: MatScalar *pv,*v,*rtmp,*pc,*w,*x;
22: MatScalar p1,p2,p3,p4,m1,m2,m3,m4,m5,m6,m7,m8,m9,x1,x2,x3,x4;
23: MatScalar p5,p6,p7,p8,p9,x5,x6,x7,x8,x9,x10,x11,x12,x13,x14,x15,x16;
24: MatScalar p10,p11,p12,p13,p14,p15,p16,m10,m11,m12;
25: MatScalar m13,m14,m15,m16;
26: MatScalar *ba = b->a,*aa = a->a;
27: PetscBool pivotinblocks = b->pivotinblocks;
28: PetscReal shift = info->shiftamount;
29: PetscBool allowzeropivot,zeropivotdetected=PETSC_FALSE;
32: ISGetIndices(isrow,&r);
33: ISGetIndices(isicol,&ic);
34: PetscMalloc1(16*(n+1),&rtmp);
35: allowzeropivot = PetscNot(A->erroriffailure);
37: for (i=0; i<n; i++) {
38: nz = bi[i+1] - bi[i];
39: ajtmp = bj + bi[i];
40: for (j=0; j<nz; j++) {
41: x = rtmp+16*ajtmp[j];
42: x[0] = x[1] = x[2] = x[3] = x[4] = x[5] = x[6] = x[7] = x[8] = x[9] = 0.0;
43: x[10] = x[11] = x[12] = x[13] = x[14] = x[15] = 0.0;
44: }
45: /* load in initial (unfactored row) */
46: idx = r[i];
47: nz = ai[idx+1] - ai[idx];
48: ajtmpold = aj + ai[idx];
49: v = aa + 16*ai[idx];
50: for (j=0; j<nz; j++) {
51: x = rtmp+16*ic[ajtmpold[j]];
52: x[0] = v[0]; x[1] = v[1]; x[2] = v[2]; x[3] = v[3];
53: x[4] = v[4]; x[5] = v[5]; x[6] = v[6]; x[7] = v[7]; x[8] = v[8];
54: x[9] = v[9]; x[10] = v[10]; x[11] = v[11]; x[12] = v[12]; x[13] = v[13];
55: x[14] = v[14]; x[15] = v[15];
56: v += 16;
57: }
58: row = *ajtmp++;
59: while (row < i) {
60: pc = rtmp + 16*row;
61: p1 = pc[0]; p2 = pc[1]; p3 = pc[2]; p4 = pc[3];
62: p5 = pc[4]; p6 = pc[5]; p7 = pc[6]; p8 = pc[7]; p9 = pc[8];
63: p10 = pc[9]; p11 = pc[10]; p12 = pc[11]; p13 = pc[12]; p14 = pc[13];
64: p15 = pc[14]; p16 = pc[15];
65: if (p1 != 0.0 || p2 != 0.0 || p3 != 0.0 || p4 != 0.0 || p5 != 0.0 ||
66: p6 != 0.0 || p7 != 0.0 || p8 != 0.0 || p9 != 0.0 || p10 != 0.0 ||
67: p11 != 0.0 || p12 != 0.0 || p13 != 0.0 || p14 != 0.0 || p15 != 0.0
68: || p16 != 0.0) {
69: pv = ba + 16*diag_offset[row];
70: pj = bj + diag_offset[row] + 1;
71: x1 = pv[0]; x2 = pv[1]; x3 = pv[2]; x4 = pv[3];
72: x5 = pv[4]; x6 = pv[5]; x7 = pv[6]; x8 = pv[7]; x9 = pv[8];
73: x10 = pv[9]; x11 = pv[10]; x12 = pv[11]; x13 = pv[12]; x14 = pv[13];
74: x15 = pv[14]; x16 = pv[15];
75: pc[0] = m1 = p1*x1 + p5*x2 + p9*x3 + p13*x4;
76: pc[1] = m2 = p2*x1 + p6*x2 + p10*x3 + p14*x4;
77: pc[2] = m3 = p3*x1 + p7*x2 + p11*x3 + p15*x4;
78: pc[3] = m4 = p4*x1 + p8*x2 + p12*x3 + p16*x4;
80: pc[4] = m5 = p1*x5 + p5*x6 + p9*x7 + p13*x8;
81: pc[5] = m6 = p2*x5 + p6*x6 + p10*x7 + p14*x8;
82: pc[6] = m7 = p3*x5 + p7*x6 + p11*x7 + p15*x8;
83: pc[7] = m8 = p4*x5 + p8*x6 + p12*x7 + p16*x8;
85: pc[8] = m9 = p1*x9 + p5*x10 + p9*x11 + p13*x12;
86: pc[9] = m10 = p2*x9 + p6*x10 + p10*x11 + p14*x12;
87: pc[10] = m11 = p3*x9 + p7*x10 + p11*x11 + p15*x12;
88: pc[11] = m12 = p4*x9 + p8*x10 + p12*x11 + p16*x12;
90: pc[12] = m13 = p1*x13 + p5*x14 + p9*x15 + p13*x16;
91: pc[13] = m14 = p2*x13 + p6*x14 + p10*x15 + p14*x16;
92: pc[14] = m15 = p3*x13 + p7*x14 + p11*x15 + p15*x16;
93: pc[15] = m16 = p4*x13 + p8*x14 + p12*x15 + p16*x16;
95: nz = bi[row+1] - diag_offset[row] - 1;
96: pv += 16;
97: for (j=0; j<nz; j++) {
98: x1 = pv[0]; x2 = pv[1]; x3 = pv[2]; x4 = pv[3];
99: x5 = pv[4]; x6 = pv[5]; x7 = pv[6]; x8 = pv[7]; x9 = pv[8];
100: x10 = pv[9]; x11 = pv[10]; x12 = pv[11]; x13 = pv[12];
101: x14 = pv[13]; x15 = pv[14]; x16 = pv[15];
102: x = rtmp + 16*pj[j];
103: x[0] -= m1*x1 + m5*x2 + m9*x3 + m13*x4;
104: x[1] -= m2*x1 + m6*x2 + m10*x3 + m14*x4;
105: x[2] -= m3*x1 + m7*x2 + m11*x3 + m15*x4;
106: x[3] -= m4*x1 + m8*x2 + m12*x3 + m16*x4;
108: x[4] -= m1*x5 + m5*x6 + m9*x7 + m13*x8;
109: x[5] -= m2*x5 + m6*x6 + m10*x7 + m14*x8;
110: x[6] -= m3*x5 + m7*x6 + m11*x7 + m15*x8;
111: x[7] -= m4*x5 + m8*x6 + m12*x7 + m16*x8;
113: x[8] -= m1*x9 + m5*x10 + m9*x11 + m13*x12;
114: x[9] -= m2*x9 + m6*x10 + m10*x11 + m14*x12;
115: x[10] -= m3*x9 + m7*x10 + m11*x11 + m15*x12;
116: x[11] -= m4*x9 + m8*x10 + m12*x11 + m16*x12;
118: x[12] -= m1*x13 + m5*x14 + m9*x15 + m13*x16;
119: x[13] -= m2*x13 + m6*x14 + m10*x15 + m14*x16;
120: x[14] -= m3*x13 + m7*x14 + m11*x15 + m15*x16;
121: x[15] -= m4*x13 + m8*x14 + m12*x15 + m16*x16;
123: pv += 16;
124: }
125: PetscLogFlops(128.0*nz+112.0);
126: }
127: row = *ajtmp++;
128: }
129: /* finished row so stick it into b->a */
130: pv = ba + 16*bi[i];
131: pj = bj + bi[i];
132: nz = bi[i+1] - bi[i];
133: for (j=0; j<nz; j++) {
134: x = rtmp+16*pj[j];
135: pv[0] = x[0]; pv[1] = x[1]; pv[2] = x[2]; pv[3] = x[3];
136: pv[4] = x[4]; pv[5] = x[5]; pv[6] = x[6]; pv[7] = x[7]; pv[8] = x[8];
137: pv[9] = x[9]; pv[10] = x[10]; pv[11] = x[11]; pv[12] = x[12];
138: pv[13] = x[13]; pv[14] = x[14]; pv[15] = x[15];
139: pv += 16;
140: }
141: /* invert diagonal block */
142: w = ba + 16*diag_offset[i];
143: if (pivotinblocks) {
144: PetscKernel_A_gets_inverse_A_4(w,shift,allowzeropivot,&zeropivotdetected);
145: if (zeropivotdetected) C->factorerrortype = MAT_FACTOR_NUMERIC_ZEROPIVOT;
146: } else {
147: PetscKernel_A_gets_inverse_A_4_nopivot(w);
148: }
149: }
151: PetscFree(rtmp);
152: ISRestoreIndices(isicol,&ic);
153: ISRestoreIndices(isrow,&r);
155: C->ops->solve = MatSolve_SeqBAIJ_4_inplace;
156: C->ops->solvetranspose = MatSolveTranspose_SeqBAIJ_4_inplace;
157: C->assembled = PETSC_TRUE;
159: PetscLogFlops(1.333333333333*4*4*4*b->mbs); /* from inverting diagonal blocks */
160: return(0);
161: }
163: /* MatLUFactorNumeric_SeqBAIJ_4 -
164: copied from MatLUFactorNumeric_SeqBAIJ_N_inplace() and manually re-implemented
165: PetscKernel_A_gets_A_times_B()
166: PetscKernel_A_gets_A_minus_B_times_C()
167: PetscKernel_A_gets_inverse_A()
168: */
170: PetscErrorCode MatLUFactorNumeric_SeqBAIJ_4(Mat B,Mat A,const MatFactorInfo *info)
171: {
172: Mat C = B;
173: Mat_SeqBAIJ *a = (Mat_SeqBAIJ*)A->data,*b=(Mat_SeqBAIJ*)C->data;
174: IS isrow = b->row,isicol = b->icol;
176: const PetscInt *r,*ic;
177: PetscInt i,j,k,nz,nzL,row;
178: const PetscInt n=a->mbs,*ai=a->i,*aj=a->j,*bi=b->i,*bj=b->j;
179: const PetscInt *ajtmp,*bjtmp,*bdiag=b->diag,*pj,bs2=a->bs2;
180: MatScalar *rtmp,*pc,*mwork,*v,*pv,*aa=a->a;
181: PetscInt flg;
182: PetscReal shift;
183: PetscBool allowzeropivot,zeropivotdetected;
186: allowzeropivot = PetscNot(A->erroriffailure);
187: ISGetIndices(isrow,&r);
188: ISGetIndices(isicol,&ic);
190: if (info->shifttype == MAT_SHIFT_NONE) {
191: shift = 0;
192: } else { /* info->shifttype == MAT_SHIFT_INBLOCKS */
193: shift = info->shiftamount;
194: }
196: /* generate work space needed by the factorization */
197: PetscMalloc2(bs2*n,&rtmp,bs2,&mwork);
198: PetscMemzero(rtmp,bs2*n*sizeof(MatScalar));
200: for (i=0; i<n; i++) {
201: /* zero rtmp */
202: /* L part */
203: nz = bi[i+1] - bi[i];
204: bjtmp = bj + bi[i];
205: for (j=0; j<nz; j++) {
206: PetscMemzero(rtmp+bs2*bjtmp[j],bs2*sizeof(MatScalar));
207: }
209: /* U part */
210: nz = bdiag[i]-bdiag[i+1];
211: bjtmp = bj + bdiag[i+1]+1;
212: for (j=0; j<nz; j++) {
213: PetscMemzero(rtmp+bs2*bjtmp[j],bs2*sizeof(MatScalar));
214: }
216: /* load in initial (unfactored row) */
217: nz = ai[r[i]+1] - ai[r[i]];
218: ajtmp = aj + ai[r[i]];
219: v = aa + bs2*ai[r[i]];
220: for (j=0; j<nz; j++) {
221: PetscMemcpy(rtmp+bs2*ic[ajtmp[j]],v+bs2*j,bs2*sizeof(MatScalar));
222: }
224: /* elimination */
225: bjtmp = bj + bi[i];
226: nzL = bi[i+1] - bi[i];
227: for (k=0; k < nzL; k++) {
228: row = bjtmp[k];
229: pc = rtmp + bs2*row;
230: for (flg=0,j=0; j<bs2; j++) {
231: if (pc[j]!=0.0) {
232: flg = 1;
233: break;
234: }
235: }
236: if (flg) {
237: pv = b->a + bs2*bdiag[row];
238: /* PetscKernel_A_gets_A_times_B(bs,pc,pv,mwork); *pc = *pc * (*pv); */
239: PetscKernel_A_gets_A_times_B_4(pc,pv,mwork);
241: pj = b->j + bdiag[row+1]+1; /* begining of U(row,:) */
242: pv = b->a + bs2*(bdiag[row+1]+1);
243: nz = bdiag[row] - bdiag[row+1] - 1; /* num of entries inU(row,:), excluding diag */
244: for (j=0; j<nz; j++) {
245: /* PetscKernel_A_gets_A_minus_B_times_C(bs,rtmp+bs2*pj[j],pc,pv+bs2*j); */
246: /* rtmp+bs2*pj[j] = rtmp+bs2*pj[j] - (*pc)*(pv+bs2*j) */
247: v = rtmp + bs2*pj[j];
248: PetscKernel_A_gets_A_minus_B_times_C_4(v,pc,pv);
249: pv += bs2;
250: }
251: PetscLogFlops(128*nz+112); /* flops = 2*bs^3*nz + 2*bs^3 - bs2) */
252: }
253: }
255: /* finished row so stick it into b->a */
256: /* L part */
257: pv = b->a + bs2*bi[i];
258: pj = b->j + bi[i];
259: nz = bi[i+1] - bi[i];
260: for (j=0; j<nz; j++) {
261: PetscMemcpy(pv+bs2*j,rtmp+bs2*pj[j],bs2*sizeof(MatScalar));
262: }
264: /* Mark diagonal and invert diagonal for simplier triangular solves */
265: pv = b->a + bs2*bdiag[i];
266: pj = b->j + bdiag[i];
267: PetscMemcpy(pv,rtmp+bs2*pj[0],bs2*sizeof(MatScalar));
268: PetscKernel_A_gets_inverse_A_4(pv,shift,allowzeropivot,&zeropivotdetected);
269: if (zeropivotdetected) C->factorerrortype = MAT_FACTOR_NUMERIC_ZEROPIVOT;
271: /* U part */
272: pv = b->a + bs2*(bdiag[i+1]+1);
273: pj = b->j + bdiag[i+1]+1;
274: nz = bdiag[i] - bdiag[i+1] - 1;
275: for (j=0; j<nz; j++) {
276: PetscMemcpy(pv+bs2*j,rtmp+bs2*pj[j],bs2*sizeof(MatScalar));
277: }
278: }
280: PetscFree2(rtmp,mwork);
281: ISRestoreIndices(isicol,&ic);
282: ISRestoreIndices(isrow,&r);
284: C->ops->solve = MatSolve_SeqBAIJ_4;
285: C->ops->solvetranspose = MatSolveTranspose_SeqBAIJ_4;
286: C->assembled = PETSC_TRUE;
288: PetscLogFlops(1.333333333333*4*4*4*n); /* from inverting diagonal blocks */
289: return(0);
290: }
292: PetscErrorCode MatLUFactorNumeric_SeqBAIJ_4_NaturalOrdering_inplace(Mat C,Mat A,const MatFactorInfo *info)
293: {
294: /*
295: Default Version for when blocks are 4 by 4 Using natural ordering
296: */
297: Mat_SeqBAIJ *a = (Mat_SeqBAIJ*)A->data,*b = (Mat_SeqBAIJ*)C->data;
299: PetscInt i,j,n = a->mbs,*bi = b->i,*bj = b->j;
300: PetscInt *ajtmpold,*ajtmp,nz,row;
301: PetscInt *diag_offset = b->diag,*ai=a->i,*aj=a->j,*pj;
302: MatScalar *pv,*v,*rtmp,*pc,*w,*x;
303: MatScalar p1,p2,p3,p4,m1,m2,m3,m4,m5,m6,m7,m8,m9,x1,x2,x3,x4;
304: MatScalar p5,p6,p7,p8,p9,x5,x6,x7,x8,x9,x10,x11,x12,x13,x14,x15,x16;
305: MatScalar p10,p11,p12,p13,p14,p15,p16,m10,m11,m12;
306: MatScalar m13,m14,m15,m16;
307: MatScalar *ba = b->a,*aa = a->a;
308: PetscBool pivotinblocks = b->pivotinblocks;
309: PetscReal shift = info->shiftamount;
310: PetscBool allowzeropivot,zeropivotdetected=PETSC_FALSE;
313: allowzeropivot = PetscNot(A->erroriffailure);
314: PetscMalloc1(16*(n+1),&rtmp);
316: for (i=0; i<n; i++) {
317: nz = bi[i+1] - bi[i];
318: ajtmp = bj + bi[i];
319: for (j=0; j<nz; j++) {
320: x = rtmp+16*ajtmp[j];
321: x[0] = x[1] = x[2] = x[3] = x[4] = x[5] = x[6] = x[7] = x[8] = x[9] = 0.0;
322: x[10] = x[11] = x[12] = x[13] = x[14] = x[15] = 0.0;
323: }
324: /* load in initial (unfactored row) */
325: nz = ai[i+1] - ai[i];
326: ajtmpold = aj + ai[i];
327: v = aa + 16*ai[i];
328: for (j=0; j<nz; j++) {
329: x = rtmp+16*ajtmpold[j];
330: x[0] = v[0]; x[1] = v[1]; x[2] = v[2]; x[3] = v[3];
331: x[4] = v[4]; x[5] = v[5]; x[6] = v[6]; x[7] = v[7]; x[8] = v[8];
332: x[9] = v[9]; x[10] = v[10]; x[11] = v[11]; x[12] = v[12]; x[13] = v[13];
333: x[14] = v[14]; x[15] = v[15];
334: v += 16;
335: }
336: row = *ajtmp++;
337: while (row < i) {
338: pc = rtmp + 16*row;
339: p1 = pc[0]; p2 = pc[1]; p3 = pc[2]; p4 = pc[3];
340: p5 = pc[4]; p6 = pc[5]; p7 = pc[6]; p8 = pc[7]; p9 = pc[8];
341: p10 = pc[9]; p11 = pc[10]; p12 = pc[11]; p13 = pc[12]; p14 = pc[13];
342: p15 = pc[14]; p16 = pc[15];
343: if (p1 != 0.0 || p2 != 0.0 || p3 != 0.0 || p4 != 0.0 || p5 != 0.0 ||
344: p6 != 0.0 || p7 != 0.0 || p8 != 0.0 || p9 != 0.0 || p10 != 0.0 ||
345: p11 != 0.0 || p12 != 0.0 || p13 != 0.0 || p14 != 0.0 || p15 != 0.0
346: || p16 != 0.0) {
347: pv = ba + 16*diag_offset[row];
348: pj = bj + diag_offset[row] + 1;
349: x1 = pv[0]; x2 = pv[1]; x3 = pv[2]; x4 = pv[3];
350: x5 = pv[4]; x6 = pv[5]; x7 = pv[6]; x8 = pv[7]; x9 = pv[8];
351: x10 = pv[9]; x11 = pv[10]; x12 = pv[11]; x13 = pv[12]; x14 = pv[13];
352: x15 = pv[14]; x16 = pv[15];
353: pc[0] = m1 = p1*x1 + p5*x2 + p9*x3 + p13*x4;
354: pc[1] = m2 = p2*x1 + p6*x2 + p10*x3 + p14*x4;
355: pc[2] = m3 = p3*x1 + p7*x2 + p11*x3 + p15*x4;
356: pc[3] = m4 = p4*x1 + p8*x2 + p12*x3 + p16*x4;
358: pc[4] = m5 = p1*x5 + p5*x6 + p9*x7 + p13*x8;
359: pc[5] = m6 = p2*x5 + p6*x6 + p10*x7 + p14*x8;
360: pc[6] = m7 = p3*x5 + p7*x6 + p11*x7 + p15*x8;
361: pc[7] = m8 = p4*x5 + p8*x6 + p12*x7 + p16*x8;
363: pc[8] = m9 = p1*x9 + p5*x10 + p9*x11 + p13*x12;
364: pc[9] = m10 = p2*x9 + p6*x10 + p10*x11 + p14*x12;
365: pc[10] = m11 = p3*x9 + p7*x10 + p11*x11 + p15*x12;
366: pc[11] = m12 = p4*x9 + p8*x10 + p12*x11 + p16*x12;
368: pc[12] = m13 = p1*x13 + p5*x14 + p9*x15 + p13*x16;
369: pc[13] = m14 = p2*x13 + p6*x14 + p10*x15 + p14*x16;
370: pc[14] = m15 = p3*x13 + p7*x14 + p11*x15 + p15*x16;
371: pc[15] = m16 = p4*x13 + p8*x14 + p12*x15 + p16*x16;
372: nz = bi[row+1] - diag_offset[row] - 1;
373: pv += 16;
374: for (j=0; j<nz; j++) {
375: x1 = pv[0]; x2 = pv[1]; x3 = pv[2]; x4 = pv[3];
376: x5 = pv[4]; x6 = pv[5]; x7 = pv[6]; x8 = pv[7]; x9 = pv[8];
377: x10 = pv[9]; x11 = pv[10]; x12 = pv[11]; x13 = pv[12];
378: x14 = pv[13]; x15 = pv[14]; x16 = pv[15];
379: x = rtmp + 16*pj[j];
380: x[0] -= m1*x1 + m5*x2 + m9*x3 + m13*x4;
381: x[1] -= m2*x1 + m6*x2 + m10*x3 + m14*x4;
382: x[2] -= m3*x1 + m7*x2 + m11*x3 + m15*x4;
383: x[3] -= m4*x1 + m8*x2 + m12*x3 + m16*x4;
385: x[4] -= m1*x5 + m5*x6 + m9*x7 + m13*x8;
386: x[5] -= m2*x5 + m6*x6 + m10*x7 + m14*x8;
387: x[6] -= m3*x5 + m7*x6 + m11*x7 + m15*x8;
388: x[7] -= m4*x5 + m8*x6 + m12*x7 + m16*x8;
390: x[8] -= m1*x9 + m5*x10 + m9*x11 + m13*x12;
391: x[9] -= m2*x9 + m6*x10 + m10*x11 + m14*x12;
392: x[10] -= m3*x9 + m7*x10 + m11*x11 + m15*x12;
393: x[11] -= m4*x9 + m8*x10 + m12*x11 + m16*x12;
395: x[12] -= m1*x13 + m5*x14 + m9*x15 + m13*x16;
396: x[13] -= m2*x13 + m6*x14 + m10*x15 + m14*x16;
397: x[14] -= m3*x13 + m7*x14 + m11*x15 + m15*x16;
398: x[15] -= m4*x13 + m8*x14 + m12*x15 + m16*x16;
400: pv += 16;
401: }
402: PetscLogFlops(128.0*nz+112.0);
403: }
404: row = *ajtmp++;
405: }
406: /* finished row so stick it into b->a */
407: pv = ba + 16*bi[i];
408: pj = bj + bi[i];
409: nz = bi[i+1] - bi[i];
410: for (j=0; j<nz; j++) {
411: x = rtmp+16*pj[j];
412: pv[0] = x[0]; pv[1] = x[1]; pv[2] = x[2]; pv[3] = x[3];
413: pv[4] = x[4]; pv[5] = x[5]; pv[6] = x[6]; pv[7] = x[7]; pv[8] = x[8];
414: pv[9] = x[9]; pv[10] = x[10]; pv[11] = x[11]; pv[12] = x[12];
415: pv[13] = x[13]; pv[14] = x[14]; pv[15] = x[15];
416: pv += 16;
417: }
418: /* invert diagonal block */
419: w = ba + 16*diag_offset[i];
420: if (pivotinblocks) {
421: PetscKernel_A_gets_inverse_A_4(w,shift,allowzeropivot,&zeropivotdetected);
422: if (zeropivotdetected) C->factorerrortype = MAT_FACTOR_NUMERIC_ZEROPIVOT;
423: } else {
424: PetscKernel_A_gets_inverse_A_4_nopivot(w);
425: }
426: }
428: PetscFree(rtmp);
430: C->ops->solve = MatSolve_SeqBAIJ_4_NaturalOrdering_inplace;
431: C->ops->solvetranspose = MatSolveTranspose_SeqBAIJ_4_NaturalOrdering_inplace;
432: C->assembled = PETSC_TRUE;
434: PetscLogFlops(1.333333333333*4*4*4*b->mbs); /* from inverting diagonal blocks */
435: return(0);
436: }
438: /*
439: MatLUFactorNumeric_SeqBAIJ_4_NaturalOrdering -
440: copied from MatLUFactorNumeric_SeqBAIJ_3_NaturalOrdering_inplace()
441: */
442: PetscErrorCode MatLUFactorNumeric_SeqBAIJ_4_NaturalOrdering(Mat B,Mat A,const MatFactorInfo *info)
443: {
444: Mat C =B;
445: Mat_SeqBAIJ *a=(Mat_SeqBAIJ*)A->data,*b=(Mat_SeqBAIJ*)C->data;
447: PetscInt i,j,k,nz,nzL,row;
448: const PetscInt n=a->mbs,*ai=a->i,*aj=a->j,*bi=b->i,*bj=b->j;
449: const PetscInt *ajtmp,*bjtmp,*bdiag=b->diag,*pj,bs2=a->bs2;
450: MatScalar *rtmp,*pc,*mwork,*v,*pv,*aa=a->a;
451: PetscInt flg;
452: PetscReal shift;
453: PetscBool allowzeropivot,zeropivotdetected;
456: allowzeropivot = PetscNot(A->erroriffailure);
458: /* generate work space needed by the factorization */
459: PetscMalloc2(bs2*n,&rtmp,bs2,&mwork);
460: PetscMemzero(rtmp,bs2*n*sizeof(MatScalar));
462: if (info->shifttype == MAT_SHIFT_NONE) {
463: shift = 0;
464: } else { /* info->shifttype == MAT_SHIFT_INBLOCKS */
465: shift = info->shiftamount;
466: }
468: for (i=0; i<n; i++) {
469: /* zero rtmp */
470: /* L part */
471: nz = bi[i+1] - bi[i];
472: bjtmp = bj + bi[i];
473: for (j=0; j<nz; j++) {
474: PetscMemzero(rtmp+bs2*bjtmp[j],bs2*sizeof(MatScalar));
475: }
477: /* U part */
478: nz = bdiag[i] - bdiag[i+1];
479: bjtmp = bj + bdiag[i+1]+1;
480: for (j=0; j<nz; j++) {
481: PetscMemzero(rtmp+bs2*bjtmp[j],bs2*sizeof(MatScalar));
482: }
484: /* load in initial (unfactored row) */
485: nz = ai[i+1] - ai[i];
486: ajtmp = aj + ai[i];
487: v = aa + bs2*ai[i];
488: for (j=0; j<nz; j++) {
489: PetscMemcpy(rtmp+bs2*ajtmp[j],v+bs2*j,bs2*sizeof(MatScalar));
490: }
492: /* elimination */
493: bjtmp = bj + bi[i];
494: nzL = bi[i+1] - bi[i];
495: for (k=0; k < nzL; k++) {
496: row = bjtmp[k];
497: pc = rtmp + bs2*row;
498: for (flg=0,j=0; j<bs2; j++) {
499: if (pc[j]!=0.0) {
500: flg = 1;
501: break;
502: }
503: }
504: if (flg) {
505: pv = b->a + bs2*bdiag[row];
506: /* PetscKernel_A_gets_A_times_B(bs,pc,pv,mwork); *pc = *pc * (*pv); */
507: PetscKernel_A_gets_A_times_B_4(pc,pv,mwork);
509: pj = b->j + bdiag[row+1]+1; /* begining of U(row,:) */
510: pv = b->a + bs2*(bdiag[row+1]+1);
511: nz = bdiag[row] - bdiag[row+1] - 1; /* num of entries inU(row,:), excluding diag */
512: for (j=0; j<nz; j++) {
513: /* PetscKernel_A_gets_A_minus_B_times_C(bs,rtmp+bs2*pj[j],pc,pv+bs2*j); */
514: /* rtmp+bs2*pj[j] = rtmp+bs2*pj[j] - (*pc)*(pv+bs2*j) */
515: v = rtmp + bs2*pj[j];
516: PetscKernel_A_gets_A_minus_B_times_C_4(v,pc,pv);
517: pv += bs2;
518: }
519: PetscLogFlops(128*nz+112); /* flops = 2*bs^3*nz + 2*bs^3 - bs2) */
520: }
521: }
523: /* finished row so stick it into b->a */
524: /* L part */
525: pv = b->a + bs2*bi[i];
526: pj = b->j + bi[i];
527: nz = bi[i+1] - bi[i];
528: for (j=0; j<nz; j++) {
529: PetscMemcpy(pv+bs2*j,rtmp+bs2*pj[j],bs2*sizeof(MatScalar));
530: }
532: /* Mark diagonal and invert diagonal for simplier triangular solves */
533: pv = b->a + bs2*bdiag[i];
534: pj = b->j + bdiag[i];
535: PetscMemcpy(pv,rtmp+bs2*pj[0],bs2*sizeof(MatScalar));
536: PetscKernel_A_gets_inverse_A_4(pv,shift,allowzeropivot,&zeropivotdetected);
537: if (zeropivotdetected) C->factorerrortype = MAT_FACTOR_NUMERIC_ZEROPIVOT;
539: /* U part */
540: pv = b->a + bs2*(bdiag[i+1]+1);
541: pj = b->j + bdiag[i+1]+1;
542: nz = bdiag[i] - bdiag[i+1] - 1;
543: for (j=0; j<nz; j++) {
544: PetscMemcpy(pv+bs2*j,rtmp+bs2*pj[j],bs2*sizeof(MatScalar));
545: }
546: }
547: PetscFree2(rtmp,mwork);
549: C->ops->solve = MatSolve_SeqBAIJ_4_NaturalOrdering;
550: C->ops->solvetranspose = MatSolveTranspose_SeqBAIJ_4_NaturalOrdering;
551: C->assembled = PETSC_TRUE;
553: PetscLogFlops(1.333333333333*4*4*4*n); /* from inverting diagonal blocks */
554: return(0);
555: }
557: #if defined(PETSC_HAVE_SSE)
559: #include PETSC_HAVE_SSE
561: /* SSE Version for when blocks are 4 by 4 Using natural ordering */
562: PetscErrorCode MatLUFactorNumeric_SeqBAIJ_4_NaturalOrdering_SSE(Mat B,Mat A,const MatFactorInfo *info)
563: {
564: Mat_SeqBAIJ *a = (Mat_SeqBAIJ*)A->data,*b = (Mat_SeqBAIJ*)C->data;
566: int i,j,n = a->mbs;
567: int *bj = b->j,*bjtmp,*pj;
568: int row;
569: int *ajtmpold,nz,*bi=b->i;
570: int *diag_offset = b->diag,*ai=a->i,*aj=a->j;
571: MatScalar *pv,*v,*rtmp,*pc,*w,*x;
572: MatScalar *ba = b->a,*aa = a->a;
573: int nonzero=0;
574: PetscBool pivotinblocks = b->pivotinblocks;
575: PetscReal shift = info->shiftamount;
576: PetscBool allowzeropivot,zeropivotdetected=PETSC_FALSE;
579: allowzeropivot = PetscNot(A->erroriffailure);
580: SSE_SCOPE_BEGIN;
582: if ((unsigned long)aa%16!=0) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_BADPTR,"Pointer aa is not 16 byte aligned. SSE will not work.");
583: if ((unsigned long)ba%16!=0) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_BADPTR,"Pointer ba is not 16 byte aligned. SSE will not work.");
584: PetscMalloc1(16*(n+1),&rtmp);
585: if ((unsigned long)rtmp%16!=0) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_BADPTR,"Pointer rtmp is not 16 byte aligned. SSE will not work.");
586: /* if ((unsigned long)bj==(unsigned long)aj) { */
587: /* colscale = 4; */
588: /* } */
589: for (i=0; i<n; i++) {
590: nz = bi[i+1] - bi[i];
591: bjtmp = bj + bi[i];
592: /* zero out the 4x4 block accumulators */
593: /* zero out one register */
594: XOR_PS(XMM7,XMM7);
595: for (j=0; j<nz; j++) {
596: x = rtmp+16*bjtmp[j];
597: /* x = rtmp+4*bjtmp[j]; */
598: SSE_INLINE_BEGIN_1(x)
599: /* Copy zero register to memory locations */
600: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
601: SSE_STOREL_PS(SSE_ARG_1,FLOAT_0,XMM7)
602: SSE_STOREH_PS(SSE_ARG_1,FLOAT_2,XMM7)
603: SSE_STOREL_PS(SSE_ARG_1,FLOAT_4,XMM7)
604: SSE_STOREH_PS(SSE_ARG_1,FLOAT_6,XMM7)
605: SSE_STOREL_PS(SSE_ARG_1,FLOAT_8,XMM7)
606: SSE_STOREH_PS(SSE_ARG_1,FLOAT_10,XMM7)
607: SSE_STOREL_PS(SSE_ARG_1,FLOAT_12,XMM7)
608: SSE_STOREH_PS(SSE_ARG_1,FLOAT_14,XMM7)
609: SSE_INLINE_END_1;
610: }
611: /* load in initial (unfactored row) */
612: nz = ai[i+1] - ai[i];
613: ajtmpold = aj + ai[i];
614: v = aa + 16*ai[i];
615: for (j=0; j<nz; j++) {
616: x = rtmp+16*ajtmpold[j];
617: /* x = rtmp+colscale*ajtmpold[j]; */
618: /* Copy v block into x block */
619: SSE_INLINE_BEGIN_2(v,x)
620: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
621: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM0)
622: SSE_STOREL_PS(SSE_ARG_2,FLOAT_0,XMM0)
624: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM1)
625: SSE_STOREH_PS(SSE_ARG_2,FLOAT_2,XMM1)
627: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM2)
628: SSE_STOREL_PS(SSE_ARG_2,FLOAT_4,XMM2)
630: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM3)
631: SSE_STOREH_PS(SSE_ARG_2,FLOAT_6,XMM3)
633: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM4)
634: SSE_STOREL_PS(SSE_ARG_2,FLOAT_8,XMM4)
636: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM5)
637: SSE_STOREH_PS(SSE_ARG_2,FLOAT_10,XMM5)
639: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM6)
640: SSE_STOREL_PS(SSE_ARG_2,FLOAT_12,XMM6)
642: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM0)
643: SSE_STOREH_PS(SSE_ARG_2,FLOAT_14,XMM0)
644: SSE_INLINE_END_2;
646: v += 16;
647: }
648: /* row = (*bjtmp++)/4; */
649: row = *bjtmp++;
650: while (row < i) {
651: pc = rtmp + 16*row;
652: SSE_INLINE_BEGIN_1(pc)
653: /* Load block from lower triangle */
654: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
655: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM0)
656: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM0)
658: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM1)
659: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM1)
661: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM2)
662: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM2)
664: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM3)
665: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM3)
667: /* Compare block to zero block */
669: SSE_COPY_PS(XMM4,XMM7)
670: SSE_CMPNEQ_PS(XMM4,XMM0)
672: SSE_COPY_PS(XMM5,XMM7)
673: SSE_CMPNEQ_PS(XMM5,XMM1)
675: SSE_COPY_PS(XMM6,XMM7)
676: SSE_CMPNEQ_PS(XMM6,XMM2)
678: SSE_CMPNEQ_PS(XMM7,XMM3)
680: /* Reduce the comparisons to one SSE register */
681: SSE_OR_PS(XMM6,XMM7)
682: SSE_OR_PS(XMM5,XMM4)
683: SSE_OR_PS(XMM5,XMM6)
684: SSE_INLINE_END_1;
686: /* Reduce the one SSE register to an integer register for branching */
687: /* Note: Since nonzero is an int, there is no INLINE block version of this call */
688: MOVEMASK(nonzero,XMM5);
690: /* If block is nonzero ... */
691: if (nonzero) {
692: pv = ba + 16*diag_offset[row];
693: PREFETCH_L1(&pv[16]);
694: pj = bj + diag_offset[row] + 1;
696: /* Form Multiplier, one column at a time (Matrix-Matrix Product) */
697: /* L_ij^(k+1) = L_ij^(k)*inv(L_jj^(k)) */
698: /* but the diagonal was inverted already */
699: /* and, L_ij^(k) is already loaded into registers XMM0-XMM3 columnwise */
701: SSE_INLINE_BEGIN_2(pv,pc)
702: /* Column 0, product is accumulated in XMM4 */
703: SSE_LOAD_SS(SSE_ARG_1,FLOAT_0,XMM4)
704: SSE_SHUFFLE(XMM4,XMM4,0x00)
705: SSE_MULT_PS(XMM4,XMM0)
707: SSE_LOAD_SS(SSE_ARG_1,FLOAT_1,XMM5)
708: SSE_SHUFFLE(XMM5,XMM5,0x00)
709: SSE_MULT_PS(XMM5,XMM1)
710: SSE_ADD_PS(XMM4,XMM5)
712: SSE_LOAD_SS(SSE_ARG_1,FLOAT_2,XMM6)
713: SSE_SHUFFLE(XMM6,XMM6,0x00)
714: SSE_MULT_PS(XMM6,XMM2)
715: SSE_ADD_PS(XMM4,XMM6)
717: SSE_LOAD_SS(SSE_ARG_1,FLOAT_3,XMM7)
718: SSE_SHUFFLE(XMM7,XMM7,0x00)
719: SSE_MULT_PS(XMM7,XMM3)
720: SSE_ADD_PS(XMM4,XMM7)
722: SSE_STOREL_PS(SSE_ARG_2,FLOAT_0,XMM4)
723: SSE_STOREH_PS(SSE_ARG_2,FLOAT_2,XMM4)
725: /* Column 1, product is accumulated in XMM5 */
726: SSE_LOAD_SS(SSE_ARG_1,FLOAT_4,XMM5)
727: SSE_SHUFFLE(XMM5,XMM5,0x00)
728: SSE_MULT_PS(XMM5,XMM0)
730: SSE_LOAD_SS(SSE_ARG_1,FLOAT_5,XMM6)
731: SSE_SHUFFLE(XMM6,XMM6,0x00)
732: SSE_MULT_PS(XMM6,XMM1)
733: SSE_ADD_PS(XMM5,XMM6)
735: SSE_LOAD_SS(SSE_ARG_1,FLOAT_6,XMM7)
736: SSE_SHUFFLE(XMM7,XMM7,0x00)
737: SSE_MULT_PS(XMM7,XMM2)
738: SSE_ADD_PS(XMM5,XMM7)
740: SSE_LOAD_SS(SSE_ARG_1,FLOAT_7,XMM6)
741: SSE_SHUFFLE(XMM6,XMM6,0x00)
742: SSE_MULT_PS(XMM6,XMM3)
743: SSE_ADD_PS(XMM5,XMM6)
745: SSE_STOREL_PS(SSE_ARG_2,FLOAT_4,XMM5)
746: SSE_STOREH_PS(SSE_ARG_2,FLOAT_6,XMM5)
748: SSE_PREFETCH_L1(SSE_ARG_1,FLOAT_24)
750: /* Column 2, product is accumulated in XMM6 */
751: SSE_LOAD_SS(SSE_ARG_1,FLOAT_8,XMM6)
752: SSE_SHUFFLE(XMM6,XMM6,0x00)
753: SSE_MULT_PS(XMM6,XMM0)
755: SSE_LOAD_SS(SSE_ARG_1,FLOAT_9,XMM7)
756: SSE_SHUFFLE(XMM7,XMM7,0x00)
757: SSE_MULT_PS(XMM7,XMM1)
758: SSE_ADD_PS(XMM6,XMM7)
760: SSE_LOAD_SS(SSE_ARG_1,FLOAT_10,XMM7)
761: SSE_SHUFFLE(XMM7,XMM7,0x00)
762: SSE_MULT_PS(XMM7,XMM2)
763: SSE_ADD_PS(XMM6,XMM7)
765: SSE_LOAD_SS(SSE_ARG_1,FLOAT_11,XMM7)
766: SSE_SHUFFLE(XMM7,XMM7,0x00)
767: SSE_MULT_PS(XMM7,XMM3)
768: SSE_ADD_PS(XMM6,XMM7)
770: SSE_STOREL_PS(SSE_ARG_2,FLOAT_8,XMM6)
771: SSE_STOREH_PS(SSE_ARG_2,FLOAT_10,XMM6)
773: /* Note: For the last column, we no longer need to preserve XMM0->XMM3 */
774: /* Column 3, product is accumulated in XMM0 */
775: SSE_LOAD_SS(SSE_ARG_1,FLOAT_12,XMM7)
776: SSE_SHUFFLE(XMM7,XMM7,0x00)
777: SSE_MULT_PS(XMM0,XMM7)
779: SSE_LOAD_SS(SSE_ARG_1,FLOAT_13,XMM7)
780: SSE_SHUFFLE(XMM7,XMM7,0x00)
781: SSE_MULT_PS(XMM1,XMM7)
782: SSE_ADD_PS(XMM0,XMM1)
784: SSE_LOAD_SS(SSE_ARG_1,FLOAT_14,XMM1)
785: SSE_SHUFFLE(XMM1,XMM1,0x00)
786: SSE_MULT_PS(XMM1,XMM2)
787: SSE_ADD_PS(XMM0,XMM1)
789: SSE_LOAD_SS(SSE_ARG_1,FLOAT_15,XMM7)
790: SSE_SHUFFLE(XMM7,XMM7,0x00)
791: SSE_MULT_PS(XMM3,XMM7)
792: SSE_ADD_PS(XMM0,XMM3)
794: SSE_STOREL_PS(SSE_ARG_2,FLOAT_12,XMM0)
795: SSE_STOREH_PS(SSE_ARG_2,FLOAT_14,XMM0)
797: /* Simplify Bookkeeping -- Completely Unnecessary Instructions */
798: /* This is code to be maintained and read by humans afterall. */
799: /* Copy Multiplier Col 3 into XMM3 */
800: SSE_COPY_PS(XMM3,XMM0)
801: /* Copy Multiplier Col 2 into XMM2 */
802: SSE_COPY_PS(XMM2,XMM6)
803: /* Copy Multiplier Col 1 into XMM1 */
804: SSE_COPY_PS(XMM1,XMM5)
805: /* Copy Multiplier Col 0 into XMM0 */
806: SSE_COPY_PS(XMM0,XMM4)
807: SSE_INLINE_END_2;
809: /* Update the row: */
810: nz = bi[row+1] - diag_offset[row] - 1;
811: pv += 16;
812: for (j=0; j<nz; j++) {
813: PREFETCH_L1(&pv[16]);
814: x = rtmp + 16*pj[j];
815: /* x = rtmp + 4*pj[j]; */
817: /* X:=X-M*PV, One column at a time */
818: /* Note: M is already loaded columnwise into registers XMM0-XMM3 */
819: SSE_INLINE_BEGIN_2(x,pv)
820: /* Load First Column of X*/
821: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM4)
822: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM4)
824: /* Matrix-Vector Product: */
825: SSE_LOAD_SS(SSE_ARG_2,FLOAT_0,XMM5)
826: SSE_SHUFFLE(XMM5,XMM5,0x00)
827: SSE_MULT_PS(XMM5,XMM0)
828: SSE_SUB_PS(XMM4,XMM5)
830: SSE_LOAD_SS(SSE_ARG_2,FLOAT_1,XMM6)
831: SSE_SHUFFLE(XMM6,XMM6,0x00)
832: SSE_MULT_PS(XMM6,XMM1)
833: SSE_SUB_PS(XMM4,XMM6)
835: SSE_LOAD_SS(SSE_ARG_2,FLOAT_2,XMM7)
836: SSE_SHUFFLE(XMM7,XMM7,0x00)
837: SSE_MULT_PS(XMM7,XMM2)
838: SSE_SUB_PS(XMM4,XMM7)
840: SSE_LOAD_SS(SSE_ARG_2,FLOAT_3,XMM5)
841: SSE_SHUFFLE(XMM5,XMM5,0x00)
842: SSE_MULT_PS(XMM5,XMM3)
843: SSE_SUB_PS(XMM4,XMM5)
845: SSE_STOREL_PS(SSE_ARG_1,FLOAT_0,XMM4)
846: SSE_STOREH_PS(SSE_ARG_1,FLOAT_2,XMM4)
848: /* Second Column */
849: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM5)
850: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM5)
852: /* Matrix-Vector Product: */
853: SSE_LOAD_SS(SSE_ARG_2,FLOAT_4,XMM6)
854: SSE_SHUFFLE(XMM6,XMM6,0x00)
855: SSE_MULT_PS(XMM6,XMM0)
856: SSE_SUB_PS(XMM5,XMM6)
858: SSE_LOAD_SS(SSE_ARG_2,FLOAT_5,XMM7)
859: SSE_SHUFFLE(XMM7,XMM7,0x00)
860: SSE_MULT_PS(XMM7,XMM1)
861: SSE_SUB_PS(XMM5,XMM7)
863: SSE_LOAD_SS(SSE_ARG_2,FLOAT_6,XMM4)
864: SSE_SHUFFLE(XMM4,XMM4,0x00)
865: SSE_MULT_PS(XMM4,XMM2)
866: SSE_SUB_PS(XMM5,XMM4)
868: SSE_LOAD_SS(SSE_ARG_2,FLOAT_7,XMM6)
869: SSE_SHUFFLE(XMM6,XMM6,0x00)
870: SSE_MULT_PS(XMM6,XMM3)
871: SSE_SUB_PS(XMM5,XMM6)
873: SSE_STOREL_PS(SSE_ARG_1,FLOAT_4,XMM5)
874: SSE_STOREH_PS(SSE_ARG_1,FLOAT_6,XMM5)
876: SSE_PREFETCH_L1(SSE_ARG_2,FLOAT_24)
878: /* Third Column */
879: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM6)
880: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM6)
882: /* Matrix-Vector Product: */
883: SSE_LOAD_SS(SSE_ARG_2,FLOAT_8,XMM7)
884: SSE_SHUFFLE(XMM7,XMM7,0x00)
885: SSE_MULT_PS(XMM7,XMM0)
886: SSE_SUB_PS(XMM6,XMM7)
888: SSE_LOAD_SS(SSE_ARG_2,FLOAT_9,XMM4)
889: SSE_SHUFFLE(XMM4,XMM4,0x00)
890: SSE_MULT_PS(XMM4,XMM1)
891: SSE_SUB_PS(XMM6,XMM4)
893: SSE_LOAD_SS(SSE_ARG_2,FLOAT_10,XMM5)
894: SSE_SHUFFLE(XMM5,XMM5,0x00)
895: SSE_MULT_PS(XMM5,XMM2)
896: SSE_SUB_PS(XMM6,XMM5)
898: SSE_LOAD_SS(SSE_ARG_2,FLOAT_11,XMM7)
899: SSE_SHUFFLE(XMM7,XMM7,0x00)
900: SSE_MULT_PS(XMM7,XMM3)
901: SSE_SUB_PS(XMM6,XMM7)
903: SSE_STOREL_PS(SSE_ARG_1,FLOAT_8,XMM6)
904: SSE_STOREH_PS(SSE_ARG_1,FLOAT_10,XMM6)
906: /* Fourth Column */
907: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM4)
908: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM4)
910: /* Matrix-Vector Product: */
911: SSE_LOAD_SS(SSE_ARG_2,FLOAT_12,XMM5)
912: SSE_SHUFFLE(XMM5,XMM5,0x00)
913: SSE_MULT_PS(XMM5,XMM0)
914: SSE_SUB_PS(XMM4,XMM5)
916: SSE_LOAD_SS(SSE_ARG_2,FLOAT_13,XMM6)
917: SSE_SHUFFLE(XMM6,XMM6,0x00)
918: SSE_MULT_PS(XMM6,XMM1)
919: SSE_SUB_PS(XMM4,XMM6)
921: SSE_LOAD_SS(SSE_ARG_2,FLOAT_14,XMM7)
922: SSE_SHUFFLE(XMM7,XMM7,0x00)
923: SSE_MULT_PS(XMM7,XMM2)
924: SSE_SUB_PS(XMM4,XMM7)
926: SSE_LOAD_SS(SSE_ARG_2,FLOAT_15,XMM5)
927: SSE_SHUFFLE(XMM5,XMM5,0x00)
928: SSE_MULT_PS(XMM5,XMM3)
929: SSE_SUB_PS(XMM4,XMM5)
931: SSE_STOREL_PS(SSE_ARG_1,FLOAT_12,XMM4)
932: SSE_STOREH_PS(SSE_ARG_1,FLOAT_14,XMM4)
933: SSE_INLINE_END_2;
934: pv += 16;
935: }
936: PetscLogFlops(128.0*nz+112.0);
937: }
938: row = *bjtmp++;
939: /* row = (*bjtmp++)/4; */
940: }
941: /* finished row so stick it into b->a */
942: pv = ba + 16*bi[i];
943: pj = bj + bi[i];
944: nz = bi[i+1] - bi[i];
946: /* Copy x block back into pv block */
947: for (j=0; j<nz; j++) {
948: x = rtmp+16*pj[j];
949: /* x = rtmp+4*pj[j]; */
951: SSE_INLINE_BEGIN_2(x,pv)
952: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
953: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM1)
954: SSE_STOREL_PS(SSE_ARG_2,FLOAT_0,XMM1)
956: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM2)
957: SSE_STOREH_PS(SSE_ARG_2,FLOAT_2,XMM2)
959: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM3)
960: SSE_STOREL_PS(SSE_ARG_2,FLOAT_4,XMM3)
962: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM4)
963: SSE_STOREH_PS(SSE_ARG_2,FLOAT_6,XMM4)
965: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM5)
966: SSE_STOREL_PS(SSE_ARG_2,FLOAT_8,XMM5)
968: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM6)
969: SSE_STOREH_PS(SSE_ARG_2,FLOAT_10,XMM6)
971: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM7)
972: SSE_STOREL_PS(SSE_ARG_2,FLOAT_12,XMM7)
974: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM0)
975: SSE_STOREH_PS(SSE_ARG_2,FLOAT_14,XMM0)
976: SSE_INLINE_END_2;
977: pv += 16;
978: }
979: /* invert diagonal block */
980: w = ba + 16*diag_offset[i];
981: if (pivotinblocks) {
982: PetscKernel_A_gets_inverse_A_4(w,shift,allowzeropivot,&zeropivotdetected);
983: if (zeropivotdetected) C->factorerrortype = MAT_FACTOR_NUMERIC_ZEROPIVOT;
984: } else {
985: PetscKernel_A_gets_inverse_A_4_nopivot(w);
986: }
987: /* PetscKernel_A_gets_inverse_A_4_SSE(w); */
988: /* Note: Using Kramer's rule, flop count below might be infairly high or low? */
989: }
991: PetscFree(rtmp);
993: C->ops->solve = MatSolve_SeqBAIJ_4_NaturalOrdering_SSE;
994: C->ops->solvetranspose = MatSolveTranspose_SeqBAIJ_4_NaturalOrdering_SSE;
995: C->assembled = PETSC_TRUE;
997: PetscLogFlops(1.333333333333*bs*bs2*b->mbs);
998: /* Flop Count from inverting diagonal blocks */
999: SSE_SCOPE_END;
1000: return(0);
1001: }
1003: PetscErrorCode MatLUFactorNumeric_SeqBAIJ_4_NaturalOrdering_SSE_usj_Inplace(Mat C)
1004: {
1005: Mat A =C;
1006: Mat_SeqBAIJ *a = (Mat_SeqBAIJ*)A->data,*b = (Mat_SeqBAIJ*)C->data;
1008: int i,j,n = a->mbs;
1009: unsigned short *bj = (unsigned short*)(b->j),*bjtmp,*pj;
1010: unsigned short *aj = (unsigned short*)(a->j),*ajtmp;
1011: unsigned int row;
1012: int nz,*bi=b->i;
1013: int *diag_offset = b->diag,*ai=a->i;
1014: MatScalar *pv,*v,*rtmp,*pc,*w,*x;
1015: MatScalar *ba = b->a,*aa = a->a;
1016: int nonzero=0;
1017: PetscBool pivotinblocks = b->pivotinblocks;
1018: PetscReal shift = info->shiftamount;
1019: PetscBool allowzeropivot,zeropivotdetected=PETSC_FALSE;
1022: allowzeropivot = PetscNot(A->erroriffailure);
1023: SSE_SCOPE_BEGIN;
1025: if ((unsigned long)aa%16!=0) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_BADPTR,"Pointer aa is not 16 byte aligned. SSE will not work.");
1026: if ((unsigned long)ba%16!=0) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_BADPTR,"Pointer ba is not 16 byte aligned. SSE will not work.");
1027: PetscMalloc1(16*(n+1),&rtmp);
1028: if ((unsigned long)rtmp%16!=0) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_BADPTR,"Pointer rtmp is not 16 byte aligned. SSE will not work.");
1029: /* if ((unsigned long)bj==(unsigned long)aj) { */
1030: /* colscale = 4; */
1031: /* } */
1033: for (i=0; i<n; i++) {
1034: nz = bi[i+1] - bi[i];
1035: bjtmp = bj + bi[i];
1036: /* zero out the 4x4 block accumulators */
1037: /* zero out one register */
1038: XOR_PS(XMM7,XMM7);
1039: for (j=0; j<nz; j++) {
1040: x = rtmp+16*((unsigned int)bjtmp[j]);
1041: /* x = rtmp+4*bjtmp[j]; */
1042: SSE_INLINE_BEGIN_1(x)
1043: /* Copy zero register to memory locations */
1044: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
1045: SSE_STOREL_PS(SSE_ARG_1,FLOAT_0,XMM7)
1046: SSE_STOREH_PS(SSE_ARG_1,FLOAT_2,XMM7)
1047: SSE_STOREL_PS(SSE_ARG_1,FLOAT_4,XMM7)
1048: SSE_STOREH_PS(SSE_ARG_1,FLOAT_6,XMM7)
1049: SSE_STOREL_PS(SSE_ARG_1,FLOAT_8,XMM7)
1050: SSE_STOREH_PS(SSE_ARG_1,FLOAT_10,XMM7)
1051: SSE_STOREL_PS(SSE_ARG_1,FLOAT_12,XMM7)
1052: SSE_STOREH_PS(SSE_ARG_1,FLOAT_14,XMM7)
1053: SSE_INLINE_END_1;
1054: }
1055: /* load in initial (unfactored row) */
1056: nz = ai[i+1] - ai[i];
1057: ajtmp = aj + ai[i];
1058: v = aa + 16*ai[i];
1059: for (j=0; j<nz; j++) {
1060: x = rtmp+16*((unsigned int)ajtmp[j]);
1061: /* x = rtmp+colscale*ajtmp[j]; */
1062: /* Copy v block into x block */
1063: SSE_INLINE_BEGIN_2(v,x)
1064: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
1065: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM0)
1066: SSE_STOREL_PS(SSE_ARG_2,FLOAT_0,XMM0)
1068: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM1)
1069: SSE_STOREH_PS(SSE_ARG_2,FLOAT_2,XMM1)
1071: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM2)
1072: SSE_STOREL_PS(SSE_ARG_2,FLOAT_4,XMM2)
1074: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM3)
1075: SSE_STOREH_PS(SSE_ARG_2,FLOAT_6,XMM3)
1077: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM4)
1078: SSE_STOREL_PS(SSE_ARG_2,FLOAT_8,XMM4)
1080: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM5)
1081: SSE_STOREH_PS(SSE_ARG_2,FLOAT_10,XMM5)
1083: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM6)
1084: SSE_STOREL_PS(SSE_ARG_2,FLOAT_12,XMM6)
1086: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM0)
1087: SSE_STOREH_PS(SSE_ARG_2,FLOAT_14,XMM0)
1088: SSE_INLINE_END_2;
1090: v += 16;
1091: }
1092: /* row = (*bjtmp++)/4; */
1093: row = (unsigned int)(*bjtmp++);
1094: while (row < i) {
1095: pc = rtmp + 16*row;
1096: SSE_INLINE_BEGIN_1(pc)
1097: /* Load block from lower triangle */
1098: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
1099: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM0)
1100: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM0)
1102: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM1)
1103: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM1)
1105: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM2)
1106: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM2)
1108: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM3)
1109: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM3)
1111: /* Compare block to zero block */
1113: SSE_COPY_PS(XMM4,XMM7)
1114: SSE_CMPNEQ_PS(XMM4,XMM0)
1116: SSE_COPY_PS(XMM5,XMM7)
1117: SSE_CMPNEQ_PS(XMM5,XMM1)
1119: SSE_COPY_PS(XMM6,XMM7)
1120: SSE_CMPNEQ_PS(XMM6,XMM2)
1122: SSE_CMPNEQ_PS(XMM7,XMM3)
1124: /* Reduce the comparisons to one SSE register */
1125: SSE_OR_PS(XMM6,XMM7)
1126: SSE_OR_PS(XMM5,XMM4)
1127: SSE_OR_PS(XMM5,XMM6)
1128: SSE_INLINE_END_1;
1130: /* Reduce the one SSE register to an integer register for branching */
1131: /* Note: Since nonzero is an int, there is no INLINE block version of this call */
1132: MOVEMASK(nonzero,XMM5);
1134: /* If block is nonzero ... */
1135: if (nonzero) {
1136: pv = ba + 16*diag_offset[row];
1137: PREFETCH_L1(&pv[16]);
1138: pj = bj + diag_offset[row] + 1;
1140: /* Form Multiplier, one column at a time (Matrix-Matrix Product) */
1141: /* L_ij^(k+1) = L_ij^(k)*inv(L_jj^(k)) */
1142: /* but the diagonal was inverted already */
1143: /* and, L_ij^(k) is already loaded into registers XMM0-XMM3 columnwise */
1145: SSE_INLINE_BEGIN_2(pv,pc)
1146: /* Column 0, product is accumulated in XMM4 */
1147: SSE_LOAD_SS(SSE_ARG_1,FLOAT_0,XMM4)
1148: SSE_SHUFFLE(XMM4,XMM4,0x00)
1149: SSE_MULT_PS(XMM4,XMM0)
1151: SSE_LOAD_SS(SSE_ARG_1,FLOAT_1,XMM5)
1152: SSE_SHUFFLE(XMM5,XMM5,0x00)
1153: SSE_MULT_PS(XMM5,XMM1)
1154: SSE_ADD_PS(XMM4,XMM5)
1156: SSE_LOAD_SS(SSE_ARG_1,FLOAT_2,XMM6)
1157: SSE_SHUFFLE(XMM6,XMM6,0x00)
1158: SSE_MULT_PS(XMM6,XMM2)
1159: SSE_ADD_PS(XMM4,XMM6)
1161: SSE_LOAD_SS(SSE_ARG_1,FLOAT_3,XMM7)
1162: SSE_SHUFFLE(XMM7,XMM7,0x00)
1163: SSE_MULT_PS(XMM7,XMM3)
1164: SSE_ADD_PS(XMM4,XMM7)
1166: SSE_STOREL_PS(SSE_ARG_2,FLOAT_0,XMM4)
1167: SSE_STOREH_PS(SSE_ARG_2,FLOAT_2,XMM4)
1169: /* Column 1, product is accumulated in XMM5 */
1170: SSE_LOAD_SS(SSE_ARG_1,FLOAT_4,XMM5)
1171: SSE_SHUFFLE(XMM5,XMM5,0x00)
1172: SSE_MULT_PS(XMM5,XMM0)
1174: SSE_LOAD_SS(SSE_ARG_1,FLOAT_5,XMM6)
1175: SSE_SHUFFLE(XMM6,XMM6,0x00)
1176: SSE_MULT_PS(XMM6,XMM1)
1177: SSE_ADD_PS(XMM5,XMM6)
1179: SSE_LOAD_SS(SSE_ARG_1,FLOAT_6,XMM7)
1180: SSE_SHUFFLE(XMM7,XMM7,0x00)
1181: SSE_MULT_PS(XMM7,XMM2)
1182: SSE_ADD_PS(XMM5,XMM7)
1184: SSE_LOAD_SS(SSE_ARG_1,FLOAT_7,XMM6)
1185: SSE_SHUFFLE(XMM6,XMM6,0x00)
1186: SSE_MULT_PS(XMM6,XMM3)
1187: SSE_ADD_PS(XMM5,XMM6)
1189: SSE_STOREL_PS(SSE_ARG_2,FLOAT_4,XMM5)
1190: SSE_STOREH_PS(SSE_ARG_2,FLOAT_6,XMM5)
1192: SSE_PREFETCH_L1(SSE_ARG_1,FLOAT_24)
1194: /* Column 2, product is accumulated in XMM6 */
1195: SSE_LOAD_SS(SSE_ARG_1,FLOAT_8,XMM6)
1196: SSE_SHUFFLE(XMM6,XMM6,0x00)
1197: SSE_MULT_PS(XMM6,XMM0)
1199: SSE_LOAD_SS(SSE_ARG_1,FLOAT_9,XMM7)
1200: SSE_SHUFFLE(XMM7,XMM7,0x00)
1201: SSE_MULT_PS(XMM7,XMM1)
1202: SSE_ADD_PS(XMM6,XMM7)
1204: SSE_LOAD_SS(SSE_ARG_1,FLOAT_10,XMM7)
1205: SSE_SHUFFLE(XMM7,XMM7,0x00)
1206: SSE_MULT_PS(XMM7,XMM2)
1207: SSE_ADD_PS(XMM6,XMM7)
1209: SSE_LOAD_SS(SSE_ARG_1,FLOAT_11,XMM7)
1210: SSE_SHUFFLE(XMM7,XMM7,0x00)
1211: SSE_MULT_PS(XMM7,XMM3)
1212: SSE_ADD_PS(XMM6,XMM7)
1214: SSE_STOREL_PS(SSE_ARG_2,FLOAT_8,XMM6)
1215: SSE_STOREH_PS(SSE_ARG_2,FLOAT_10,XMM6)
1217: /* Note: For the last column, we no longer need to preserve XMM0->XMM3 */
1218: /* Column 3, product is accumulated in XMM0 */
1219: SSE_LOAD_SS(SSE_ARG_1,FLOAT_12,XMM7)
1220: SSE_SHUFFLE(XMM7,XMM7,0x00)
1221: SSE_MULT_PS(XMM0,XMM7)
1223: SSE_LOAD_SS(SSE_ARG_1,FLOAT_13,XMM7)
1224: SSE_SHUFFLE(XMM7,XMM7,0x00)
1225: SSE_MULT_PS(XMM1,XMM7)
1226: SSE_ADD_PS(XMM0,XMM1)
1228: SSE_LOAD_SS(SSE_ARG_1,FLOAT_14,XMM1)
1229: SSE_SHUFFLE(XMM1,XMM1,0x00)
1230: SSE_MULT_PS(XMM1,XMM2)
1231: SSE_ADD_PS(XMM0,XMM1)
1233: SSE_LOAD_SS(SSE_ARG_1,FLOAT_15,XMM7)
1234: SSE_SHUFFLE(XMM7,XMM7,0x00)
1235: SSE_MULT_PS(XMM3,XMM7)
1236: SSE_ADD_PS(XMM0,XMM3)
1238: SSE_STOREL_PS(SSE_ARG_2,FLOAT_12,XMM0)
1239: SSE_STOREH_PS(SSE_ARG_2,FLOAT_14,XMM0)
1241: /* Simplify Bookkeeping -- Completely Unnecessary Instructions */
1242: /* This is code to be maintained and read by humans afterall. */
1243: /* Copy Multiplier Col 3 into XMM3 */
1244: SSE_COPY_PS(XMM3,XMM0)
1245: /* Copy Multiplier Col 2 into XMM2 */
1246: SSE_COPY_PS(XMM2,XMM6)
1247: /* Copy Multiplier Col 1 into XMM1 */
1248: SSE_COPY_PS(XMM1,XMM5)
1249: /* Copy Multiplier Col 0 into XMM0 */
1250: SSE_COPY_PS(XMM0,XMM4)
1251: SSE_INLINE_END_2;
1253: /* Update the row: */
1254: nz = bi[row+1] - diag_offset[row] - 1;
1255: pv += 16;
1256: for (j=0; j<nz; j++) {
1257: PREFETCH_L1(&pv[16]);
1258: x = rtmp + 16*((unsigned int)pj[j]);
1259: /* x = rtmp + 4*pj[j]; */
1261: /* X:=X-M*PV, One column at a time */
1262: /* Note: M is already loaded columnwise into registers XMM0-XMM3 */
1263: SSE_INLINE_BEGIN_2(x,pv)
1264: /* Load First Column of X*/
1265: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM4)
1266: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM4)
1268: /* Matrix-Vector Product: */
1269: SSE_LOAD_SS(SSE_ARG_2,FLOAT_0,XMM5)
1270: SSE_SHUFFLE(XMM5,XMM5,0x00)
1271: SSE_MULT_PS(XMM5,XMM0)
1272: SSE_SUB_PS(XMM4,XMM5)
1274: SSE_LOAD_SS(SSE_ARG_2,FLOAT_1,XMM6)
1275: SSE_SHUFFLE(XMM6,XMM6,0x00)
1276: SSE_MULT_PS(XMM6,XMM1)
1277: SSE_SUB_PS(XMM4,XMM6)
1279: SSE_LOAD_SS(SSE_ARG_2,FLOAT_2,XMM7)
1280: SSE_SHUFFLE(XMM7,XMM7,0x00)
1281: SSE_MULT_PS(XMM7,XMM2)
1282: SSE_SUB_PS(XMM4,XMM7)
1284: SSE_LOAD_SS(SSE_ARG_2,FLOAT_3,XMM5)
1285: SSE_SHUFFLE(XMM5,XMM5,0x00)
1286: SSE_MULT_PS(XMM5,XMM3)
1287: SSE_SUB_PS(XMM4,XMM5)
1289: SSE_STOREL_PS(SSE_ARG_1,FLOAT_0,XMM4)
1290: SSE_STOREH_PS(SSE_ARG_1,FLOAT_2,XMM4)
1292: /* Second Column */
1293: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM5)
1294: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM5)
1296: /* Matrix-Vector Product: */
1297: SSE_LOAD_SS(SSE_ARG_2,FLOAT_4,XMM6)
1298: SSE_SHUFFLE(XMM6,XMM6,0x00)
1299: SSE_MULT_PS(XMM6,XMM0)
1300: SSE_SUB_PS(XMM5,XMM6)
1302: SSE_LOAD_SS(SSE_ARG_2,FLOAT_5,XMM7)
1303: SSE_SHUFFLE(XMM7,XMM7,0x00)
1304: SSE_MULT_PS(XMM7,XMM1)
1305: SSE_SUB_PS(XMM5,XMM7)
1307: SSE_LOAD_SS(SSE_ARG_2,FLOAT_6,XMM4)
1308: SSE_SHUFFLE(XMM4,XMM4,0x00)
1309: SSE_MULT_PS(XMM4,XMM2)
1310: SSE_SUB_PS(XMM5,XMM4)
1312: SSE_LOAD_SS(SSE_ARG_2,FLOAT_7,XMM6)
1313: SSE_SHUFFLE(XMM6,XMM6,0x00)
1314: SSE_MULT_PS(XMM6,XMM3)
1315: SSE_SUB_PS(XMM5,XMM6)
1317: SSE_STOREL_PS(SSE_ARG_1,FLOAT_4,XMM5)
1318: SSE_STOREH_PS(SSE_ARG_1,FLOAT_6,XMM5)
1320: SSE_PREFETCH_L1(SSE_ARG_2,FLOAT_24)
1322: /* Third Column */
1323: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM6)
1324: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM6)
1326: /* Matrix-Vector Product: */
1327: SSE_LOAD_SS(SSE_ARG_2,FLOAT_8,XMM7)
1328: SSE_SHUFFLE(XMM7,XMM7,0x00)
1329: SSE_MULT_PS(XMM7,XMM0)
1330: SSE_SUB_PS(XMM6,XMM7)
1332: SSE_LOAD_SS(SSE_ARG_2,FLOAT_9,XMM4)
1333: SSE_SHUFFLE(XMM4,XMM4,0x00)
1334: SSE_MULT_PS(XMM4,XMM1)
1335: SSE_SUB_PS(XMM6,XMM4)
1337: SSE_LOAD_SS(SSE_ARG_2,FLOAT_10,XMM5)
1338: SSE_SHUFFLE(XMM5,XMM5,0x00)
1339: SSE_MULT_PS(XMM5,XMM2)
1340: SSE_SUB_PS(XMM6,XMM5)
1342: SSE_LOAD_SS(SSE_ARG_2,FLOAT_11,XMM7)
1343: SSE_SHUFFLE(XMM7,XMM7,0x00)
1344: SSE_MULT_PS(XMM7,XMM3)
1345: SSE_SUB_PS(XMM6,XMM7)
1347: SSE_STOREL_PS(SSE_ARG_1,FLOAT_8,XMM6)
1348: SSE_STOREH_PS(SSE_ARG_1,FLOAT_10,XMM6)
1350: /* Fourth Column */
1351: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM4)
1352: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM4)
1354: /* Matrix-Vector Product: */
1355: SSE_LOAD_SS(SSE_ARG_2,FLOAT_12,XMM5)
1356: SSE_SHUFFLE(XMM5,XMM5,0x00)
1357: SSE_MULT_PS(XMM5,XMM0)
1358: SSE_SUB_PS(XMM4,XMM5)
1360: SSE_LOAD_SS(SSE_ARG_2,FLOAT_13,XMM6)
1361: SSE_SHUFFLE(XMM6,XMM6,0x00)
1362: SSE_MULT_PS(XMM6,XMM1)
1363: SSE_SUB_PS(XMM4,XMM6)
1365: SSE_LOAD_SS(SSE_ARG_2,FLOAT_14,XMM7)
1366: SSE_SHUFFLE(XMM7,XMM7,0x00)
1367: SSE_MULT_PS(XMM7,XMM2)
1368: SSE_SUB_PS(XMM4,XMM7)
1370: SSE_LOAD_SS(SSE_ARG_2,FLOAT_15,XMM5)
1371: SSE_SHUFFLE(XMM5,XMM5,0x00)
1372: SSE_MULT_PS(XMM5,XMM3)
1373: SSE_SUB_PS(XMM4,XMM5)
1375: SSE_STOREL_PS(SSE_ARG_1,FLOAT_12,XMM4)
1376: SSE_STOREH_PS(SSE_ARG_1,FLOAT_14,XMM4)
1377: SSE_INLINE_END_2;
1378: pv += 16;
1379: }
1380: PetscLogFlops(128.0*nz+112.0);
1381: }
1382: row = (unsigned int)(*bjtmp++);
1383: /* row = (*bjtmp++)/4; */
1384: /* bjtmp++; */
1385: }
1386: /* finished row so stick it into b->a */
1387: pv = ba + 16*bi[i];
1388: pj = bj + bi[i];
1389: nz = bi[i+1] - bi[i];
1391: /* Copy x block back into pv block */
1392: for (j=0; j<nz; j++) {
1393: x = rtmp+16*((unsigned int)pj[j]);
1394: /* x = rtmp+4*pj[j]; */
1396: SSE_INLINE_BEGIN_2(x,pv)
1397: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
1398: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM1)
1399: SSE_STOREL_PS(SSE_ARG_2,FLOAT_0,XMM1)
1401: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM2)
1402: SSE_STOREH_PS(SSE_ARG_2,FLOAT_2,XMM2)
1404: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM3)
1405: SSE_STOREL_PS(SSE_ARG_2,FLOAT_4,XMM3)
1407: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM4)
1408: SSE_STOREH_PS(SSE_ARG_2,FLOAT_6,XMM4)
1410: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM5)
1411: SSE_STOREL_PS(SSE_ARG_2,FLOAT_8,XMM5)
1413: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM6)
1414: SSE_STOREH_PS(SSE_ARG_2,FLOAT_10,XMM6)
1416: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM7)
1417: SSE_STOREL_PS(SSE_ARG_2,FLOAT_12,XMM7)
1419: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM0)
1420: SSE_STOREH_PS(SSE_ARG_2,FLOAT_14,XMM0)
1421: SSE_INLINE_END_2;
1422: pv += 16;
1423: }
1424: /* invert diagonal block */
1425: w = ba + 16*diag_offset[i];
1426: if (pivotinblocks) {
1427: PetscKernel_A_gets_inverse_A_4(w,shift,allowzeropivot,&zeropivotdetected);
1428: if (zeropivotdetected) C->factorerrortype = MAT_FACTOR_NUMERIC_ZEROPIVOT;
1429: } else {
1430: PetscKernel_A_gets_inverse_A_4_nopivot(w);
1431: }
1432: /* Note: Using Kramer's rule, flop count below might be infairly high or low? */
1433: }
1435: PetscFree(rtmp);
1437: C->ops->solve = MatSolve_SeqBAIJ_4_NaturalOrdering_SSE;
1438: C->ops->solvetranspose = MatSolveTranspose_SeqBAIJ_4_NaturalOrdering_SSE;
1439: C->assembled = PETSC_TRUE;
1441: PetscLogFlops(1.333333333333*bs*bs2*b->mbs);
1442: /* Flop Count from inverting diagonal blocks */
1443: SSE_SCOPE_END;
1444: return(0);
1445: }
1447: PetscErrorCode MatLUFactorNumeric_SeqBAIJ_4_NaturalOrdering_SSE_usj(Mat C,Mat A,const MatFactorInfo *info)
1448: {
1449: Mat_SeqBAIJ *a = (Mat_SeqBAIJ*)A->data,*b = (Mat_SeqBAIJ*)C->data;
1451: int i,j,n = a->mbs;
1452: unsigned short *bj = (unsigned short*)(b->j),*bjtmp,*pj;
1453: unsigned int row;
1454: int *ajtmpold,nz,*bi=b->i;
1455: int *diag_offset = b->diag,*ai=a->i,*aj=a->j;
1456: MatScalar *pv,*v,*rtmp,*pc,*w,*x;
1457: MatScalar *ba = b->a,*aa = a->a;
1458: int nonzero=0;
1459: PetscBool pivotinblocks = b->pivotinblocks;
1460: PetscReal shift = info->shiftamount;
1461: PetscBool allowzeropivot,zeropivotdetected=PETSC_FALSE;
1464: allowzeropivot = PetscNot(A->erroriffailure);
1465: SSE_SCOPE_BEGIN;
1467: if ((unsigned long)aa%16!=0) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_BADPTR,"Pointer aa is not 16 byte aligned. SSE will not work.");
1468: if ((unsigned long)ba%16!=0) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_BADPTR,"Pointer ba is not 16 byte aligned. SSE will not work.");
1469: PetscMalloc1(16*(n+1),&rtmp);
1470: if ((unsigned long)rtmp%16!=0) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_BADPTR,"Pointer rtmp is not 16 byte aligned. SSE will not work.");
1471: /* if ((unsigned long)bj==(unsigned long)aj) { */
1472: /* colscale = 4; */
1473: /* } */
1474: if ((unsigned long)bj==(unsigned long)aj) {
1475: return(MatLUFactorNumeric_SeqBAIJ_4_NaturalOrdering_SSE_usj_Inplace(C));
1476: }
1478: for (i=0; i<n; i++) {
1479: nz = bi[i+1] - bi[i];
1480: bjtmp = bj + bi[i];
1481: /* zero out the 4x4 block accumulators */
1482: /* zero out one register */
1483: XOR_PS(XMM7,XMM7);
1484: for (j=0; j<nz; j++) {
1485: x = rtmp+16*((unsigned int)bjtmp[j]);
1486: /* x = rtmp+4*bjtmp[j]; */
1487: SSE_INLINE_BEGIN_1(x)
1488: /* Copy zero register to memory locations */
1489: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
1490: SSE_STOREL_PS(SSE_ARG_1,FLOAT_0,XMM7)
1491: SSE_STOREH_PS(SSE_ARG_1,FLOAT_2,XMM7)
1492: SSE_STOREL_PS(SSE_ARG_1,FLOAT_4,XMM7)
1493: SSE_STOREH_PS(SSE_ARG_1,FLOAT_6,XMM7)
1494: SSE_STOREL_PS(SSE_ARG_1,FLOAT_8,XMM7)
1495: SSE_STOREH_PS(SSE_ARG_1,FLOAT_10,XMM7)
1496: SSE_STOREL_PS(SSE_ARG_1,FLOAT_12,XMM7)
1497: SSE_STOREH_PS(SSE_ARG_1,FLOAT_14,XMM7)
1498: SSE_INLINE_END_1;
1499: }
1500: /* load in initial (unfactored row) */
1501: nz = ai[i+1] - ai[i];
1502: ajtmpold = aj + ai[i];
1503: v = aa + 16*ai[i];
1504: for (j=0; j<nz; j++) {
1505: x = rtmp+16*ajtmpold[j];
1506: /* x = rtmp+colscale*ajtmpold[j]; */
1507: /* Copy v block into x block */
1508: SSE_INLINE_BEGIN_2(v,x)
1509: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
1510: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM0)
1511: SSE_STOREL_PS(SSE_ARG_2,FLOAT_0,XMM0)
1513: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM1)
1514: SSE_STOREH_PS(SSE_ARG_2,FLOAT_2,XMM1)
1516: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM2)
1517: SSE_STOREL_PS(SSE_ARG_2,FLOAT_4,XMM2)
1519: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM3)
1520: SSE_STOREH_PS(SSE_ARG_2,FLOAT_6,XMM3)
1522: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM4)
1523: SSE_STOREL_PS(SSE_ARG_2,FLOAT_8,XMM4)
1525: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM5)
1526: SSE_STOREH_PS(SSE_ARG_2,FLOAT_10,XMM5)
1528: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM6)
1529: SSE_STOREL_PS(SSE_ARG_2,FLOAT_12,XMM6)
1531: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM0)
1532: SSE_STOREH_PS(SSE_ARG_2,FLOAT_14,XMM0)
1533: SSE_INLINE_END_2;
1535: v += 16;
1536: }
1537: /* row = (*bjtmp++)/4; */
1538: row = (unsigned int)(*bjtmp++);
1539: while (row < i) {
1540: pc = rtmp + 16*row;
1541: SSE_INLINE_BEGIN_1(pc)
1542: /* Load block from lower triangle */
1543: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
1544: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM0)
1545: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM0)
1547: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM1)
1548: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM1)
1550: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM2)
1551: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM2)
1553: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM3)
1554: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM3)
1556: /* Compare block to zero block */
1558: SSE_COPY_PS(XMM4,XMM7)
1559: SSE_CMPNEQ_PS(XMM4,XMM0)
1561: SSE_COPY_PS(XMM5,XMM7)
1562: SSE_CMPNEQ_PS(XMM5,XMM1)
1564: SSE_COPY_PS(XMM6,XMM7)
1565: SSE_CMPNEQ_PS(XMM6,XMM2)
1567: SSE_CMPNEQ_PS(XMM7,XMM3)
1569: /* Reduce the comparisons to one SSE register */
1570: SSE_OR_PS(XMM6,XMM7)
1571: SSE_OR_PS(XMM5,XMM4)
1572: SSE_OR_PS(XMM5,XMM6)
1573: SSE_INLINE_END_1;
1575: /* Reduce the one SSE register to an integer register for branching */
1576: /* Note: Since nonzero is an int, there is no INLINE block version of this call */
1577: MOVEMASK(nonzero,XMM5);
1579: /* If block is nonzero ... */
1580: if (nonzero) {
1581: pv = ba + 16*diag_offset[row];
1582: PREFETCH_L1(&pv[16]);
1583: pj = bj + diag_offset[row] + 1;
1585: /* Form Multiplier, one column at a time (Matrix-Matrix Product) */
1586: /* L_ij^(k+1) = L_ij^(k)*inv(L_jj^(k)) */
1587: /* but the diagonal was inverted already */
1588: /* and, L_ij^(k) is already loaded into registers XMM0-XMM3 columnwise */
1590: SSE_INLINE_BEGIN_2(pv,pc)
1591: /* Column 0, product is accumulated in XMM4 */
1592: SSE_LOAD_SS(SSE_ARG_1,FLOAT_0,XMM4)
1593: SSE_SHUFFLE(XMM4,XMM4,0x00)
1594: SSE_MULT_PS(XMM4,XMM0)
1596: SSE_LOAD_SS(SSE_ARG_1,FLOAT_1,XMM5)
1597: SSE_SHUFFLE(XMM5,XMM5,0x00)
1598: SSE_MULT_PS(XMM5,XMM1)
1599: SSE_ADD_PS(XMM4,XMM5)
1601: SSE_LOAD_SS(SSE_ARG_1,FLOAT_2,XMM6)
1602: SSE_SHUFFLE(XMM6,XMM6,0x00)
1603: SSE_MULT_PS(XMM6,XMM2)
1604: SSE_ADD_PS(XMM4,XMM6)
1606: SSE_LOAD_SS(SSE_ARG_1,FLOAT_3,XMM7)
1607: SSE_SHUFFLE(XMM7,XMM7,0x00)
1608: SSE_MULT_PS(XMM7,XMM3)
1609: SSE_ADD_PS(XMM4,XMM7)
1611: SSE_STOREL_PS(SSE_ARG_2,FLOAT_0,XMM4)
1612: SSE_STOREH_PS(SSE_ARG_2,FLOAT_2,XMM4)
1614: /* Column 1, product is accumulated in XMM5 */
1615: SSE_LOAD_SS(SSE_ARG_1,FLOAT_4,XMM5)
1616: SSE_SHUFFLE(XMM5,XMM5,0x00)
1617: SSE_MULT_PS(XMM5,XMM0)
1619: SSE_LOAD_SS(SSE_ARG_1,FLOAT_5,XMM6)
1620: SSE_SHUFFLE(XMM6,XMM6,0x00)
1621: SSE_MULT_PS(XMM6,XMM1)
1622: SSE_ADD_PS(XMM5,XMM6)
1624: SSE_LOAD_SS(SSE_ARG_1,FLOAT_6,XMM7)
1625: SSE_SHUFFLE(XMM7,XMM7,0x00)
1626: SSE_MULT_PS(XMM7,XMM2)
1627: SSE_ADD_PS(XMM5,XMM7)
1629: SSE_LOAD_SS(SSE_ARG_1,FLOAT_7,XMM6)
1630: SSE_SHUFFLE(XMM6,XMM6,0x00)
1631: SSE_MULT_PS(XMM6,XMM3)
1632: SSE_ADD_PS(XMM5,XMM6)
1634: SSE_STOREL_PS(SSE_ARG_2,FLOAT_4,XMM5)
1635: SSE_STOREH_PS(SSE_ARG_2,FLOAT_6,XMM5)
1637: SSE_PREFETCH_L1(SSE_ARG_1,FLOAT_24)
1639: /* Column 2, product is accumulated in XMM6 */
1640: SSE_LOAD_SS(SSE_ARG_1,FLOAT_8,XMM6)
1641: SSE_SHUFFLE(XMM6,XMM6,0x00)
1642: SSE_MULT_PS(XMM6,XMM0)
1644: SSE_LOAD_SS(SSE_ARG_1,FLOAT_9,XMM7)
1645: SSE_SHUFFLE(XMM7,XMM7,0x00)
1646: SSE_MULT_PS(XMM7,XMM1)
1647: SSE_ADD_PS(XMM6,XMM7)
1649: SSE_LOAD_SS(SSE_ARG_1,FLOAT_10,XMM7)
1650: SSE_SHUFFLE(XMM7,XMM7,0x00)
1651: SSE_MULT_PS(XMM7,XMM2)
1652: SSE_ADD_PS(XMM6,XMM7)
1654: SSE_LOAD_SS(SSE_ARG_1,FLOAT_11,XMM7)
1655: SSE_SHUFFLE(XMM7,XMM7,0x00)
1656: SSE_MULT_PS(XMM7,XMM3)
1657: SSE_ADD_PS(XMM6,XMM7)
1659: SSE_STOREL_PS(SSE_ARG_2,FLOAT_8,XMM6)
1660: SSE_STOREH_PS(SSE_ARG_2,FLOAT_10,XMM6)
1662: /* Note: For the last column, we no longer need to preserve XMM0->XMM3 */
1663: /* Column 3, product is accumulated in XMM0 */
1664: SSE_LOAD_SS(SSE_ARG_1,FLOAT_12,XMM7)
1665: SSE_SHUFFLE(XMM7,XMM7,0x00)
1666: SSE_MULT_PS(XMM0,XMM7)
1668: SSE_LOAD_SS(SSE_ARG_1,FLOAT_13,XMM7)
1669: SSE_SHUFFLE(XMM7,XMM7,0x00)
1670: SSE_MULT_PS(XMM1,XMM7)
1671: SSE_ADD_PS(XMM0,XMM1)
1673: SSE_LOAD_SS(SSE_ARG_1,FLOAT_14,XMM1)
1674: SSE_SHUFFLE(XMM1,XMM1,0x00)
1675: SSE_MULT_PS(XMM1,XMM2)
1676: SSE_ADD_PS(XMM0,XMM1)
1678: SSE_LOAD_SS(SSE_ARG_1,FLOAT_15,XMM7)
1679: SSE_SHUFFLE(XMM7,XMM7,0x00)
1680: SSE_MULT_PS(XMM3,XMM7)
1681: SSE_ADD_PS(XMM0,XMM3)
1683: SSE_STOREL_PS(SSE_ARG_2,FLOAT_12,XMM0)
1684: SSE_STOREH_PS(SSE_ARG_2,FLOAT_14,XMM0)
1686: /* Simplify Bookkeeping -- Completely Unnecessary Instructions */
1687: /* This is code to be maintained and read by humans afterall. */
1688: /* Copy Multiplier Col 3 into XMM3 */
1689: SSE_COPY_PS(XMM3,XMM0)
1690: /* Copy Multiplier Col 2 into XMM2 */
1691: SSE_COPY_PS(XMM2,XMM6)
1692: /* Copy Multiplier Col 1 into XMM1 */
1693: SSE_COPY_PS(XMM1,XMM5)
1694: /* Copy Multiplier Col 0 into XMM0 */
1695: SSE_COPY_PS(XMM0,XMM4)
1696: SSE_INLINE_END_2;
1698: /* Update the row: */
1699: nz = bi[row+1] - diag_offset[row] - 1;
1700: pv += 16;
1701: for (j=0; j<nz; j++) {
1702: PREFETCH_L1(&pv[16]);
1703: x = rtmp + 16*((unsigned int)pj[j]);
1704: /* x = rtmp + 4*pj[j]; */
1706: /* X:=X-M*PV, One column at a time */
1707: /* Note: M is already loaded columnwise into registers XMM0-XMM3 */
1708: SSE_INLINE_BEGIN_2(x,pv)
1709: /* Load First Column of X*/
1710: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM4)
1711: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM4)
1713: /* Matrix-Vector Product: */
1714: SSE_LOAD_SS(SSE_ARG_2,FLOAT_0,XMM5)
1715: SSE_SHUFFLE(XMM5,XMM5,0x00)
1716: SSE_MULT_PS(XMM5,XMM0)
1717: SSE_SUB_PS(XMM4,XMM5)
1719: SSE_LOAD_SS(SSE_ARG_2,FLOAT_1,XMM6)
1720: SSE_SHUFFLE(XMM6,XMM6,0x00)
1721: SSE_MULT_PS(XMM6,XMM1)
1722: SSE_SUB_PS(XMM4,XMM6)
1724: SSE_LOAD_SS(SSE_ARG_2,FLOAT_2,XMM7)
1725: SSE_SHUFFLE(XMM7,XMM7,0x00)
1726: SSE_MULT_PS(XMM7,XMM2)
1727: SSE_SUB_PS(XMM4,XMM7)
1729: SSE_LOAD_SS(SSE_ARG_2,FLOAT_3,XMM5)
1730: SSE_SHUFFLE(XMM5,XMM5,0x00)
1731: SSE_MULT_PS(XMM5,XMM3)
1732: SSE_SUB_PS(XMM4,XMM5)
1734: SSE_STOREL_PS(SSE_ARG_1,FLOAT_0,XMM4)
1735: SSE_STOREH_PS(SSE_ARG_1,FLOAT_2,XMM4)
1737: /* Second Column */
1738: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM5)
1739: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM5)
1741: /* Matrix-Vector Product: */
1742: SSE_LOAD_SS(SSE_ARG_2,FLOAT_4,XMM6)
1743: SSE_SHUFFLE(XMM6,XMM6,0x00)
1744: SSE_MULT_PS(XMM6,XMM0)
1745: SSE_SUB_PS(XMM5,XMM6)
1747: SSE_LOAD_SS(SSE_ARG_2,FLOAT_5,XMM7)
1748: SSE_SHUFFLE(XMM7,XMM7,0x00)
1749: SSE_MULT_PS(XMM7,XMM1)
1750: SSE_SUB_PS(XMM5,XMM7)
1752: SSE_LOAD_SS(SSE_ARG_2,FLOAT_6,XMM4)
1753: SSE_SHUFFLE(XMM4,XMM4,0x00)
1754: SSE_MULT_PS(XMM4,XMM2)
1755: SSE_SUB_PS(XMM5,XMM4)
1757: SSE_LOAD_SS(SSE_ARG_2,FLOAT_7,XMM6)
1758: SSE_SHUFFLE(XMM6,XMM6,0x00)
1759: SSE_MULT_PS(XMM6,XMM3)
1760: SSE_SUB_PS(XMM5,XMM6)
1762: SSE_STOREL_PS(SSE_ARG_1,FLOAT_4,XMM5)
1763: SSE_STOREH_PS(SSE_ARG_1,FLOAT_6,XMM5)
1765: SSE_PREFETCH_L1(SSE_ARG_2,FLOAT_24)
1767: /* Third Column */
1768: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM6)
1769: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM6)
1771: /* Matrix-Vector Product: */
1772: SSE_LOAD_SS(SSE_ARG_2,FLOAT_8,XMM7)
1773: SSE_SHUFFLE(XMM7,XMM7,0x00)
1774: SSE_MULT_PS(XMM7,XMM0)
1775: SSE_SUB_PS(XMM6,XMM7)
1777: SSE_LOAD_SS(SSE_ARG_2,FLOAT_9,XMM4)
1778: SSE_SHUFFLE(XMM4,XMM4,0x00)
1779: SSE_MULT_PS(XMM4,XMM1)
1780: SSE_SUB_PS(XMM6,XMM4)
1782: SSE_LOAD_SS(SSE_ARG_2,FLOAT_10,XMM5)
1783: SSE_SHUFFLE(XMM5,XMM5,0x00)
1784: SSE_MULT_PS(XMM5,XMM2)
1785: SSE_SUB_PS(XMM6,XMM5)
1787: SSE_LOAD_SS(SSE_ARG_2,FLOAT_11,XMM7)
1788: SSE_SHUFFLE(XMM7,XMM7,0x00)
1789: SSE_MULT_PS(XMM7,XMM3)
1790: SSE_SUB_PS(XMM6,XMM7)
1792: SSE_STOREL_PS(SSE_ARG_1,FLOAT_8,XMM6)
1793: SSE_STOREH_PS(SSE_ARG_1,FLOAT_10,XMM6)
1795: /* Fourth Column */
1796: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM4)
1797: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM4)
1799: /* Matrix-Vector Product: */
1800: SSE_LOAD_SS(SSE_ARG_2,FLOAT_12,XMM5)
1801: SSE_SHUFFLE(XMM5,XMM5,0x00)
1802: SSE_MULT_PS(XMM5,XMM0)
1803: SSE_SUB_PS(XMM4,XMM5)
1805: SSE_LOAD_SS(SSE_ARG_2,FLOAT_13,XMM6)
1806: SSE_SHUFFLE(XMM6,XMM6,0x00)
1807: SSE_MULT_PS(XMM6,XMM1)
1808: SSE_SUB_PS(XMM4,XMM6)
1810: SSE_LOAD_SS(SSE_ARG_2,FLOAT_14,XMM7)
1811: SSE_SHUFFLE(XMM7,XMM7,0x00)
1812: SSE_MULT_PS(XMM7,XMM2)
1813: SSE_SUB_PS(XMM4,XMM7)
1815: SSE_LOAD_SS(SSE_ARG_2,FLOAT_15,XMM5)
1816: SSE_SHUFFLE(XMM5,XMM5,0x00)
1817: SSE_MULT_PS(XMM5,XMM3)
1818: SSE_SUB_PS(XMM4,XMM5)
1820: SSE_STOREL_PS(SSE_ARG_1,FLOAT_12,XMM4)
1821: SSE_STOREH_PS(SSE_ARG_1,FLOAT_14,XMM4)
1822: SSE_INLINE_END_2;
1823: pv += 16;
1824: }
1825: PetscLogFlops(128.0*nz+112.0);
1826: }
1827: row = (unsigned int)(*bjtmp++);
1828: /* row = (*bjtmp++)/4; */
1829: /* bjtmp++; */
1830: }
1831: /* finished row so stick it into b->a */
1832: pv = ba + 16*bi[i];
1833: pj = bj + bi[i];
1834: nz = bi[i+1] - bi[i];
1836: /* Copy x block back into pv block */
1837: for (j=0; j<nz; j++) {
1838: x = rtmp+16*((unsigned int)pj[j]);
1839: /* x = rtmp+4*pj[j]; */
1841: SSE_INLINE_BEGIN_2(x,pv)
1842: /* Note: on future SSE architectures, STORE might be more efficient than STOREL/H */
1843: SSE_LOADL_PS(SSE_ARG_1,FLOAT_0,XMM1)
1844: SSE_STOREL_PS(SSE_ARG_2,FLOAT_0,XMM1)
1846: SSE_LOADH_PS(SSE_ARG_1,FLOAT_2,XMM2)
1847: SSE_STOREH_PS(SSE_ARG_2,FLOAT_2,XMM2)
1849: SSE_LOADL_PS(SSE_ARG_1,FLOAT_4,XMM3)
1850: SSE_STOREL_PS(SSE_ARG_2,FLOAT_4,XMM3)
1852: SSE_LOADH_PS(SSE_ARG_1,FLOAT_6,XMM4)
1853: SSE_STOREH_PS(SSE_ARG_2,FLOAT_6,XMM4)
1855: SSE_LOADL_PS(SSE_ARG_1,FLOAT_8,XMM5)
1856: SSE_STOREL_PS(SSE_ARG_2,FLOAT_8,XMM5)
1858: SSE_LOADH_PS(SSE_ARG_1,FLOAT_10,XMM6)
1859: SSE_STOREH_PS(SSE_ARG_2,FLOAT_10,XMM6)
1861: SSE_LOADL_PS(SSE_ARG_1,FLOAT_12,XMM7)
1862: SSE_STOREL_PS(SSE_ARG_2,FLOAT_12,XMM7)
1864: SSE_LOADH_PS(SSE_ARG_1,FLOAT_14,XMM0)
1865: SSE_STOREH_PS(SSE_ARG_2,FLOAT_14,XMM0)
1866: SSE_INLINE_END_2;
1867: pv += 16;
1868: }
1869: /* invert diagonal block */
1870: w = ba + 16*diag_offset[i];
1871: if (pivotinblocks) {
1872: PetscKernel_A_gets_inverse_A_4(w,shift,allowzeropivot,,&zeropivotdetected);
1873: if (zeropivotdetected) C->factorerrortype = MAT_FACTOR_NUMERIC_ZEROPIVOT;
1874: } else {
1875: PetscKernel_A_gets_inverse_A_4_nopivot(w);
1876: }
1877: /* PetscKernel_A_gets_inverse_A_4_SSE(w); */
1878: /* Note: Using Kramer's rule, flop count below might be infairly high or low? */
1879: }
1881: PetscFree(rtmp);
1883: C->ops->solve = MatSolve_SeqBAIJ_4_NaturalOrdering_SSE;
1884: C->ops->solvetranspose = MatSolveTranspose_SeqBAIJ_4_NaturalOrdering_SSE;
1885: C->assembled = PETSC_TRUE;
1887: PetscLogFlops(1.333333333333*bs*bs2*b->mbs);
1888: /* Flop Count from inverting diagonal blocks */
1889: SSE_SCOPE_END;
1890: return(0);
1891: }
1893: #endif