Actual source code: ex147.c
petsc-3.8.4 2018-03-24
1: /* This program illustrates use of parallel real FFT */
2: static char help[]="This program illustrates the use of parallel real multi-dimensional fftw (without PETSc interface)";
3: #include <petscmat.h>
4: #include <fftw3.h>
5: #include <fftw3-mpi.h>
7: PetscInt main(PetscInt argc,char **args)
8: {
9: ptrdiff_t N0=2,N1=2,N2=2,N3=2,dim[4],N,D;
10: fftw_plan bplan,fplan;
11: fftw_complex *out;
12: double *in1,*in2;
13: ptrdiff_t alloc_local,local_n0,local_0_start;
14: ptrdiff_t local_n1,local_1_start;
15: PetscInt i,j,indx[100],n1;
16: PetscInt size,rank,n,*in,N_factor;
17: PetscScalar *data_fin,value1,one=1.0,zero=0.0;
18: PetscScalar a,*x_arr,*y_arr,*z_arr,enorm;
19: Vec fin,fout,fout1,x,y;
20: PetscRandom rnd;
23: PetscInitialize(&argc,&args,(char*)0,help);if (ierr) return ierr;
24: #if defined(PETSC_USE_COMPLEX)
25: SETERRQ(PETSC_COMM_WORLD,PETSC_ERR_SUP, "This example requires real numbers. Your current scalar type is complex");
26: #endif
27: MPI_Comm_size(PETSC_COMM_WORLD, &size);
28: MPI_Comm_rank(PETSC_COMM_WORLD, &rank);
30: PetscRandomCreate(PETSC_COMM_WORLD,&rnd);
31: D =4;
32: dim[0]=N0;dim[1]=N1;dim[2]=N2;dim[3]=N3/2+1;
35: alloc_local = fftw_mpi_local_size_transposed(D,dim,PETSC_COMM_WORLD,&local_n0,&local_0_start,&local_n1,&local_1_start);
37: printf("The value alloc_local is %ld from process %d\n",alloc_local,rank);
38: printf("The value local_n0 is %ld from process %d\n",local_n0,rank);
39: printf("The value local_0_start is %ld from process %d\n",local_0_start,rank);
40: printf("The value local_n1 is %ld from process %d\n",local_n1,rank);
41: printf("The value local_1_start is %ld from process %d\n",local_1_start,rank);
43: /* Allocate space for input and output arrays */
45: in1=(double*)fftw_malloc(sizeof(double)*alloc_local*2);
46: in2=(double*)fftw_malloc(sizeof(double)*alloc_local*2);
47: out=(fftw_complex*)fftw_malloc(sizeof(fftw_complex)*alloc_local);
50: N=2*N0*N1*N2*(N3/2+1);N_factor=N0*N1*N2*N3;
51: n=2*local_n0*N1*N2*(N3/2+1);n1=local_n1*N0*2*N1*N2;
53: /* printf("The value N is %d from process %d\n",N,rank); */
54: /* printf("The value n is %d from process %d\n",n,rank); */
55: /* printf("The value n1 is %d from process %d\n",n1,rank); */
56: /* Creating data vector and accompanying array with VeccreateMPIWithArray */
57: VecCreateMPIWithArray(PETSC_COMM_WORLD,1,n,N,(PetscScalar*)in1,&fin);
58: VecCreateMPIWithArray(PETSC_COMM_WORLD,1,n,N,(PetscScalar*)out,&fout);
59: VecCreateMPIWithArray(PETSC_COMM_WORLD,1,n,N,(PetscScalar*)in2,&fout1);
61: /* VecGetSize(fin,&size); */
62: /* printf("The size is %d\n",size); */
64: VecSet(fin,one);
65: /* VecAssemblyBegin(fin); */
66: /* VecAssemblyEnd(fin); */
67: /* VecView(fin,PETSC_VIEWER_STDOUT_WORLD); */
70: VecGetArray(fin,&x_arr);
71: VecGetArray(fout1,&z_arr);
72: VecGetArray(fout,&y_arr);
74: dim[3]=N3;
76: fplan=fftw_mpi_plan_dft_r2c(D,dim,(double*)x_arr,(fftw_complex*)y_arr,PETSC_COMM_WORLD,FFTW_ESTIMATE);
77: bplan=fftw_mpi_plan_dft_c2r(D,dim,(fftw_complex*)y_arr,(double*)z_arr,PETSC_COMM_WORLD,FFTW_ESTIMATE);
79: fftw_execute(fplan);
80: fftw_execute(bplan);
82: VecRestoreArray(fin,&x_arr);
83: VecRestoreArray(fout1,&z_arr);
84: VecRestoreArray(fout,&y_arr);
86: /* a = 1.0/(PetscReal)N_factor; */
87: /* VecScale(fout1,a); */
89: VecAssemblyBegin(fout1);
90: VecAssemblyEnd(fout1);
92: VecView(fout1,PETSC_VIEWER_STDOUT_WORLD);
94: fftw_destroy_plan(fplan);
95: fftw_destroy_plan(bplan);
96: fftw_free(in1); VecDestroy(&fin);
97: fftw_free(out); VecDestroy(&fout);
98: fftw_free(in2); VecDestroy(&fout1);
100: PetscFinalize();
101: return ierr;
102: }