Loading gridding_cpu.c +15 −10 Original line number Diff line number Diff line Loading @@ -94,6 +94,7 @@ void gridding_data() // allocate sector arrays // note: we use the largest allocation among all sectors // unsigned long long int mem_size = (Nsec*3)*sizeof(double_t) + (Nvissec*2+Nweightss)*sizeof(float_t); double_t *memory = (double*) malloc ( (Nsec*3)*sizeof(double_t) + (Nvissec*2+Nweightss)*sizeof(float_t) ); Loading @@ -113,6 +114,8 @@ void gridding_data() double start = CPU_TIME_wt; memset( memory, 0, mem:size ); // select data for this sector uint icount = 0; uint ip = 0; Loading Loading @@ -182,6 +185,12 @@ void gridding_data() start = CPU_TIME_wt; double *stacking_target_array; if ( size > 1 ) stacking_target_array; = gridss; else stacking_target_array = grid; //We have to call different GPUs per MPI task!!! [GL] wstack(param.num_w_planes, Nsec, Loading @@ -198,7 +207,7 @@ void gridding_data() param.w_support, xaxis, yaxis, gridss, stacking_target_array, param.num_threads, rank); Loading Loading @@ -239,13 +248,9 @@ void gridding_data() } timing_wt.reduce += CPU_TIME_wt - start; } else *grid += *gridss; // Go to next sector memset ( gridss, 0, 2*param.num_w_planes*xaxis*yaxis * sizeof(double) ); } } Loading Loading
gridding_cpu.c +15 −10 Original line number Diff line number Diff line Loading @@ -94,6 +94,7 @@ void gridding_data() // allocate sector arrays // note: we use the largest allocation among all sectors // unsigned long long int mem_size = (Nsec*3)*sizeof(double_t) + (Nvissec*2+Nweightss)*sizeof(float_t); double_t *memory = (double*) malloc ( (Nsec*3)*sizeof(double_t) + (Nvissec*2+Nweightss)*sizeof(float_t) ); Loading @@ -113,6 +114,8 @@ void gridding_data() double start = CPU_TIME_wt; memset( memory, 0, mem:size ); // select data for this sector uint icount = 0; uint ip = 0; Loading Loading @@ -182,6 +185,12 @@ void gridding_data() start = CPU_TIME_wt; double *stacking_target_array; if ( size > 1 ) stacking_target_array; = gridss; else stacking_target_array = grid; //We have to call different GPUs per MPI task!!! [GL] wstack(param.num_w_planes, Nsec, Loading @@ -198,7 +207,7 @@ void gridding_data() param.w_support, xaxis, yaxis, gridss, stacking_target_array, param.num_threads, rank); Loading Loading @@ -239,13 +248,9 @@ void gridding_data() } timing_wt.reduce += CPU_TIME_wt - start; } else *grid += *gridss; // Go to next sector memset ( gridss, 0, 2*param.num_w_planes*xaxis*yaxis * sizeof(double) ); } } Loading