ISCE_INSAR/contrib/PyCuAmpcor/src/cuDeramp.cu

161 lines
5.6 KiB
Plaintext
Raw Normal View History

2019-01-16 19:40:08 +00:00
/*
* @file cuDeramp.cu
* @brief Derampling a batch of 2D complex images with GPU
*
* A phase ramp is equivalent to a frequency shift in frequency domain,
* which needs to be removed (deramping) in order to move the band center
* to zero. This is necessary before oversampling a complex signal.
* Method 1: each signal is decomposed into real and imaginary parts,
* and the average phase shift is obtained as atan(\sum imag / \sum real).
* The average is weighted by the amplitudes (coherence).
* Method 0 or else: skip deramping
*
2019-01-16 19:40:08 +00:00
*/
#include "cuArrays.h"
#include "float2.h"
#include <cfloat>
#include "cudaError.h"
#include "cudaUtil.h"
#include "cuAmpcorUtil.h"
#include <iostream>
#include <iomanip>
#include <cmath>
#include <limits>
// cuda does not have a good support on volatile vector struct, e.g. float2
// have to use regular float type for shared memory (volatile) data
// the following methods are defined to operate float2/complex objects through float
2019-01-16 19:40:08 +00:00
inline static __device__ void copyToShared(volatile float *s, const int i, const float2 x, const int block)
{ s[i] = x.x; s[i+block] = x.y; }
inline static __device__ void copyFromShared(float2 &x, volatile float *s, const int i, const int block)
{ x.x = s[i]; x.y = s[i+block]; }
inline static __device__ void addInShared(volatile float *s, const int i, const int j, const int block)
{ s[i] += s[i+j]; s[i+block] += s[i+j+block];}
// kernel to do sum reduction for float2 within a block
2019-01-16 19:40:08 +00:00
template <const int nthreads>
__device__ void complexSumReduceBlock(float2& sum, volatile float *shmem)
{
const int tid = threadIdx.x;
copyToShared(shmem, tid, sum, nthreads);
__syncthreads();
if (nthreads >=1024) { if (tid < 512) { addInShared(shmem, tid, 512, nthreads); } __syncthreads(); }
if (nthreads >= 512) { if (tid < 256) { addInShared(shmem, tid, 256, nthreads); } __syncthreads(); }
if (nthreads >= 256) { if (tid < 128) { addInShared(shmem, tid, 128, nthreads); } __syncthreads(); }
if (nthreads >= 128) { if (tid < 64) { addInShared(shmem, tid, 64, nthreads); } __syncthreads(); }
if (tid < 32)
{
addInShared(shmem, tid, 32, nthreads);
addInShared(shmem, tid, 16, nthreads);
addInShared(shmem, tid, 8, nthreads);
addInShared(shmem, tid, 4, nthreads);
addInShared(shmem, tid, 2, nthreads);
addInShared(shmem, tid, 1, nthreads);
}
__syncthreads();
copyFromShared(sum, shmem, 0, nthreads);
}
// cuda kernel for cuDerampMethod1
2019-01-16 19:40:08 +00:00
template<const int nthreads>
__global__ void cuDerampMethod1_kernel(float2 *images, const int imageNX, int const imageNY,
const int imageSize, const int nImages, const float normCoef)
{
__shared__ float shmem[2*nthreads];
int pixelIdx, pixelIdxX, pixelIdxY;
const int bid = blockIdx.x;
if(bid >= nImages) return;
float2 *image = images+ bid*imageSize;
const int tid = threadIdx.x;
float2 phaseDiffY = make_float2(0.0f, 0.0f);
for (int i = tid; i < imageSize; i += nthreads) {
pixelIdxY = i % imageNY;
if(pixelIdxY < imageNY -1) {
pixelIdx = i;
float2 cprod = complexMulConj( image[pixelIdx], image[pixelIdx+1]);
phaseDiffY += cprod;
}
}
complexSumReduceBlock<nthreads>(phaseDiffY, shmem);
//phaseDiffY *= normCoef;
float phaseY=atan2f(phaseDiffY.y, phaseDiffY.x);
float2 phaseDiffX = make_float2(0.0f, 0.0f);
for (int i = tid; i < imageSize; i += nthreads) {
pixelIdxX = i / imageNY;
if(pixelIdxX < imageNX -1) {
pixelIdx = i;
float2 cprod = complexMulConj(image[i], image[i+imageNY]);
phaseDiffX += cprod;
}
}
complexSumReduceBlock<nthreads>(phaseDiffX, shmem);
//phaseDiffX *= normCoef;
float phaseX = atan2f(phaseDiffX.y, phaseDiffX.x); //+FLT_EPSILON
for (int i = tid; i < imageSize; i += nthreads)
{
pixelIdxX = i%imageNY;
pixelIdxY = i/imageNY;
float phase = pixelIdxX*phaseX + pixelIdxY*phaseY;
float2 phase_factor = make_float2(cosf(phase), sinf(phase));
image[i] *= phase_factor;
}
}
/**
* Deramp a complex signal with Method 1
* @brief Each signal is decomposed into real and imaginary parts,
* and the average phase shift is obtained as atan(\sum imag / \sum real).
* @param[inout] images input/output complex signals
* @param[in] stream cuda stream
*/
2019-01-16 19:40:08 +00:00
void cuDerampMethod1(cuArrays<float2> *images, cudaStream_t stream)
{
const dim3 grid(images->count);
const int imageSize = images->width*images->height;
const float invSize = 1.0f/imageSize;
if(imageSize <=64) {
cuDerampMethod1_kernel<64> <<<grid, 64, 0, stream>>>
(images->devData, images->height, images->width,
imageSize, images->count, invSize); }
else if(imageSize <=128) {
cuDerampMethod1_kernel<128> <<<grid, 128, 0, stream>>>
(images->devData, images->height, images->width,
imageSize, images->count, invSize); }
else if(imageSize <=256) {
cuDerampMethod1_kernel<256> <<<grid, 256, 0, stream>>>
(images->devData, images->height, images->width,
imageSize, images->count, invSize); }
else {
cuDerampMethod1_kernel<512> <<<grid, 512, 0, stream>>>
(images->devData, images->height, images->width,
imageSize, images->count, invSize); }
2019-01-16 19:40:08 +00:00
getLastCudaError("cuDerampMethod1 kernel error\n");
}
void cuDeramp(int method, cuArrays<float2> *images, cudaStream_t stream)
{
switch(method) {
case 1:
cuDerampMethod1(images, stream);
break;
default:
break;
}
}
// end of file