| | |
| | | |
| | | __global__ void forward_maxpool_layer_kernel(int n, int in_h, int in_w, int in_c, int stride, int size, int pad, float *input, float *output, int *indexes) |
| | | { |
| | | int h = (in_h + 2*pad)/stride; |
| | | int w = (in_w + 2*pad)/stride; |
| | | int h = (in_h + 2 * pad - size) / stride + 1; |
| | | int w = (in_w + 2 * pad - size) / stride + 1; |
| | | int c = in_c; |
| | | |
| | | int id = (blockIdx.x + blockIdx.y*gridDim.x) * blockDim.x + threadIdx.x; |
| | |
| | | |
| | | __global__ void backward_maxpool_layer_kernel(int n, int in_h, int in_w, int in_c, int stride, int size, int pad, float *delta, float *prev_delta, int *indexes) |
| | | { |
| | | int h = (in_h + 2*pad)/stride; |
| | | int w = (in_w + 2*pad)/stride; |
| | | int h = (in_h + 2 * pad - size) / stride + 1; |
| | | int w = (in_w + 2 * pad - size) / stride + 1; |
| | | int c = in_c; |
| | | int area = (size-1)/stride; |
| | | |
| | |
| | | |
| | | size_t n = h*w*c*layer.batch; |
| | | |
| | | forward_maxpool_layer_kernel<<<cuda_gridsize(n), BLOCK>>>(n, layer.h, layer.w, layer.c, layer.stride, layer.size, layer.pad, state.input, layer.output_gpu, layer.indexes_gpu); |
| | | forward_maxpool_layer_kernel<<<cuda_gridsize(n), BLOCK, 0, get_cuda_stream()>>>(n, layer.h, layer.w, layer.c, layer.stride, layer.size, layer.pad, state.input, layer.output_gpu, layer.indexes_gpu); |
| | | check_error(cudaPeekAtLastError()); |
| | | } |
| | | |