From d0b9326a352ed2fbc3ae66fdef40b4533a2f211d Mon Sep 17 00:00:00 2001
From: Joseph Redmon <pjreddie@gmail.com>
Date: Tue, 11 Aug 2015 06:22:27 +0000
Subject: [PATCH] Hacks to get nightmare to not break gridsizing
---
src/network_kernels.cu | 9 +++++++++
1 files changed, 9 insertions(+), 0 deletions(-)
diff --git a/src/network_kernels.cu b/src/network_kernels.cu
index 9cc8be8..3340afa 100644
--- a/src/network_kernels.cu
+++ b/src/network_kernels.cu
@@ -15,11 +15,13 @@
#include "convolutional_layer.h"
#include "deconvolutional_layer.h"
#include "maxpool_layer.h"
+#include "avgpool_layer.h"
#include "normalization_layer.h"
#include "cost_layer.h"
#include "softmax_layer.h"
#include "dropout_layer.h"
#include "route_layer.h"
+#include "blas.h"
}
float * get_network_output_gpu_layer(network net, int i);
@@ -31,6 +33,9 @@
int i;
for(i = 0; i < net.n; ++i){
layer l = net.layers[i];
+ if(l.delta_gpu){
+ scal_ongpu(l.outputs * l.batch, 0, l.delta_gpu, 1);
+ }
if(l.type == CONVOLUTIONAL){
forward_convolutional_layer_gpu(l, state);
} else if(l.type == DECONVOLUTIONAL){
@@ -49,6 +54,8 @@
forward_normalization_layer_gpu(l, state);
} else if(l.type == MAXPOOL){
forward_maxpool_layer_gpu(l, state);
+ } else if(l.type == AVGPOOL){
+ forward_avgpool_layer_gpu(l, state);
} else if(l.type == DROPOUT){
forward_dropout_layer_gpu(l, state);
} else if(l.type == ROUTE){
@@ -79,6 +86,8 @@
backward_deconvolutional_layer_gpu(l, state);
} else if(l.type == MAXPOOL){
if(i != 0) backward_maxpool_layer_gpu(l, state);
+ } else if(l.type == AVGPOOL){
+ if(i != 0) backward_avgpool_layer_gpu(l, state);
} else if(l.type == DROPOUT){
backward_dropout_layer_gpu(l, state);
} else if(l.type == DETECTION){
--
Gitblit v1.10.0