From dda993f3dd3c753dfd580d485b39c1001830fee4 Mon Sep 17 00:00:00 2001
From: AlexeyAB <alexeyab84@gmail.com>
Date: Thu, 22 Feb 2018 19:54:40 +0000
Subject: [PATCH] Use half_float16 instead of float32 if defined both CUDNN and CUDNN_HALF. Use Tensor Cores.
---
src/convolutional_layer.h | 57 ++++++++++++++++++++++++++++++++++++---------------------
1 files changed, 36 insertions(+), 21 deletions(-)
diff --git a/src/convolutional_layer.h b/src/convolutional_layer.h
index ab414ec..da98dce 100644
--- a/src/convolutional_layer.h
+++ b/src/convolutional_layer.h
@@ -1,37 +1,52 @@
#ifndef CONVOLUTIONAL_LAYER_H
#define CONVOLUTIONAL_LAYER_H
+#include "cuda.h"
#include "image.h"
#include "activations.h"
+#include "layer.h"
+#include "network.h"
-typedef struct {
- int h,w,c;
- int n;
- int stride;
- image *kernels;
- image *kernel_updates;
- double *biases;
- double *bias_updates;
- image upsampled;
- double *delta;
- double *output;
+typedef layer convolutional_layer;
- double (* activation)();
- double (* gradient)();
-} convolutional_layer;
+#ifdef GPU
+void forward_convolutional_layer_gpu(convolutional_layer layer, network_state state);
+void backward_convolutional_layer_gpu(convolutional_layer layer, network_state state);
+void update_convolutional_layer_gpu(convolutional_layer layer, int batch, float learning_rate, float momentum, float decay);
-convolutional_layer *make_convolutional_layer(int h, int w, int c, int n, int size, int stride, ACTIVATION activator);
-void forward_convolutional_layer(const convolutional_layer layer, double *in);
-void backward_convolutional_layer(convolutional_layer layer, double *input, double *delta);
-void learn_convolutional_layer(convolutional_layer layer, double *input);
+void push_convolutional_layer(convolutional_layer layer);
+void pull_convolutional_layer(convolutional_layer layer);
-void update_convolutional_layer(convolutional_layer layer, double step);
+void add_bias_gpu(float *output, float *biases, int batch, int n, int size);
+void backward_bias_gpu(float *bias_updates, float *delta, int batch, int n, int size);
+#ifdef CUDNN
+void cudnn_convolutional_setup(layer *l, int cudnn_preference);
+#endif
+#endif
-void backpropagate_convolutional_layer_convolve(image input, convolutional_layer layer);
-void visualize_convolutional_layer(convolutional_layer layer);
+convolutional_layer make_convolutional_layer(int batch, int h, int w, int c, int n, int size, int stride, int padding, ACTIVATION activation, int batch_normalize, int binary, int xnor, int adam);
+void denormalize_convolutional_layer(convolutional_layer l);
+void resize_convolutional_layer(convolutional_layer *layer, int w, int h);
+void forward_convolutional_layer(const convolutional_layer layer, network_state state);
+void update_convolutional_layer(convolutional_layer layer, int batch, float learning_rate, float momentum, float decay);
+image *visualize_convolutional_layer(convolutional_layer layer, char *window, image *prev_weights);
+void binarize_weights(float *weights, int n, int size, float *binary);
+void swap_binary(convolutional_layer *l);
+void binarize_weights2(float *weights, int n, int size, char *binary, float *scales);
+
+void backward_convolutional_layer(convolutional_layer layer, network_state state);
+
+void add_bias(float *output, float *biases, int batch, int n, int size);
+void backward_bias(float *bias_updates, float *delta, int batch, int n, int size);
image get_convolutional_image(convolutional_layer layer);
image get_convolutional_delta(convolutional_layer layer);
+image get_convolutional_weight(convolutional_layer layer, int i);
+
+int convolutional_out_height(convolutional_layer layer);
+int convolutional_out_width(convolutional_layer layer);
+void rescale_weights(convolutional_layer l, float scale, float trans);
+void rgbgr_weights(convolutional_layer l);
#endif
--
Gitblit v1.10.0