From f047cfff99e00e28c02eb59b6d32386c122f9af6 Mon Sep 17 00:00:00 2001
From: Joseph Redmon <pjreddie@gmail.com>
Date: Sun, 08 Mar 2015 18:31:12 +0000
Subject: [PATCH] renamed sigmoid to logistic
---
src/activation_kernels.cu | 12 ++++++------
1 files changed, 6 insertions(+), 6 deletions(-)
diff --git a/src/activation_kernels.cu b/src/activation_kernels.cu
index a15d64b..5ee1524 100644
--- a/src/activation_kernels.cu
+++ b/src/activation_kernels.cu
@@ -4,13 +4,13 @@
}
__device__ float linear_activate_kernel(float x){return x;}
-__device__ float sigmoid_activate_kernel(float x){return 1./(1. + exp(-x));}
+__device__ float logistic_activate_kernel(float x){return 1./(1. + exp(-x));}
__device__ float relu_activate_kernel(float x){return x*(x>0);}
__device__ float ramp_activate_kernel(float x){return x*(x>0)+.1*x;}
__device__ float tanh_activate_kernel(float x){return (exp(2*x)-1)/(exp(2*x)+1);}
__device__ float linear_gradient_kernel(float x){return 1;}
-__device__ float sigmoid_gradient_kernel(float x){return (1-x)*x;}
+__device__ float logistic_gradient_kernel(float x){return (1-x)*x;}
__device__ float relu_gradient_kernel(float x){return (x>0);}
__device__ float ramp_gradient_kernel(float x){return (x>0)+.1;}
__device__ float tanh_gradient_kernel(float x){return 1-x*x;}
@@ -20,8 +20,8 @@
switch(a){
case LINEAR:
return linear_activate_kernel(x);
- case SIGMOID:
- return sigmoid_activate_kernel(x);
+ case LOGISTIC:
+ return logistic_activate_kernel(x);
case RELU:
return relu_activate_kernel(x);
case RAMP:
@@ -37,8 +37,8 @@
switch(a){
case LINEAR:
return linear_gradient_kernel(x);
- case SIGMOID:
- return sigmoid_gradient_kernel(x);
+ case LOGISTIC:
+ return logistic_gradient_kernel(x);
case RELU:
return relu_gradient_kernel(x);
case RAMP:
--
Gitblit v1.10.0