| | |
| | | } else if(net.types[i] == CONNECTED){ |
| | | connected_layer layer = *(connected_layer *)net.layers[i]; |
| | | return layer.output; |
| | | } else if(net.types[i] == CROP){ |
| | | crop_layer layer = *(crop_layer *)net.layers[i]; |
| | | return layer.output; |
| | | } else if(net.types[i] == NORMALIZATION){ |
| | | normalization_layer layer = *(normalization_layer *)net.layers[i]; |
| | | return layer.output; |
| | |
| | | maxpool_layer layer = *(maxpool_layer *)net.layers[i]; |
| | | if(i != 0) backward_maxpool_layer(layer, prev_delta); |
| | | } |
| | | else if(net.types[i] == DROPOUT){ |
| | | dropout_layer layer = *(dropout_layer *)net.layers[i]; |
| | | backward_dropout_layer(layer, prev_delta); |
| | | } |
| | | else if(net.types[i] == NORMALIZATION){ |
| | | normalization_layer layer = *(normalization_layer *)net.layers[i]; |
| | | if(i != 0) backward_normalization_layer(layer, prev_input, prev_delta); |
| | |
| | | } |
| | | } |
| | | |
| | | |
| | | |
| | | |
| | | float train_network_datum(network net, float *x, float *y) |
| | | { |
| | | #ifdef GPU |
| | | if(gpu_index >= 0) return train_network_datum_gpu(net, x, y); |
| | | #endif |
| | | forward_network(net, x, y, 1); |
| | | //int class = get_predicted_class_network(net); |
| | | backward_network(net, x); |
| | | float error = get_network_cost(net); |
| | | update_network(net); |
| | | //return (y[class]?1:0); |
| | | return error; |
| | | } |
| | | |
| | |
| | | return (float)sum/(n*batch); |
| | | } |
| | | |
| | | float train_network(network net, data d) |
| | | { |
| | | int batch = net.batch; |
| | | int n = d.X.rows / batch; |
| | | float *X = calloc(batch*d.X.cols, sizeof(float)); |
| | | float *y = calloc(batch*d.y.cols, sizeof(float)); |
| | | |
| | | int i; |
| | | float sum = 0; |
| | | for(i = 0; i < n; ++i){ |
| | | get_next_batch(d, batch, i*batch, X, y); |
| | | float err = train_network_datum(net, X, y); |
| | | sum += err; |
| | | } |
| | | free(X); |
| | | free(y); |
| | | return (float)sum/(n*batch); |
| | | } |
| | | |
| | | float train_network_batch(network net, data d, int n) |
| | | { |
| | | int i,j; |
| | |
| | | return (float)sum/(n*batch); |
| | | } |
| | | |
| | | float train_network_data_cpu(network net, data d, int n) |
| | | { |
| | | int batch = net.batch; |
| | | float *X = calloc(batch*d.X.cols, sizeof(float)); |
| | | float *y = calloc(batch*d.y.cols, sizeof(float)); |
| | | |
| | | int i; |
| | | float sum = 0; |
| | | for(i = 0; i < n; ++i){ |
| | | get_next_batch(d, batch, i*batch, X, y); |
| | | float err = train_network_datum(net, X, y); |
| | | sum += err; |
| | | } |
| | | free(X); |
| | | free(y); |
| | | return (float)sum/(n*batch); |
| | | } |
| | | |
| | | void train_network(network net, data d) |
| | | { |
| | | int i; |
| | | int correct = 0; |
| | | for(i = 0; i < d.X.rows; ++i){ |
| | | correct += train_network_datum(net, d.X.vals[i], d.y.vals[i]); |
| | | if(i%100 == 0){ |
| | | visualize_network(net); |
| | | cvWaitKey(10); |
| | | } |
| | | } |
| | | visualize_network(net); |
| | | cvWaitKey(100); |
| | | fprintf(stderr, "Accuracy: %f\n", (float)correct/d.X.rows); |
| | | } |
| | | |
| | | void set_learning_network(network *net, float rate, float momentum, float decay) |
| | | { |
| | | int i; |
| | |
| | | } else if(net.types[i] == DROPOUT){ |
| | | dropout_layer layer = *(dropout_layer *) net.layers[i]; |
| | | return layer.inputs; |
| | | } else if(net.types[i] == CROP){ |
| | | crop_layer layer = *(crop_layer *) net.layers[i]; |
| | | return layer.c*layer.h*layer.w; |
| | | } |
| | | else if(net.types[i] == FREEWEIGHT){ |
| | | freeweight_layer layer = *(freeweight_layer *) net.layers[i]; |
| | |
| | | softmax_layer layer = *(softmax_layer *)net.layers[i]; |
| | | return layer.inputs; |
| | | } |
| | | printf("Can't find input size\n"); |
| | | return 0; |
| | | } |
| | | |
| | |
| | | image output = get_maxpool_image(layer); |
| | | return output.h*output.w*output.c; |
| | | } |
| | | else if(net.types[i] == CROP){ |
| | | crop_layer layer = *(crop_layer *) net.layers[i]; |
| | | return layer.c*layer.crop_height*layer.crop_width; |
| | | } |
| | | else if(net.types[i] == CONNECTED){ |
| | | connected_layer layer = *(connected_layer *)net.layers[i]; |
| | | return layer.outputs; |
| | |
| | | softmax_layer layer = *(softmax_layer *)net.layers[i]; |
| | | return layer.inputs; |
| | | } |
| | | printf("Can't find output size\n"); |
| | | return 0; |
| | | } |
| | | |
| | |
| | | |
| | | float *network_predict(network net, float *input) |
| | | { |
| | | #ifdef GPU |
| | | if(gpu_index >= 0) return network_predict_gpu(net, input); |
| | | #endif |
| | | |
| | | forward_network(net, input, 0, 0); |
| | | float *out = get_network_output(net); |
| | | return out; |