tipe/src/cnn/main.c

133 lines
4.7 KiB
C
Raw Normal View History

2022-06-30 10:27:42 +02:00
#include <stdlib.h>
#include <stdio.h>
#include <math.h>
#include <float.h>
#include "initialisation.c"
#include "function.c"
#include "creation.c"
#include "make.c"
2022-09-16 14:53:35 +02:00
#include "include/main.h"
2022-06-30 10:27:42 +02:00
2022-07-05 08:13:25 +02:00
// Augmente les dimensions de l'image d'entrée
#define PADDING_INPUT 2
2022-06-30 10:27:42 +02:00
int will_be_drop(int dropout_prob) {
return (rand() % 100) < dropout_prob;
2022-06-30 10:27:42 +02:00
}
void write_image_in_network_32(int** image, int height, int width, float** input) {
for (int i=0; i < height+2*PADDING_INPUT; i++) {
2022-09-16 14:53:35 +02:00
for (int j=0; j < width+2*PADDING_INPUT; j++) {
if (i < PADDING_INPUT || i >= height+PADDING_INPUT || j < PADDING_INPUT || j >= width+PADDING_INPUT) {
2022-06-30 10:27:42 +02:00
input[i][j] = 0.;
}
else {
input[i][j] = (float)image[i][j] / 255.0f;
}
}
}
}
void forward_propagation(Network* network) {
int output_dim, output_depth;
float*** output;
2022-06-30 10:27:42 +02:00
for (int i=0; i < network->size-1; i++) {
2022-09-10 17:17:49 +02:00
if (network->kernel[i]->nn==NULL && network->kernel[i]->cnn!=NULL) { //CNN
output = network->input[i+1];
2022-09-12 17:56:44 +02:00
output_dim = network->width[i+1];
output_depth = network->depth[i+1];
2022-09-10 17:17:49 +02:00
make_convolution(network->input[i], network->kernel[i]->cnn, output, output_dim);
choose_apply_function_input(network->kernel[i]->activation, output, output_depth, output_dim, output_dim);
2022-06-30 10:27:42 +02:00
}
2022-09-10 17:17:49 +02:00
else if (network->kernel[i]->nn!=NULL && network->kernel[i]->cnn==NULL) { //NN
2022-09-12 17:56:44 +02:00
make_fully_connected(network->input[i][0][0], network->kernel[i]->nn, network->input[i+1][0][0], network->width[i], network->width[i+1]);
choose_apply_function_input(network->kernel[i]->activation, network->input[i+1], 1, 1, network->width[i+1]);
2022-06-30 10:27:42 +02:00
}
else { //Pooling
2022-06-30 10:27:42 +02:00
if (network->size-2==i) {
printf("Le réseau ne peut pas finir par une pooling layer");
return;
}
2022-09-10 17:17:49 +02:00
if (network->kernel[i+1]->nn!=NULL && network->kernel[i+1]->cnn==NULL) {
2022-09-12 17:56:44 +02:00
make_average_pooling_flattened(network->input[i], network->input[i+1][0][0], network->kernel[i]->activation/100, network->depth[i], network->width[i]);
choose_apply_function_input(network->kernel[i]->activation%100, network->input[i+1], 1, 1, network->width[i+1]);
2022-06-30 10:27:42 +02:00
}
2022-09-10 17:17:49 +02:00
else if (network->kernel[i+1]->nn==NULL && network->kernel[i+1]->cnn!=NULL) {
2022-09-12 17:56:44 +02:00
make_average_pooling(network->input[i], network->input[i+1], network->kernel[i]->activation/100, network->depth[i+1], network->width[i+1]);
choose_apply_function_input(network->kernel[i]->activation%100, network->input[i+1], network->depth[i+1], network->width[i+1], network->width[i+1]);
2022-05-13 15:28:45 +02:00
}
else {
2022-09-10 17:17:49 +02:00
printf("Le réseau ne peut pas contenir deux pooling layers collées");
2022-06-30 10:27:42 +02:00
return;
2022-05-13 15:28:45 +02:00
}
}
}
}
2022-06-30 10:27:42 +02:00
void backward_propagation(Network* network, float wanted_number) {
float* wanted_output = generate_wanted_output(wanted_number);
int n = network->size-1;
2022-09-12 17:56:44 +02:00
float loss = compute_cross_entropy_loss(network->input[n][0][0], wanted_output, network->width[n]);
for (int i=n; i >= 0; i--) {
2022-06-30 10:27:42 +02:00
if (i==n) {
2022-09-10 17:17:49 +02:00
if (network->kernel[i]->activation == SOFTMAX) {
2022-09-12 17:56:44 +02:00
int l2 = network->width[i]; // Taille de la dernière couche
int l1 = network->width[i-1];
for (int j=0; j < l2; j++) {
2022-06-30 10:27:42 +02:00
}
}
else {
2022-09-10 17:17:49 +02:00
printf("Erreur, seule la fonction SOFTMAX est implémentée pour la dernière couche");
2022-06-30 10:27:42 +02:00
return;
}
}
else {
2022-09-10 17:17:49 +02:00
if (network->kernel[i]->activation == SIGMOID) {
2022-06-30 10:27:42 +02:00
}
2022-09-10 17:17:49 +02:00
else if (network->kernel[i]->activation == TANH) {
2022-06-30 10:27:42 +02:00
}
2022-09-10 17:17:49 +02:00
else if (network->kernel[i]->activation == RELU) {
2022-06-30 10:27:42 +02:00
}
}
}
free(wanted_output);
}
float compute_cross_entropy_loss(float* output, float* wanted_output, int len) {
float loss=0.;
for (int i=0; i < len ; i++) {
2022-06-30 10:27:42 +02:00
if (wanted_output[i]==1) {
if (output[i]==0.) {
loss -= log(FLT_EPSILON);
}
else {
loss -= log(output[i]);
}
}
}
return loss;
}
float* generate_wanted_output(float wanted_number) {
float* wanted_output = (float*)malloc(sizeof(float)*10);
for (int i=0; i < 10; i++) {
2022-06-30 10:27:42 +02:00
if (i==wanted_number) {
wanted_output[i]=1;
}
else {
wanted_output[i]=0;
}
}
return wanted_output;
}
int main() {
2022-09-10 17:17:49 +02:00
Network* network = create_network_lenet5(0, TANH, GLOROT_NORMAL);
forward_propagation(network);
return 0;
}