121 lines
4.4 KiB
C
Raw Normal View History

2022-06-30 10:27:42 +02:00
#include <stdlib.h>
#include <stdio.h>
#include <math.h>
#include <float.h>
2022-07-05 08:13:25 +02:00
#include "function.h"
#include "make.h"
2022-06-30 10:27:42 +02:00
#include "cnn.h"
2022-07-05 08:13:25 +02:00
// Augmente les dimensions de l'image d'entrée
#define PADING_INPUT 2
2022-06-30 10:27:42 +02:00
int will_be_drop(int dropout_prob) {
return (rand() % 100)<dropout_prob;
}
void write_image_in_newtork_32(int** image, int height, int width, float** input) {
2022-05-13 15:28:45 +02:00
for (int i=0; i < height+2*PADING_INPUT; i++) {
for (int j=PADING_INPUT; j < width+2*PADING_INPUT; j++) {
2022-06-30 10:27:42 +02:00
if (i<PADING_INPUT || i>height+PADING_INPUT || j<PADING_INPUT || j>width+PADING_INPUT) {
input[i][j] = 0.;
}
else {
input[i][j] = (float)image[i][j] / 255.0f;
}
}
}
}
void forward_propagation(Network* network) {
for (int i=0; i < network->size-1; i++) {
if (network->kernel[i].nn==NULL && network->kernel[i].cnn!=NULL) {
make_convolution(network->input[i], network->kernel[i].cnn, network->input[i+1], network->dim[i+1][0]);
choose_apply_function_input(network->kernel[i].activation, network->input[i+1], network->dim[i+1][1], network->dim[i+1][0], network->dim[i+1][0]);
}
else if (network->kernel[i].nn!=NULL && network->kernel[i].cnn==NULL) {
make_fully_connected(network->input[i][0][0], network->kernel[i].nn, network->input[i+1][0][0], network->dim[i][0], network->dim[i+1][0]);
choose_apply_function_input(network->kernel[i].activation, network->input[i+1], 1, 1, network->dim[i+1][0]);
}
else {
if (network->size-2==i) {
printf("Le réseau ne peut pas finir par une pooling layer");
return;
}
if (network->kernel[i+1].nn!=NULL && network->kernel[i+1].cnn==NULL) {
make_average_pooling_flattened(network->input[i], network->input[i+1][0][0], network->kernel[i].activation/100, network->dim[i][1], network->dim[i][0]);
choose_apply_function_input(network->kernel[i].activation%100, network->input[i+1], 1, 1, network->dim[i+1][0]);
}
else if (network->kernel[i+1].nn==NULL && network->kernel[i+1].cnn!=NULL) {
make_average_pooling(network->input[i], network->input[i+1], network->kernel[i].activation/100, network->dim[i+1][1], network->dim[i+1][0]);
choose_apply_function_input(network->kernel[i].activation%100, network->input[i+1], network->dim[i+1][1], network->dim[i+1][0], network->dim[i+1][0]);
2022-05-13 15:28:45 +02:00
}
else {
2022-06-30 10:27:42 +02:00
printf("Le réseau ne peut pas contenir deux poolings layers collées");
return;
2022-05-13 15:28:45 +02:00
}
}
}
}
2022-06-30 10:27:42 +02:00
void backward_propagation(Network* network, float wanted_number) {
float* wanted_output = generate_wanted_output(wanted_number);
int n = network->size-1;
float loss = compute_cross_entropy_loss(network->input[n][0][0], wanted_output, network->dim[n][0]);
int i, j;
for (i=n; i>=0; i--) {
if (i==n) {
if (network->kernel[i].activation == SOFTMAX) {
int l2 = network->dim[i][0]; // Taille de la dernière couche
int l1 = network->dim[i-1][0];
for (j=0; j<l2; j++) {
}
}
else {
printf("Erreur, seule la fonction softmax est implémentée pour la dernière couche");
return;
}
}
else {
if (network->kernel[i].activation == SIGMOID) {
}
else if (network->kernel[i].activation == TANH) {
}
else if (network->kernel[i].activation == RELU) {
}
}
}
free(wanted_output);
}
float compute_cross_entropy_loss(float* output, float* wanted_output, int len) {
float loss=0.;
for (int i=0; i<len ; i++) {
if (wanted_output[i]==1) {
if (output[i]==0.) {
loss -= log(FLT_EPSILON);
}
else {
loss -= log(output[i]);
}
}
}
return loss;
}
float* generate_wanted_output(float wanted_number) {
float* wanted_output = malloc(sizeof(float)*10);
for (int i=0; i<10; i++) {
if (i==wanted_number) {
wanted_output[i]=1;
}
else {
wanted_output[i]=0;
}
}
return wanted_output;
}