Rede Neural Convolucional para reconhecimento de gestos em LIBRAS (Alfabeto) Projeto 01/2019 - Ciência da Computação (Universidade Anhembi Morumbi)
-
Updated
May 18, 2020 - Python
Rede Neural Convolucional para reconhecimento de gestos em LIBRAS (Alfabeto) Projeto 01/2019 - Ciência da Computação (Universidade Anhembi Morumbi)
The project aimed to implement Deep NN / RNN based solution in order to develop flexible methods that are able to adaptively fillin, backfill, and predict time-series using a large number of heterogeneous training datasets.
Convolutional autoencoder for encoding/decoding RGB images in TensorFlow with high compression ratio
Avoiding the vanishing gradients problem by adding random noise and batch normalization
This package is a Tensorflow2/Keras implementation for Graph Attention Network embeddings and also provides a Trainable layer for Multihead Graph Attention.
Deep Learning Projects
Neural Network implemented with different Activation Functions i.e, sigmoid, relu, leaky-relu, softmax and different Optimizers i.e, Gradient Descent, AdaGrad, RMSProp, Adam. You can choose different loss functions as well i.e, cross-entropy loss, hinge-loss, mean squared error (MSE)
Creating your own custom layers(Leaky ReLU) with Keras
ReLU++: A modified ReLU activation function with enhanced performance for deep learning models.
Building Generative Adversarial Networks
Generic L-layer 'straight in Python' fully connected Neural Network implementation using numpy.
Lightweight neural network library written in ANSI-C supporting prediction and backpropagation for Convolutional- and Fully Connected neural networks
CNeuron is a simple singular neural network neuron implementation in C, designed for easy integration into C projects.
The code implements a neural network model, PricePredictor, trained on historical stock price data to predict future stock prices, visualizing the predictions alongside historical prices and calculating the average of the predicted prices.
Modifies a neural network's hyperparameters, activation functions, cost functions, and regularization methods to improve training performance and generalization.
Deep Learning concepts practice using Cifar-10 dataset
This repository summarizes the basic concepts, types and usage scenarios of activation functions in deep learning.
This repository delves into the role of activation functions in perceptron-based classification models. It features a comprehensive Jupyter notebook demonstrating different activation functions, their mathematical foundations, and their impact on model performance.
Modifies a neural network's hyperparameters, activation functions, cost functions, and regularization methods to improve training performance and generalization.
Add a description, image, and links to the leaky-relu topic page so that developers can more easily learn about it.
To associate your repository with the leaky-relu topic, visit your repo's landing page and select "manage topics."