Found 44 repositories(showing 30)
ApelSYN
Recognizing hand-written digits from MNIST dataset with Brain.js. Neural Networks example.
yingcongshaw
驭风计划,深度学习: 1. Softmax实现手写数字识别(MNIST); 2. 多层感知机实现手写数字识别(MNIST); 3. PyTorch实战: CIFAR图像分类(CIFAR10); 4. 脑部MRI图像分割(Kaggle-Brain MRI); 5. 滴滴出行-交通场景目标检测(MMDetection,MMCV,COCO); 6. 图像自然语言描述生成(COCO2014); 7. SRGAN图像超分辨(DIV2K)。
CAPRDZV
论文 Hinton等的论文 Matrix capsules with EM routing - Hinton, G. E., Sabour, S. and Frosst, N. (2018) Dynamic Routing Between Capsules - Sabour, S., Frosst, N. and Hinton, G.E. (2017) Transforming Auto-encoders - Hinton, G. E., Krizhevsky, A. and Wang, S. D. (2011) A parallel computation that assigns canonical object-based frames of reference. - Hinton, G.E. (1981) Shape representation in parallel systems - Hinton, G.E. (1981) Optimizing Neural Networks that Generate Images - Tijmen Tieleman’s disseration 其他论文 Capsule Network Performance on Complex Data - Xi, E., Bing, S. and Jin, Y. (2017) Accurate reconstruction of image stimuli from human fMRI based on the decoding model with capsule network architecture - Qiao, K., Zhang, C., Wang, L., Yan, B., Chen, J., Zeng, L. and Tong, L., (2018) An Optimization View on Dynamic Routing Between Capsules - Wang, D., Liu, E. (2018) CapsuleGAN: Generative Adversarial Capsule Network Ayush Jaiswal, Wael AbdAlmageed, Premkumar Natarajan. (2018) Spectral Capsule Networks - Bahadori, M. T. (2018) 博客 Max Pechyonkin的胶囊网络入门系列: 胶囊网络背后的直觉 胶囊如何工作 囊间动态路由算法 胶囊网络架构 基于TensorFlow实现胶囊网络 Debarko De的胶囊网络教程,包括注释详尽的胶囊网络实现代码 基于CUDA为胶囊网络实现TensorFlow定制操作 Jos van de Wolfshaar的文章,定制胶囊网络运算的CUDA支持 ISI新研究:胶囊生成对抗网络 用胶囊网络替换CNN作为GAN的判别网络,在MNIST数据集上取得了比卷积GAN更好的表现 Uncovering the Intuition behind Capsule Networks and Inverse Graphic Tanay Kothari的长篇教程 A Visual Representation of Capsule Connections in Dynamic Routing Between Capsules Mike Ross的胶囊网络示意图 Capsule Networks Are Shaking up AI — Here’s How to Use Them Nick Bourdakos的介绍 Capsule Networks Explained Kendrick Tan的解释 Understanding Dynamic Routing between Capsules (Capsule Networks) Jonathan Hui的教程,包括注释详尽的基于Keras的胶囊网络实现代码 Matrix capsules with EM routing Adrian Colyer关于EM路由的文章 Capsule Networks: A Glossary Sebastian Kwiatkowski的胶囊网络术语表 Overview of awesome articles 点评三篇胶囊网络教程 视频 Geoffrey Hinton’s talk: What is wrong with convolutional neural nets? - Geoffrey Hinton在MIT. Brain & Cognitive Sciences的演讲《卷积神经网络的问题在哪里?》 Capsule Networks (CapsNets) – Tutorial - “这视频棒极了。我本希望我能把胶囊解释得这么清楚。”Geoffrey Hinton Capsule networks: overview - 胶囊网络概览,包括向量和矩阵胶囊。 Overview of awesome videos 对以上3个视频的点评。 Capsule Networks: An Improvement to Convolutional Networks Siraj Raval介绍胶囊网络的视频 动态路由实现 官方实现 Sarasra/models 《Dynamic Routing Between Capsules》论文所用的代码 TensorFlow alisure-ml/CapsNet bourdakos1/capsule-networks etendue/CapsNet_TF InnerPeace-Wu/CapsNet-tensorflow jaesik817/adv_attack_capsnet jostosh/capsnet JunYeopLee/capsule-networks laodar/tf_CapsNet leoniloris/CapsNet naturomics/CapsNet-Tensorflow rrqq/CapsNet-tensorflow-jupyter thibo73800/capsnet-traffic-sign-classifier tjiang31/CapsNet winwinJJiang/capsNet-Tensorflow PyTorch acburigo/CapsNet adambielski/CapsNet-pytorch AlexHex7/CapsNet_pytorch aliasvishnu/Capsule-Networks-Notebook-MNIST andreaazzini/capsnet.pytorch cedrickchee/capsule-net-pytorch dragen1860/CapsNet-Pytorch gram-ai/capsule-networks higgsfield/Capsule-Network-Tutorial laubonghaudoi/CapsNet_guide_PyTorch leftthomas/CapsNet nishnik/CapsNet-PyTorch tonysy/CapsuleNet-PyTorch Ujjwal-9/CapsNet Keras fengwang/minimal-capsule gusgad/capsule-GAN mitiku1/Emopy-CapsNet ruslangrimov/capsnet-with-capsulewise-convolution streamride/CapsNet-keras-imdb sunxirui310/CapsNet-Keras theblackcat102/dynamic-routing-capsule-cifar XifengGuo/CapsNet-Keras XifengGuo/CapsNet-Fashion-MNIST Chainer soskek/dynamic_routing_between_capsules Torch mrkulk/Unsupervised-Capsule-Network MXNet AaronLeong/CapsNet_Mxnet GarrickLin/Capsnet.Gluon Soonhwan-Kwon/capsnet.mxnet CNTK Southworkscom/CapsNet-CNTK Lasagne DeniskaMazur/CapsNet-Lasagne Matlab yechengxi/LightCapsNet R dfalbel/capsnet JavaScript alseambusher/capsnet.js Vulcan moothyknight/CapsNet-for-Graphics-Rendering-Optimization EM路由实现 TensorFlow gyang274/capsulesEM www0wwwjs1/Matrix-Capsules-EM-Tensorflow PyTorch shzygmyx/Matrix-Capsules-pytorch 其他资源 Capsule Networks discussion Facebook讨论组 CapsNet-Tensorflow CapsNet-Tensorflow的gitter.im讨论组 Will capsule networks replace neural networks? Quora问答“胶囊网络会取代神经网络吗?” Could GANs work with Hinton’s capsule theory? Quora问答“GAN可以应用Hinton的胶囊理论吗?” Dynamic Routing Between Capsules Kyuhwan Jung对论文《Dynamic routing between Capsules》的评论(slideshare)
alik604
Many-in-one repo: The "MNIST" of Brain Digits - Thought classification, Motor movement classification, 3D cancer detection, and Covid detection
BrainJS
Brain.js mnist demo
averyzgriffin
No description available
kakaobrain
brain cloud hyperopt example (mnist)
pinae
Slightly extended MNIST example from brainstorm
ABazshoushtari
Image Classification projects on MNIST dataset and brain tumors images using PyTorch
hasithbasnayake
A computer vision model to classify Fashion-MNIST using spiking neural network architecture and a brain-inspired learning rule
biancasama
Le Wagon batch 941 project
MinerNJ
A Neural Network in Javascript using Brain.JS on the MNIST dataset.
This is a simple code to test the libs Brain.JS and Mnist.
solanki1993
Brain inspired AI-Boltzmann machine trained on the MNIST data and temporal difference learning model for navigating Morris water-maze task
Real-time embedded brain–machine interface on nRF5340 DK: displays on ILI9488 TFT, runs MNIST & Sleep-EDF EEG ML models via Edge Impulse. Quantized models enable real-time inference; external QSPI flash explored to reduce RAM for large models.
SamboDutta20
Neural network is a supervised classification model which mimics the structure of human brain. They are a model of interconnected nodes or neurons where one arrow denotes how the output from one node becomes an input for the next. There are several models of neural networks. The most used one are \textit{feed-forward neural networks} and \textit{recurrent neural networks}. In this project we mainly work with various types of \textit{feed-forward neural networks}. Using the MNIST-Fashion data set we show how the results vary for one-hidden layer NN, multi-layer NN and convolution NN.
ARKAtheVLOGGER
Introduction: Handwritten digit recognition using MNIST dataset is a major project made with the help of Neural Network. It basically detects the scanned images of handwritten digits. We have taken this a step further where our handwritten digit recognition system not only detects scanned images of handwritten digits but also allows writing digits on the screen with the help of an integrated GUI for recognition. Approach: We will approach this project by using a three-layered Neural Network. The input layer: It distributes the features of our examples to the next layer for calculation of activations of the next layer. The hidden layer: They are made of hidden units called activations providing nonlinear ties for the network. A number of hidden layers can vary according to our requirements. The output layer: The nodes here are called output units. It provides us with the final prediction of the Neural Network on the basis of which final predictions can be made. A neural network is a model inspired by how the brain works. It consists of multiple layers having many activations, this activation resembles neurons of our brain. A neural network tries to learn a set of parameters in a set of data which could help to recognize the underlying relationships. Neural networks can adapt to changing input; so the network generates the best possible result without needing to redesign the output criteria. Methodology: We have implemented a Neural Network with 1 hidden layer having 100 activation units (excluding bias units). The data is loaded from a .mat file, features(X) and labels(y) were extracted. Then features are divided by 255 to rescale them into a range of [0,1] to avoid overflow during computation. Data is split up into 60,000 training and 10,000 testing examples. Feedforward is performed with the training set for calculating the hypothesis and then backpropagation is done in order to reduce the error between the layers. The regularization parameter lambda is set to 0.1 to address the problem of overfitting. Optimizer is run for 70 iterations to find the best fit model.
vaezyq
No description available
geopopos
Mnist but for your brain
vaezyq
No description available
ducan114
No description available
AdamGuesmia
No description available
Sepehrshirani
No description available
neuroph12
Brain MNIST: Digit classification for EEG signals
daruai
MNIST digit classification from EEG records
MatveiZhartun
MNIST Digit Recognition built on Brain.JS
Robin-WZQ
Brainlike learning system for MNist.(企图通过类脑神经元学习的方式,应用到MNist数据集上)
KyleSeaman201
No description available
jakeweholt
No description available
adplays21
Human Brain Detection on Fashion MNIST Dataset with TensorFlow