19 Commits (fa0bc0be9f935d59aa83ab83f80fe2f1bf089e34)

Author SHA1 Message Date
Shad Amethyst fa0bc0be9f Lock layers
2 years ago
Shad Amethyst d40098d2ef 🔥 Refactor of NeuraTrainableLayer, split it into multiple traits
2 years ago
Shad Amethyst c1473a6d5c Add integration test for training
2 years ago
Shad Amethyst b3b97f76bd Return training and validation losses in train(), plot them out
2 years ago
Shad Amethyst a5237a8ef1 Inefficient but working forward-forward implementation
2 years ago
Shad Amethyst cb862f12cc Softmax layer
2 years ago
Shad Amethyst 0c97a65013 🎨 Remove From<f64> requirement in dropout, working bivariate layer, add builder pattern
2 years ago
Shad Amethyst 9b821b92b0 🎨 Clean up NeuraSequential
2 years ago
Shad Amethyst 2edbff860c 🔥 🚚 ♻️ Refactoring the previous layer system
2 years ago
Shad Amethyst cc7686569a Block convolution, max pooling
2 years ago
Shad Amethyst a6da11b125 Pooling layers
2 years ago
Shad Amethyst d7eb6de34e 1D convolution layer
2 years ago
Shad Amethyst 6c1d6874d7 ♻️ Implement and transition to NeuraMatrix and NeuraVector, to prevent stack overflows
2 years ago
Shad Amethyst 920bca4a48 🚚 Move files and traits around, extract stuff out of train.rs
2 years ago
Shad Amethyst b4a97694a6 🎨 Clean things up, add unit tests, add one hot layer
2 years ago
Shad Amethyst bca56a5557 Re-order arguments of neura_layer, implement softmax and normalization
2 years ago
Shad Amethyst 220c61ff6b Dropout layers
2 years ago
Shad Amethyst 8ac82e20e2 Working backpropagation :3
2 years ago
Shad Amethyst 7a6921a1c1 🔥 Semi-working training, although it seems to be only want to converge to zero
2 years ago