neuralnetlib 3.3.6
- feat: add Transformer model and layer architecture (wip)
- fix(Transformer): gradient propagation between layers
- fix(Transformer): tokenization, sequence handling and shapes
- fix(callbacks): now compatible with every model architecture
- fix_later: find why the Transformer output won't work
- ci: bump version to 3.3.6