- New layer normalized GRU layer (
LayerNormGRU
). - New IndRNN layer.
- CPU support for all PyTorch layers.
- Support for building PyTorch API on Windows.
- Added
state
argument to PyTorch layers to specify initial state. - Added weight transforms to TensorFlow API (see docs for details).
- Added
get_weights
method to extract weights from RNN layers (TensorFlow). - Added
to_native_weights
andfrom_native_weights
to PyTorch API forLSTM
andGRU
layers. - Validation tests to check for correctness.
- Performance improvements to GRU layer.
- BREAKING CHANGE: PyTorch layers default to CPU instead of GPU.
- BREAKING CHANGE:
h
must not be transposed before passing it togru::BackwardPass::Iterate
.
- Multi-GPU training with TensorFlow caused by invalid sharing of
cublasHandle_t
.
- PyTorch support.
- New layer normalized LSTM layer (
LayerNormLSTM
). - New fused layer normalization layer.
- Occasional uninitialized memory use in TensorFlow LSTM implementation.
- New time-fused API for LSTM (
lstm::ForwardPass::Run
,lstm::BackwardPass::Run
). - Benchmarking code to evaluate the performance of an implementation.
- Performance improvements to existing iterative LSTM API.
- BREAKING CHANGE:
h
must not be transposed before passing it tolstm::BackwardPass::Iterate
. - BREAKING CHANGE:
dv
does not need to be allocated andv
must be passed instead tolstm::BackwardPass::Iterate
.
- Initial release of Haste.