-
Notifications
You must be signed in to change notification settings - Fork 26
Conversation
Causing an additional (unwanted) channel for non grayscale images; not yet discovered because only tested with mnist
…into jit-backend
Change default kwargs for mixed precision (only used if `mixed_precision` is set to `True`)
Causing an additional (unwanted) channel for non grayscale images; not yet discovered because only tested with mnist
…ve because seeking is not supported in in-memory HDF5-Files
@ORippler Tests are passing now. It would be great if you could make some time for a review |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
- change
delira/delira/training/base_trainer.py
Line 207 in cdee72f
self.save_state(os.path.join(self.save_path, "checkpoint_epoch_0")) - Look at the docstrings. Oftentimes, e.g. for basetrainer, they are wrong (still wrongly states to raise NotImplementedError. I think this still comes from before the Trainer merger.)
In general, tfeager and tfgraph seem to look quite nice. One note, we don't need to actually do the tf.convert_to_tensor, as tf accepts np.ndarrays. But that is nitpicking.
I also remarked some points for discussion
I'll have a look at all of them and either address them here or in separate PRs: Regarding the conversation to tensor type: I added this on purpose, since explicit is better than implicit and one can do some optimization (and maybe even quantization) after converting to tensors. Also this shows where the conversion is done, as tensors and numpy arrays are not completely identical regarding their syntax. Also it is important to notice, that for tf eager one will probably start gradient taping right after this. But if you prefer removing this part it's fine for me. |
This PR will add additional backends (TorchScript in #98 and TF Eager Execution in #101 ) as well as the move to the new
apex.amp
API (#97 ).Currently this PR is just open to track conversations and the Progress of the different Parts of it. It will be merged, after #66 , #97 , #98 and #101 are merged.
Progress so far:
master
apex.amp
API enabled (Move to new APEX.amp API #97 )TorchScript
Backend (including new dependency resolution) (TorchScript backend #98 ) mergedTF Eager Execution
Backend (Add TF Eager Execution Backend #101 ) mergedSkLearn
Backend (Scikit-Learn Compatibility #109 ) mergedchainer
Backend (Chainer backend #110 ) mergeddelira.training
including train-utils, trainer and experiment; 1 File / backend indelira.models
including base network and other necessary stuff (like data parallel forchainer
)closure
andprepare_batch
to each abstract networkThe following changes will be done in subsequent PRs.
MXNet
Backend (WIP: Mxnet backend #119 ) mergedOther possible backends include
If single backends of this list will be included, this should be done in separate issues/prs afterwards, since it is not yet clear which backends we want to integrate from here