Discussion around precision
!608 (merged) switched from float16 to float32 for the writing of the final train samples. It's clear now from the algs meeting that this doesn't affect training performance, so we should switch back to float16 to save space and improve dataloading times during training.
In a separate problem, dumping from the TDD at half precision results in reduced tagger performance. This suggests something is going wrong in the umami preprocessing (probably in the scaling). To try and mitigate, we should make sure all operations are performed at full precision, but only ever write out at half precision (since this will approx 2x i/o speeds).
Edited by Samuel Van Stroud