Tag: tensorflow
-
Speed Up your Keras Sequence Pipeline
Reading Time: < 1 minute
When using tf.keras.utils.Sequence to generate batches, the data copy overhead between processed can be very high. This leads to worker processes being blocked most of the time, and decline in batch generation. A common solution is the use of shared memory to share data between processes. PyTorch uses it. With Python 3.8, you can use shared_memory…