diff --git a/cv/detection/ssd/tensorflow/dataset/convert_voc_sample_tfrecords.py b/cv/detection/ssd/tensorflow/dataset/convert_voc_sample_tfrecords.py index 6fe35a4a67467fc5b0a8aea2f237548915722dee..8231e7a3d92457e5ac67ddaef5cefbb30dc9d6cb 100644 --- a/cv/detection/ssd/tensorflow/dataset/convert_voc_sample_tfrecords.py +++ b/cv/detection/ssd/tensorflow/dataset/convert_voc_sample_tfrecords.py @@ -326,7 +326,7 @@ def _process_image_files(name, directory, all_records, num_shards): num_shards: integer number of shards for this data set. """ # Break all images into batches with a [ranges[i][0], ranges[i][1]]. - spacing = np.linspace(0, len(all_records), FLAGS.num_threads + 1).astype(np.int) + spacing = np.linspace(0, len(all_records), FLAGS.num_threads + 1).astype(np.int32) ranges = [] threads = [] for i in range(len(spacing) - 1):