repo
stringlengths
2
99
file
stringlengths
13
225
code
stringlengths
0
18.3M
file_length
int64
0
18.3M
avg_line_length
float64
0
1.36M
max_line_length
int64
0
4.26M
extension_type
stringclasses
1 value
models
models-master/official/legacy/image_classification/efficientnet/efficientnet_model.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Contains definitions for EfficientNet model. [1] Mingxing Tan, Quoc V. Le EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. ICML'19, https://arxiv.org/abs/1905.11946 """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import dataclasses import math from typing import Any, Dict, Optional, Text, Tuple from absl import logging import tensorflow as tf from official.legacy.image_classification import preprocessing from official.legacy.image_classification.efficientnet import common_modules from official.modeling import tf_utils from official.modeling.hyperparams import base_config @dataclasses.dataclass class BlockConfig(base_config.Config): """Config for a single MB Conv Block.""" input_filters: int = 0 output_filters: int = 0 kernel_size: int = 3 num_repeat: int = 1 expand_ratio: int = 1 strides: Tuple[int, int] = (1, 1) se_ratio: Optional[float] = None id_skip: bool = True fused_conv: bool = False conv_type: str = 'depthwise' @dataclasses.dataclass class ModelConfig(base_config.Config): """Default Config for Efficientnet-B0.""" width_coefficient: float = 1.0 depth_coefficient: float = 1.0 resolution: int = 224 dropout_rate: float = 0.2 blocks: Tuple[BlockConfig, ...] = ( # (input_filters, output_filters, kernel_size, num_repeat, # expand_ratio, strides, se_ratio) # pylint: disable=bad-whitespace BlockConfig.from_args(32, 16, 3, 1, 1, (1, 1), 0.25), BlockConfig.from_args(16, 24, 3, 2, 6, (2, 2), 0.25), BlockConfig.from_args(24, 40, 5, 2, 6, (2, 2), 0.25), BlockConfig.from_args(40, 80, 3, 3, 6, (2, 2), 0.25), BlockConfig.from_args(80, 112, 5, 3, 6, (1, 1), 0.25), BlockConfig.from_args(112, 192, 5, 4, 6, (2, 2), 0.25), BlockConfig.from_args(192, 320, 3, 1, 6, (1, 1), 0.25), # pylint: enable=bad-whitespace ) stem_base_filters: int = 32 top_base_filters: int = 1280 activation: str = 'simple_swish' batch_norm: str = 'default' bn_momentum: float = 0.99 bn_epsilon: float = 1e-3 # While the original implementation used a weight decay of 1e-5, # tf.nn.l2_loss divides it by 2, so we halve this to compensate in Keras weight_decay: float = 5e-6 drop_connect_rate: float = 0.2 depth_divisor: int = 8 min_depth: Optional[int] = None use_se: bool = True input_channels: int = 3 num_classes: int = 1000 model_name: str = 'efficientnet' rescale_input: bool = True data_format: str = 'channels_last' dtype: str = 'float32' MODEL_CONFIGS = { # (width, depth, resolution, dropout) 'efficientnet-b0': ModelConfig.from_args(1.0, 1.0, 224, 0.2), 'efficientnet-b1': ModelConfig.from_args(1.0, 1.1, 240, 0.2), 'efficientnet-b2': ModelConfig.from_args(1.1, 1.2, 260, 0.3), 'efficientnet-b3': ModelConfig.from_args(1.2, 1.4, 300, 0.3), 'efficientnet-b4': ModelConfig.from_args(1.4, 1.8, 380, 0.4), 'efficientnet-b5': ModelConfig.from_args(1.6, 2.2, 456, 0.4), 'efficientnet-b6': ModelConfig.from_args(1.8, 2.6, 528, 0.5), 'efficientnet-b7': ModelConfig.from_args(2.0, 3.1, 600, 0.5), 'efficientnet-b8': ModelConfig.from_args(2.2, 3.6, 672, 0.5), 'efficientnet-l2': ModelConfig.from_args(4.3, 5.3, 800, 0.5), } CONV_KERNEL_INITIALIZER = { 'class_name': 'VarianceScaling', 'config': { 'scale': 2.0, 'mode': 'fan_out', # Note: this is a truncated normal distribution 'distribution': 'normal' } } DENSE_KERNEL_INITIALIZER = { 'class_name': 'VarianceScaling', 'config': { 'scale': 1 / 3.0, 'mode': 'fan_out', 'distribution': 'uniform' } } def round_filters(filters: int, config: ModelConfig) -> int: """Round number of filters based on width coefficient.""" width_coefficient = config.width_coefficient min_depth = config.min_depth divisor = config.depth_divisor orig_filters = filters if not width_coefficient: return filters filters *= width_coefficient min_depth = min_depth or divisor new_filters = max(min_depth, int(filters + divisor / 2) // divisor * divisor) # Make sure that round down does not go down by more than 10%. if new_filters < 0.9 * filters: new_filters += divisor logging.info('round_filter input=%s output=%s', orig_filters, new_filters) return int(new_filters) def round_repeats(repeats: int, depth_coefficient: float) -> int: """Round number of repeats based on depth coefficient.""" return int(math.ceil(depth_coefficient * repeats)) def conv2d_block(inputs: tf.Tensor, conv_filters: Optional[int], config: ModelConfig, kernel_size: Any = (1, 1), strides: Any = (1, 1), use_batch_norm: bool = True, use_bias: bool = False, activation: Optional[Any] = None, depthwise: bool = False, name: Optional[Text] = None): """A conv2d followed by batch norm and an activation.""" batch_norm = common_modules.get_batch_norm(config.batch_norm) bn_momentum = config.bn_momentum bn_epsilon = config.bn_epsilon data_format = tf.keras.backend.image_data_format() weight_decay = config.weight_decay name = name or '' # Collect args based on what kind of conv2d block is desired init_kwargs = { 'kernel_size': kernel_size, 'strides': strides, 'use_bias': use_bias, 'padding': 'same', 'name': name + '_conv2d', 'kernel_regularizer': tf.keras.regularizers.l2(weight_decay), 'bias_regularizer': tf.keras.regularizers.l2(weight_decay), } if depthwise: conv2d = tf.keras.layers.DepthwiseConv2D init_kwargs.update({'depthwise_initializer': CONV_KERNEL_INITIALIZER}) else: conv2d = tf.keras.layers.Conv2D init_kwargs.update({ 'filters': conv_filters, 'kernel_initializer': CONV_KERNEL_INITIALIZER }) x = conv2d(**init_kwargs)(inputs) if use_batch_norm: bn_axis = 1 if data_format == 'channels_first' else -1 x = batch_norm( axis=bn_axis, momentum=bn_momentum, epsilon=bn_epsilon, name=name + '_bn')( x) if activation is not None: x = tf.keras.layers.Activation(activation, name=name + '_activation')(x) return x def mb_conv_block(inputs: tf.Tensor, block: BlockConfig, config: ModelConfig, prefix: Optional[Text] = None): """Mobile Inverted Residual Bottleneck. Args: inputs: the Keras input to the block block: BlockConfig, arguments to create a Block config: ModelConfig, a set of model parameters prefix: prefix for naming all layers Returns: the output of the block """ use_se = config.use_se activation = tf_utils.get_activation(config.activation) drop_connect_rate = config.drop_connect_rate data_format = tf.keras.backend.image_data_format() use_depthwise = block.conv_type != 'no_depthwise' prefix = prefix or '' filters = block.input_filters * block.expand_ratio x = inputs if block.fused_conv: # If we use fused mbconv, skip expansion and use regular conv. x = conv2d_block( x, filters, config, kernel_size=block.kernel_size, strides=block.strides, activation=activation, name=prefix + 'fused') else: if block.expand_ratio != 1: # Expansion phase kernel_size = (1, 1) if use_depthwise else (3, 3) x = conv2d_block( x, filters, config, kernel_size=kernel_size, activation=activation, name=prefix + 'expand') # Depthwise Convolution if use_depthwise: x = conv2d_block( x, conv_filters=None, config=config, kernel_size=block.kernel_size, strides=block.strides, activation=activation, depthwise=True, name=prefix + 'depthwise') # Squeeze and Excitation phase if use_se: assert block.se_ratio is not None assert 0 < block.se_ratio <= 1 num_reduced_filters = max(1, int(block.input_filters * block.se_ratio)) if data_format == 'channels_first': se_shape = (filters, 1, 1) else: se_shape = (1, 1, filters) se = tf.keras.layers.GlobalAveragePooling2D(name=prefix + 'se_squeeze')(x) se = tf.keras.layers.Reshape(se_shape, name=prefix + 'se_reshape')(se) se = conv2d_block( se, num_reduced_filters, config, use_bias=True, use_batch_norm=False, activation=activation, name=prefix + 'se_reduce') se = conv2d_block( se, filters, config, use_bias=True, use_batch_norm=False, activation='sigmoid', name=prefix + 'se_expand') x = tf.keras.layers.multiply([x, se], name=prefix + 'se_excite') # Output phase x = conv2d_block( x, block.output_filters, config, activation=None, name=prefix + 'project') # Add identity so that quantization-aware training can insert quantization # ops correctly. x = tf.keras.layers.Activation( tf_utils.get_activation('identity'), name=prefix + 'id')( x) if (block.id_skip and all(s == 1 for s in block.strides) and block.input_filters == block.output_filters): if drop_connect_rate and drop_connect_rate > 0: # Apply dropconnect # The only difference between dropout and dropconnect in TF is scaling by # drop_connect_rate during training. See: # https://github.com/keras-team/keras/pull/9898#issuecomment-380577612 x = tf.keras.layers.Dropout( drop_connect_rate, noise_shape=(None, 1, 1, 1), name=prefix + 'drop')( x) x = tf.keras.layers.add([x, inputs], name=prefix + 'add') return x def efficientnet(image_input: tf.keras.layers.Input, config: ModelConfig): # pytype: disable=invalid-annotation # typed-keras """Creates an EfficientNet graph given the model parameters. This function is wrapped by the `EfficientNet` class to make a tf.keras.Model. Args: image_input: the input batch of images config: the model config Returns: the output of efficientnet """ depth_coefficient = config.depth_coefficient blocks = config.blocks stem_base_filters = config.stem_base_filters top_base_filters = config.top_base_filters activation = tf_utils.get_activation(config.activation) dropout_rate = config.dropout_rate drop_connect_rate = config.drop_connect_rate num_classes = config.num_classes input_channels = config.input_channels rescale_input = config.rescale_input data_format = tf.keras.backend.image_data_format() dtype = config.dtype weight_decay = config.weight_decay x = image_input if data_format == 'channels_first': # Happens on GPU/TPU if available. x = tf.keras.layers.Permute((3, 1, 2))(x) if rescale_input: x = preprocessing.normalize_images( x, num_channels=input_channels, dtype=dtype, data_format=data_format) # Build stem x = conv2d_block( x, round_filters(stem_base_filters, config), config, kernel_size=[3, 3], strides=[2, 2], activation=activation, name='stem') # Build blocks num_blocks_total = sum( round_repeats(block.num_repeat, depth_coefficient) for block in blocks) block_num = 0 for stack_idx, block in enumerate(blocks): assert block.num_repeat > 0 # Update block input and output filters based on depth multiplier block = block.replace( input_filters=round_filters(block.input_filters, config), output_filters=round_filters(block.output_filters, config), num_repeat=round_repeats(block.num_repeat, depth_coefficient)) # The first block needs to take care of stride and filter size increase drop_rate = drop_connect_rate * float(block_num) / num_blocks_total config = config.replace(drop_connect_rate=drop_rate) block_prefix = 'stack_{}/block_0/'.format(stack_idx) x = mb_conv_block(x, block, config, block_prefix) block_num += 1 if block.num_repeat > 1: block = block.replace(input_filters=block.output_filters, strides=[1, 1]) for block_idx in range(block.num_repeat - 1): drop_rate = drop_connect_rate * float(block_num) / num_blocks_total config = config.replace(drop_connect_rate=drop_rate) block_prefix = 'stack_{}/block_{}/'.format(stack_idx, block_idx + 1) x = mb_conv_block(x, block, config, prefix=block_prefix) block_num += 1 # Build top x = conv2d_block( x, round_filters(top_base_filters, config), config, activation=activation, name='top') # Build classifier x = tf.keras.layers.GlobalAveragePooling2D(name='top_pool')(x) if dropout_rate and dropout_rate > 0: x = tf.keras.layers.Dropout(dropout_rate, name='top_dropout')(x) x = tf.keras.layers.Dense( num_classes, kernel_initializer=DENSE_KERNEL_INITIALIZER, kernel_regularizer=tf.keras.regularizers.l2(weight_decay), bias_regularizer=tf.keras.regularizers.l2(weight_decay), name='logits')( x) x = tf.keras.layers.Activation('softmax', name='probs')(x) return x class EfficientNet(tf.keras.Model): """Wrapper class for an EfficientNet Keras model. Contains helper methods to build, manage, and save metadata about the model. """ def __init__(self, config: Optional[ModelConfig] = None, overrides: Optional[Dict[Text, Any]] = None): """Create an EfficientNet model. Args: config: (optional) the main model parameters to create the model overrides: (optional) a dict containing keys that can override config """ overrides = overrides or {} config = config or ModelConfig() self.config = config.replace(**overrides) input_channels = self.config.input_channels model_name = self.config.model_name input_shape = (None, None, input_channels) # Should handle any size image image_input = tf.keras.layers.Input(shape=input_shape) output = efficientnet(image_input, self.config) # Cast to float32 in case we have a different model dtype output = tf.cast(output, tf.float32) logging.info('Building model %s with params %s', model_name, self.config) super(EfficientNet, self).__init__( inputs=image_input, outputs=output, name=model_name) @classmethod def from_name(cls, model_name: Text, model_weights_path: Optional[Text] = None, weights_format: Text = 'saved_model', overrides: Optional[Dict[Text, Any]] = None): """Construct an EfficientNet model from a predefined model name. E.g., `EfficientNet.from_name('efficientnet-b0')`. Args: model_name: the predefined model name model_weights_path: the path to the weights (h5 file or saved model dir) weights_format: the model weights format. One of 'saved_model', 'h5', or 'checkpoint'. overrides: (optional) a dict containing keys that can override config Returns: A constructed EfficientNet instance. """ model_configs = dict(MODEL_CONFIGS) overrides = dict(overrides) if overrides else {} # One can define their own custom models if necessary model_configs.update(overrides.pop('model_config', {})) if model_name not in model_configs: raise ValueError('Unknown model name {}'.format(model_name)) config = model_configs[model_name] model = cls(config=config, overrides=overrides) if model_weights_path: common_modules.load_weights( model, model_weights_path, weights_format=weights_format) return model
16,417
32.100806
127
py
models
models-master/official/legacy/image_classification/efficientnet/efficientnet_config.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Configuration definitions for EfficientNet losses, learning rates, and optimizers.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import dataclasses from official.legacy.image_classification.configs import base_configs from official.modeling.hyperparams import base_config @dataclasses.dataclass class EfficientNetModelConfig(base_configs.ModelConfig): """Configuration for the EfficientNet model. This configuration will default to settings used for training efficientnet-b0 on a v3-8 TPU on ImageNet. Attributes: name: The name of the model. Defaults to 'EfficientNet'. num_classes: The number of classes in the model. model_params: A dictionary that represents the parameters of the EfficientNet model. These will be passed in to the "from_name" function. loss: The configuration for loss. Defaults to a categorical cross entropy implementation. optimizer: The configuration for optimizations. Defaults to an RMSProp configuration. learning_rate: The configuration for learning rate. Defaults to an exponential configuration. """ name: str = 'EfficientNet' num_classes: int = 1000 model_params: base_config.Config = dataclasses.field( default_factory=lambda: { 'model_name': 'efficientnet-b0', 'model_weights_path': '', 'weights_format': 'saved_model', 'overrides': { 'batch_norm': 'default', 'rescale_input': True, 'num_classes': 1000, 'activation': 'swish', 'dtype': 'float32', } }) loss: base_configs.LossConfig = dataclasses.field( default_factory=lambda: base_configs.LossConfig( # pylint: disable=g-long-lambda name='categorical_crossentropy', label_smoothing=0.1 ) ) optimizer: base_configs.OptimizerConfig = dataclasses.field( default_factory=lambda: base_configs.OptimizerConfig( # pylint: disable=g-long-lambda name='rmsprop', decay=0.9, epsilon=0.001, momentum=0.9, moving_average_decay=None, ) ) learning_rate: base_configs.LearningRateConfig = dataclasses.field( default_factory=lambda: base_configs.LearningRateConfig( # pylint: disable=g-long-lambda name='exponential', initial_lr=0.008, decay_epochs=2.4, decay_rate=0.97, warmup_epochs=5, scale_by_batch_size=1.0 / 128.0, staircase=True, ) )
3,169
37.192771
95
py
models
models-master/official/legacy/image_classification/efficientnet/tfhub_export.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """A script to export TF-Hub SavedModel.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import os from absl import app from absl import flags import tensorflow as tf from official.legacy.image_classification.efficientnet import efficientnet_model FLAGS = flags.FLAGS flags.DEFINE_string("model_name", None, "EfficientNet model name.") flags.DEFINE_string("model_path", None, "File path to TF model checkpoint.") flags.DEFINE_string("export_path", None, "TF-Hub SavedModel destination path to export.") def export_tfhub(model_path, hub_destination, model_name): """Restores a tf.keras.Model and saves for TF-Hub.""" model_configs = dict(efficientnet_model.MODEL_CONFIGS) config = model_configs[model_name] image_input = tf.keras.layers.Input( shape=(None, None, 3), name="image_input", dtype=tf.float32) x = image_input * 255.0 outputs = efficientnet_model.efficientnet(x, config) hub_model = tf.keras.Model(image_input, outputs) ckpt = tf.train.Checkpoint(model=hub_model) ckpt.restore(model_path).assert_existing_objects_matched() hub_model.save( os.path.join(hub_destination, "classification"), include_optimizer=False) feature_vector_output = hub_model.get_layer(name="top_pool").get_output_at(0) hub_model2 = tf.keras.Model(image_input, feature_vector_output) hub_model2.save( os.path.join(hub_destination, "feature-vector"), include_optimizer=False) def main(argv): if len(argv) > 1: raise app.UsageError("Too many command-line arguments.") export_tfhub(FLAGS.model_path, FLAGS.export_path, FLAGS.model_name) if __name__ == "__main__": app.run(main)
2,319
33.117647
80
py
models
models-master/official/legacy/image_classification/efficientnet/common_modules.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Common modeling utilities.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from typing import Optional, Text import numpy as np import tensorflow as tf import tensorflow.compat.v1 as tf1 from tensorflow.python.tpu import tpu_function @tf.keras.utils.register_keras_serializable(package='Vision') class TpuBatchNormalization(tf.keras.layers.BatchNormalization): """Cross replica batch normalization.""" def __init__(self, fused: Optional[bool] = False, **kwargs): if fused in (True, None): raise ValueError('TpuBatchNormalization does not support fused=True.') super(TpuBatchNormalization, self).__init__(fused=fused, **kwargs) def _cross_replica_average(self, t: tf.Tensor, num_shards_per_group: int): """Calculates the average value of input tensor across TPU replicas.""" num_shards = tpu_function.get_tpu_context().number_of_shards group_assignment = None if num_shards_per_group > 1: if num_shards % num_shards_per_group != 0: raise ValueError( 'num_shards: %d mod shards_per_group: %d, should be 0' % (num_shards, num_shards_per_group)) num_groups = num_shards // num_shards_per_group group_assignment = [[ x for x in range(num_shards) if x // num_shards_per_group == y ] for y in range(num_groups)] return tf1.tpu.cross_replica_sum(t, group_assignment) / tf.cast( num_shards_per_group, t.dtype) def _moments(self, inputs: tf.Tensor, reduction_axes: int, keep_dims: int): """Compute the mean and variance: it overrides the original _moments.""" shard_mean, shard_variance = super(TpuBatchNormalization, self)._moments( inputs, reduction_axes, keep_dims=keep_dims) num_shards = tpu_function.get_tpu_context().number_of_shards or 1 if num_shards <= 8: # Skip cross_replica for 2x2 or smaller slices. num_shards_per_group = 1 else: num_shards_per_group = max(8, num_shards // 8) if num_shards_per_group > 1: # Compute variance using: Var[X]= E[X^2] - E[X]^2. shard_square_of_mean = tf.math.square(shard_mean) shard_mean_of_square = shard_variance + shard_square_of_mean group_mean = self._cross_replica_average(shard_mean, num_shards_per_group) group_mean_of_square = self._cross_replica_average( shard_mean_of_square, num_shards_per_group) group_variance = group_mean_of_square - tf.math.square(group_mean) return (group_mean, group_variance) else: return (shard_mean, shard_variance) def get_batch_norm(batch_norm_type: Text) -> tf.keras.layers.BatchNormalization: """A helper to create a batch normalization getter. Args: batch_norm_type: The type of batch normalization layer implementation. `tpu` will use `TpuBatchNormalization`. Returns: An instance of `tf.keras.layers.BatchNormalization`. """ if batch_norm_type == 'tpu': return TpuBatchNormalization return tf.keras.layers.BatchNormalization # pytype: disable=bad-return-type # typed-keras def count_params(model, trainable_only=True): """Returns the count of all model parameters, or just trainable ones.""" if not trainable_only: return model.count_params() else: return int( np.sum([ tf.keras.backend.count_params(p) for p in model.trainable_weights ])) def load_weights(model: tf.keras.Model, model_weights_path: Text, weights_format: Text = 'saved_model'): """Load model weights from the given file path. Args: model: the model to load weights into model_weights_path: the path of the model weights weights_format: the model weights format. One of 'saved_model', 'h5', or 'checkpoint'. """ if weights_format == 'saved_model': loaded_model = tf.keras.models.load_model(model_weights_path) model.set_weights(loaded_model.get_weights()) else: model.load_weights(model_weights_path)
4,609
38.401709
93
py
models
models-master/official/legacy/image_classification/efficientnet/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/image_classification/configs/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/image_classification/configs/base_configs.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Definitions for high level configuration groups..""" import dataclasses from typing import Any, List, Optional from official.core import config_definitions from official.modeling import hyperparams RuntimeConfig = config_definitions.RuntimeConfig @dataclasses.dataclass class TensorBoardConfig(hyperparams.Config): """Configuration for TensorBoard. Attributes: track_lr: Whether or not to track the learning rate in TensorBoard. Defaults to True. write_model_weights: Whether or not to write the model weights as images in TensorBoard. Defaults to False. """ track_lr: bool = True write_model_weights: bool = False @dataclasses.dataclass class CallbacksConfig(hyperparams.Config): """Configuration for Callbacks. Attributes: enable_checkpoint_and_export: Whether or not to enable checkpoints as a Callback. Defaults to True. enable_backup_and_restore: Whether or not to add BackupAndRestore callback. Defaults to True. enable_tensorboard: Whether or not to enable TensorBoard as a Callback. Defaults to True. enable_time_history: Whether or not to enable TimeHistory Callbacks. Defaults to True. """ enable_checkpoint_and_export: bool = True enable_backup_and_restore: bool = False enable_tensorboard: bool = True enable_time_history: bool = True @dataclasses.dataclass class ExportConfig(hyperparams.Config): """Configuration for exports. Attributes: checkpoint: the path to the checkpoint to export. destination: the path to where the checkpoint should be exported. """ checkpoint: str = None destination: str = None @dataclasses.dataclass class MetricsConfig(hyperparams.Config): """Configuration for Metrics. Attributes: accuracy: Whether or not to track accuracy as a Callback. Defaults to None. top_5: Whether or not to track top_5_accuracy as a Callback. Defaults to None. """ accuracy: bool = None top_5: bool = None @dataclasses.dataclass class TimeHistoryConfig(hyperparams.Config): """Configuration for the TimeHistory callback. Attributes: log_steps: Interval of steps between logging of batch level stats. """ log_steps: int = None @dataclasses.dataclass class TrainConfig(hyperparams.Config): """Configuration for training. Attributes: resume_checkpoint: Whether or not to enable load checkpoint loading. Defaults to None. epochs: The number of training epochs to run. Defaults to None. steps: The number of steps to run per epoch. If None, then this will be inferred based on the number of images and batch size. Defaults to None. callbacks: An instance of CallbacksConfig. metrics: An instance of MetricsConfig. tensorboard: An instance of TensorBoardConfig. set_epoch_loop: Whether or not to set `steps_per_execution` to equal the number of training steps in `model.compile`. This reduces the number of callbacks run per epoch which significantly improves end-to-end TPU training time. """ resume_checkpoint: bool = None epochs: int = None steps: int = None callbacks: CallbacksConfig = dataclasses.field( default_factory=CallbacksConfig ) metrics: MetricsConfig = None tensorboard: TensorBoardConfig = dataclasses.field( default_factory=TensorBoardConfig ) time_history: TimeHistoryConfig = dataclasses.field( default_factory=TimeHistoryConfig ) set_epoch_loop: bool = False @dataclasses.dataclass class EvalConfig(hyperparams.Config): """Configuration for evaluation. Attributes: epochs_between_evals: The number of train epochs to run between evaluations. Defaults to None. steps: The number of eval steps to run during evaluation. If None, this will be inferred based on the number of images and batch size. Defaults to None. skip_eval: Whether or not to skip evaluation. """ epochs_between_evals: int = None steps: int = None skip_eval: bool = False @dataclasses.dataclass class LossConfig(hyperparams.Config): """Configuration for Loss. Attributes: name: The name of the loss. Defaults to None. label_smoothing: Whether or not to apply label smoothing to the loss. This only applies to 'categorical_cross_entropy'. """ name: str = None label_smoothing: float = None @dataclasses.dataclass class OptimizerConfig(hyperparams.Config): """Configuration for Optimizers. Attributes: name: The name of the optimizer. Defaults to None. decay: Decay or rho, discounting factor for gradient. Defaults to None. epsilon: Small value used to avoid 0 denominator. Defaults to None. momentum: Plain momentum constant. Defaults to None. nesterov: Whether or not to apply Nesterov momentum. Defaults to None. moving_average_decay: The amount of decay to apply. If 0 or None, then exponential moving average is not used. Defaults to None. lookahead: Whether or not to apply the lookahead optimizer. Defaults to None. beta_1: The exponential decay rate for the 1st moment estimates. Used in the Adam optimizers. Defaults to None. beta_2: The exponential decay rate for the 2nd moment estimates. Used in the Adam optimizers. Defaults to None. epsilon: Small value used to avoid 0 denominator. Defaults to 1e-7. """ name: str = None decay: float = None epsilon: float = None momentum: float = None nesterov: bool = None moving_average_decay: Optional[float] = None lookahead: Optional[bool] = None beta_1: float = None beta_2: float = None epsilon: float = None @dataclasses.dataclass class LearningRateConfig(hyperparams.Config): """Configuration for learning rates. Attributes: name: The name of the learning rate. Defaults to None. initial_lr: The initial learning rate. Defaults to None. decay_epochs: The number of decay epochs. Defaults to None. decay_rate: The rate of decay. Defaults to None. warmup_epochs: The number of warmup epochs. Defaults to None. batch_lr_multiplier: The multiplier to apply to the base learning rate, if necessary. Defaults to None. examples_per_epoch: the number of examples in a single epoch. Defaults to None. boundaries: boundaries used in piecewise constant decay with warmup. multipliers: multipliers used in piecewise constant decay with warmup. scale_by_batch_size: Scale the learning rate by a fraction of the batch size. Set to 0 for no scaling (default). staircase: Apply exponential decay at discrete values instead of continuous. """ name: str = None initial_lr: float = None decay_epochs: float = None decay_rate: float = None warmup_epochs: int = None examples_per_epoch: int = None boundaries: List[int] = None multipliers: List[float] = None scale_by_batch_size: float = 0. staircase: bool = None @dataclasses.dataclass class ModelConfig(hyperparams.Config): """Configuration for Models. Attributes: name: The name of the model. Defaults to None. model_params: The parameters used to create the model. Defaults to None. num_classes: The number of classes in the model. Defaults to None. loss: A `LossConfig` instance. Defaults to None. optimizer: An `OptimizerConfig` instance. Defaults to None. """ name: str = None model_params: hyperparams.Config = None num_classes: int = None loss: LossConfig = None optimizer: OptimizerConfig = None @dataclasses.dataclass class ExperimentConfig(hyperparams.Config): """Base configuration for an image classification experiment. Attributes: model_dir: The directory to use when running an experiment. mode: e.g. 'train_and_eval', 'export' runtime: A `RuntimeConfig` instance. train: A `TrainConfig` instance. evaluation: An `EvalConfig` instance. model: A `ModelConfig` instance. export: An `ExportConfig` instance. """ model_dir: str = None model_name: str = None mode: str = None runtime: RuntimeConfig = None train_dataset: Any = None validation_dataset: Any = None train: TrainConfig = None evaluation: EvalConfig = None model: ModelConfig = None export: ExportConfig = None
8,797
32.452471
80
py
models
models-master/official/legacy/image_classification/configs/configs.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Configuration utils for image classification experiments.""" import dataclasses from official.legacy.image_classification import dataset_factory from official.legacy.image_classification.configs import base_configs from official.legacy.image_classification.efficientnet import efficientnet_config from official.legacy.image_classification.resnet import resnet_config from official.legacy.image_classification.vgg import vgg_config @dataclasses.dataclass class EfficientNetImageNetConfig(base_configs.ExperimentConfig): """Base configuration to train efficientnet-b0 on ImageNet. Attributes: export: An `ExportConfig` instance runtime: A `RuntimeConfig` instance. dataset: A `DatasetConfig` instance. train: A `TrainConfig` instance. evaluation: An `EvalConfig` instance. model: A `ModelConfig` instance. """ export: base_configs.ExportConfig = dataclasses.field( default_factory=base_configs.ExportConfig ) runtime: base_configs.RuntimeConfig = dataclasses.field( default_factory=base_configs.RuntimeConfig ) train_dataset: dataset_factory.DatasetConfig = dataclasses.field( default_factory=lambda: dataset_factory.ImageNetConfig(split='train') ) validation_dataset: dataset_factory.DatasetConfig = dataclasses.field( default_factory=lambda: dataset_factory.ImageNetConfig(split='validation') ) train: base_configs.TrainConfig = dataclasses.field( default_factory=lambda: base_configs.TrainConfig( # pylint: disable=g-long-lambda resume_checkpoint=True, epochs=500, steps=None, callbacks=base_configs.CallbacksConfig( enable_checkpoint_and_export=True, enable_tensorboard=True ), metrics=['accuracy', 'top_5'], time_history=base_configs.TimeHistoryConfig(log_steps=100), tensorboard=base_configs.TensorBoardConfig( track_lr=True, write_model_weights=False ), set_epoch_loop=False, ) ) evaluation: base_configs.EvalConfig = dataclasses.field( default_factory=lambda: base_configs.EvalConfig( # pylint: disable=g-long-lambda epochs_between_evals=1, steps=None ) ) model: base_configs.ModelConfig = dataclasses.field( default_factory=efficientnet_config.EfficientNetModelConfig ) @dataclasses.dataclass class ResNetImagenetConfig(base_configs.ExperimentConfig): """Base configuration to train resnet-50 on ImageNet.""" export: base_configs.ExportConfig = dataclasses.field( default_factory=base_configs.ExportConfig ) runtime: base_configs.RuntimeConfig = dataclasses.field( default_factory=base_configs.RuntimeConfig ) train_dataset: dataset_factory.DatasetConfig = dataclasses.field( default_factory=lambda: dataset_factory.ImageNetConfig( # pylint: disable=g-long-lambda split='train', one_hot=False, mean_subtract=True, standardize=True ) ) validation_dataset: dataset_factory.DatasetConfig = dataclasses.field( default_factory=lambda: dataset_factory.ImageNetConfig( # pylint: disable=g-long-lambda split='validation', one_hot=False, mean_subtract=True, standardize=True, ) ) train: base_configs.TrainConfig = dataclasses.field( default_factory=lambda: base_configs.TrainConfig( # pylint: disable=g-long-lambda resume_checkpoint=True, epochs=90, steps=None, callbacks=base_configs.CallbacksConfig( enable_checkpoint_and_export=True, enable_tensorboard=True ), metrics=['accuracy', 'top_5'], time_history=base_configs.TimeHistoryConfig(log_steps=100), tensorboard=base_configs.TensorBoardConfig( track_lr=True, write_model_weights=False ), set_epoch_loop=False, ) ) evaluation: base_configs.EvalConfig = dataclasses.field( default_factory=lambda: base_configs.EvalConfig( # pylint: disable=g-long-lambda epochs_between_evals=1, steps=None ) ) model: base_configs.ModelConfig = dataclasses.field( default_factory=resnet_config.ResNetModelConfig ) @dataclasses.dataclass class VGGImagenetConfig(base_configs.ExperimentConfig): """Base configuration to train vgg-16 on ImageNet.""" export: base_configs.ExportConfig = dataclasses.field( default_factory=base_configs.ExportConfig ) runtime: base_configs.RuntimeConfig = dataclasses.field( default_factory=base_configs.RuntimeConfig ) train_dataset: dataset_factory.DatasetConfig = dataclasses.field( default_factory=lambda: dataset_factory.ImageNetConfig( # pylint: disable=g-long-lambda split='train', one_hot=False, mean_subtract=True, standardize=True ) ) validation_dataset: dataset_factory.DatasetConfig = dataclasses.field( default_factory=lambda: dataset_factory.ImageNetConfig( # pylint: disable=g-long-lambda split='validation', one_hot=False, mean_subtract=True, standardize=True, ) ) train: base_configs.TrainConfig = dataclasses.field( default_factory=lambda: base_configs.TrainConfig( # pylint: disable=g-long-lambda resume_checkpoint=True, epochs=90, steps=None, callbacks=base_configs.CallbacksConfig( enable_checkpoint_and_export=True, enable_tensorboard=True ), metrics=['accuracy', 'top_5'], time_history=base_configs.TimeHistoryConfig(log_steps=100), tensorboard=base_configs.TensorBoardConfig( track_lr=True, write_model_weights=False ), set_epoch_loop=False, ) ) evaluation: base_configs.EvalConfig = dataclasses.field( default_factory=lambda: base_configs.EvalConfig( # pylint: disable=g-long-lambda epochs_between_evals=1, steps=None ) ) model: base_configs.ModelConfig = dataclasses.field( default_factory=vgg_config.VGGModelConfig ) def get_config(model: str, dataset: str) -> base_configs.ExperimentConfig: """Given model and dataset names, return the ExperimentConfig.""" dataset_model_config_map = { 'imagenet': { 'efficientnet': EfficientNetImageNetConfig(), 'resnet': ResNetImagenetConfig(), 'vgg': VGGImagenetConfig(), } } try: return dataset_model_config_map[dataset][model] except KeyError: if dataset not in dataset_model_config_map: raise KeyError('Invalid dataset received. Received: {}. Supported ' 'datasets include: {}'.format( dataset, ', '.join(dataset_model_config_map.keys()))) raise KeyError('Invalid model received. Received: {}. Supported models for' '{} include: {}'.format( model, dataset, ', '.join(dataset_model_config_map[dataset].keys())))
7,571
38.4375
94
py
models
models-master/official/legacy/image_classification/resnet/resnet_ctl_imagenet_main.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Runs a ResNet model on the ImageNet dataset using custom training loops.""" import math import os # Import libraries from absl import app from absl import flags from absl import logging import orbit import tensorflow as tf from official.common import distribute_utils from official.legacy.image_classification.resnet import common from official.legacy.image_classification.resnet import imagenet_preprocessing from official.legacy.image_classification.resnet import resnet_runnable from official.modeling import performance from official.utils.flags import core as flags_core from official.utils.misc import keras_utils from official.utils.misc import model_helpers flags.DEFINE_boolean(name='use_tf_function', default=True, help='Wrap the train and test step inside a ' 'tf.function.') flags.DEFINE_boolean(name='single_l2_loss_op', default=False, help='Calculate L2_loss on concatenated weights, ' 'instead of using Keras per-layer L2 loss.') def build_stats(runnable, time_callback): """Normalizes and returns dictionary of stats. Args: runnable: The module containing all the training and evaluation metrics. time_callback: Time tracking callback instance. Returns: Dictionary of normalized results. """ stats = {} if not runnable.flags_obj.skip_eval: stats['eval_loss'] = runnable.test_loss.result().numpy() stats['eval_acc'] = runnable.test_accuracy.result().numpy() stats['train_loss'] = runnable.train_loss.result().numpy() stats['train_acc'] = runnable.train_accuracy.result().numpy() if time_callback: timestamp_log = time_callback.timestamp_log stats['step_timestamp_log'] = timestamp_log stats['train_finish_time'] = time_callback.train_finish_time if time_callback.epoch_runtime_log: stats['avg_exp_per_second'] = time_callback.average_examples_per_second return stats def get_num_train_iterations(flags_obj): """Returns the number of training steps, train and test epochs.""" train_steps = ( imagenet_preprocessing.NUM_IMAGES['train'] // flags_obj.batch_size) train_epochs = flags_obj.train_epochs if flags_obj.train_steps: train_steps = min(flags_obj.train_steps, train_steps) train_epochs = 1 eval_steps = math.ceil(1.0 * imagenet_preprocessing.NUM_IMAGES['validation'] / flags_obj.batch_size) return train_steps, train_epochs, eval_steps def run(flags_obj): """Run ResNet ImageNet training and eval loop using custom training loops. Args: flags_obj: An object containing parsed flag values. Raises: ValueError: If fp16 is passed as it is not currently supported. Returns: Dictionary of training and eval stats. """ keras_utils.set_session_config() performance.set_mixed_precision_policy(flags_core.get_tf_dtype(flags_obj)) if tf.config.list_physical_devices('GPU'): if flags_obj.tf_gpu_thread_mode: keras_utils.set_gpu_thread_mode_and_count( per_gpu_thread_count=flags_obj.per_gpu_thread_count, gpu_thread_mode=flags_obj.tf_gpu_thread_mode, num_gpus=flags_obj.num_gpus, datasets_num_private_threads=flags_obj.datasets_num_private_threads) common.set_cudnn_batchnorm_mode() data_format = flags_obj.data_format if data_format is None: data_format = ('channels_first' if tf.config.list_physical_devices('GPU') else 'channels_last') tf.keras.backend.set_image_data_format(data_format) strategy = distribute_utils.get_distribution_strategy( distribution_strategy=flags_obj.distribution_strategy, num_gpus=flags_obj.num_gpus, all_reduce_alg=flags_obj.all_reduce_alg, num_packs=flags_obj.num_packs, tpu_address=flags_obj.tpu) per_epoch_steps, train_epochs, eval_steps = get_num_train_iterations( flags_obj) if flags_obj.steps_per_loop is None: steps_per_loop = per_epoch_steps elif flags_obj.steps_per_loop > per_epoch_steps: steps_per_loop = per_epoch_steps logging.warn('Setting steps_per_loop to %d to respect epoch boundary.', steps_per_loop) else: steps_per_loop = flags_obj.steps_per_loop logging.info( 'Training %d epochs, each epoch has %d steps, ' 'total steps: %d; Eval %d steps', train_epochs, per_epoch_steps, train_epochs * per_epoch_steps, eval_steps) time_callback = keras_utils.TimeHistory( flags_obj.batch_size, flags_obj.log_steps, logdir=flags_obj.model_dir if flags_obj.enable_tensorboard else None) with distribute_utils.get_strategy_scope(strategy): runnable = resnet_runnable.ResnetRunnable(flags_obj, time_callback, per_epoch_steps) eval_interval = flags_obj.epochs_between_evals * per_epoch_steps checkpoint_interval = ( steps_per_loop * 5 if flags_obj.enable_checkpoint_and_export else None) summary_interval = steps_per_loop if flags_obj.enable_tensorboard else None checkpoint_manager = tf.train.CheckpointManager( runnable.checkpoint, directory=flags_obj.model_dir, max_to_keep=10, step_counter=runnable.global_step, checkpoint_interval=checkpoint_interval) resnet_controller = orbit.Controller( strategy=strategy, trainer=runnable, evaluator=runnable if not flags_obj.skip_eval else None, global_step=runnable.global_step, steps_per_loop=steps_per_loop, checkpoint_manager=checkpoint_manager, summary_interval=summary_interval, summary_dir=flags_obj.model_dir, eval_summary_dir=os.path.join(flags_obj.model_dir, 'eval')) time_callback.on_train_begin() if not flags_obj.skip_eval: resnet_controller.train_and_evaluate( train_steps=per_epoch_steps * train_epochs, eval_steps=eval_steps, eval_interval=eval_interval) else: resnet_controller.train(steps=per_epoch_steps * train_epochs) time_callback.on_train_end() stats = build_stats(runnable, time_callback) return stats def main(_): model_helpers.apply_clean(flags.FLAGS) stats = run(flags.FLAGS) logging.info('Run stats:\n%s', stats) if __name__ == '__main__': logging.set_verbosity(logging.INFO) common.define_keras_flags() app.run(main)
6,916
34.290816
80
py
models
models-master/official/legacy/image_classification/resnet/imagenet_preprocessing.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Provides utilities to preprocess images. Training images are sampled using the provided bounding boxes, and subsequently cropped to the sampled bounding box. Images are additionally flipped randomly, then resized to the target output size (without aspect-ratio preservation). Images used during evaluation are resized (with aspect-ratio preservation) and centrally cropped. All images undergo mean color subtraction. Note that these steps are colloquially referred to as "ResNet preprocessing," and they differ from "VGG preprocessing," which does not use bounding boxes and instead does an aspect-preserving resize followed by random crop during training. (These both differ from "Inception preprocessing," which introduces color distortion steps.) """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import os from absl import logging import tensorflow as tf DEFAULT_IMAGE_SIZE = 224 NUM_CHANNELS = 3 NUM_CLASSES = 1001 NUM_IMAGES = { 'train': 1281167, 'validation': 50000, } _NUM_TRAIN_FILES = 1024 _SHUFFLE_BUFFER = 10000 _R_MEAN = 123.68 _G_MEAN = 116.78 _B_MEAN = 103.94 CHANNEL_MEANS = [_R_MEAN, _G_MEAN, _B_MEAN] # The lower bound for the smallest side of the image for aspect-preserving # resizing. For example, if an image is 500 x 1000, it will be resized to # _RESIZE_MIN x (_RESIZE_MIN * 2). _RESIZE_MIN = 256 def process_record_dataset(dataset, is_training, batch_size, shuffle_buffer, parse_record_fn, dtype=tf.float32, datasets_num_private_threads=None, drop_remainder=False, tf_data_experimental_slack=False): """Given a Dataset with raw records, return an iterator over the records. Args: dataset: A Dataset representing raw records is_training: A boolean denoting whether the input is for training. batch_size: The number of samples per batch. shuffle_buffer: The buffer size to use when shuffling records. A larger value results in better randomness, but smaller values reduce startup time and use less memory. parse_record_fn: A function that takes a raw record and returns the corresponding (image, label) pair. dtype: Data type to use for images/features. datasets_num_private_threads: Number of threads for a private threadpool created for all datasets computation. drop_remainder: A boolean indicates whether to drop the remainder of the batches. If True, the batch dimension will be static. tf_data_experimental_slack: Whether to enable tf.data's `experimental_slack` option. Returns: Dataset of (image, label) pairs ready for iteration. """ # Defines a specific size thread pool for tf.data operations. if datasets_num_private_threads: options = tf.data.Options() options.experimental_threading.private_threadpool_size = ( datasets_num_private_threads) dataset = dataset.with_options(options) logging.info('datasets_num_private_threads: %s', datasets_num_private_threads) if is_training: # Shuffles records before repeating to respect epoch boundaries. dataset = dataset.shuffle(buffer_size=shuffle_buffer) # Repeats the dataset for the number of epochs to train. dataset = dataset.repeat() # Parses the raw records into images and labels. dataset = dataset.map( lambda value: parse_record_fn(value, is_training, dtype), num_parallel_calls=tf.data.experimental.AUTOTUNE) dataset = dataset.batch(batch_size, drop_remainder=drop_remainder) # Operations between the final prefetch and the get_next call to the iterator # will happen synchronously during run time. We prefetch here again to # background all of the above processing work and keep it out of the # critical training path. Setting buffer_size to tf.data.experimental.AUTOTUNE # allows DistributionStrategies to adjust how many batches to fetch based # on how many devices are present. dataset = dataset.prefetch(buffer_size=tf.data.experimental.AUTOTUNE) options = tf.data.Options() options.experimental_slack = tf_data_experimental_slack dataset = dataset.with_options(options) return dataset def get_filenames(is_training, data_dir): """Return filenames for dataset.""" if is_training: return [ os.path.join(data_dir, 'train-%05d-of-01024' % i) for i in range(_NUM_TRAIN_FILES) ] else: return [ os.path.join(data_dir, 'validation-%05d-of-00128' % i) for i in range(128) ] def parse_example_proto(example_serialized): """Parses an Example proto containing a training example of an image. The output of the build_image_data.py image preprocessing script is a dataset containing serialized Example protocol buffers. Each Example proto contains the following fields (values are included as examples): image/height: 462 image/width: 581 image/colorspace: 'RGB' image/channels: 3 image/class/label: 615 image/class/synset: 'n03623198' image/class/text: 'knee pad' image/object/bbox/xmin: 0.1 image/object/bbox/xmax: 0.9 image/object/bbox/ymin: 0.2 image/object/bbox/ymax: 0.6 image/object/bbox/label: 615 image/format: 'JPEG' image/filename: 'ILSVRC2012_val_00041207.JPEG' image/encoded: <JPEG encoded string> Args: example_serialized: scalar Tensor tf.string containing a serialized Example protocol buffer. Returns: image_buffer: Tensor tf.string containing the contents of a JPEG file. label: Tensor tf.int32 containing the label. bbox: 3-D float Tensor of bounding boxes arranged [1, num_boxes, coords] where each coordinate is [0, 1) and the coordinates are arranged as [ymin, xmin, ymax, xmax]. """ # Dense features in Example proto. feature_map = { 'image/encoded': tf.io.FixedLenFeature([], dtype=tf.string, default_value=''), 'image/class/label': tf.io.FixedLenFeature([], dtype=tf.int64, default_value=-1), 'image/class/text': tf.io.FixedLenFeature([], dtype=tf.string, default_value=''), } sparse_float32 = tf.io.VarLenFeature(dtype=tf.float32) # Sparse features in Example proto. feature_map.update({ k: sparse_float32 for k in [ 'image/object/bbox/xmin', 'image/object/bbox/ymin', 'image/object/bbox/xmax', 'image/object/bbox/ymax' ] }) features = tf.io.parse_single_example( serialized=example_serialized, features=feature_map) label = tf.cast(features['image/class/label'], dtype=tf.int32) xmin = tf.expand_dims(features['image/object/bbox/xmin'].values, 0) ymin = tf.expand_dims(features['image/object/bbox/ymin'].values, 0) xmax = tf.expand_dims(features['image/object/bbox/xmax'].values, 0) ymax = tf.expand_dims(features['image/object/bbox/ymax'].values, 0) # Note that we impose an ordering of (y, x) just to make life difficult. bbox = tf.concat([ymin, xmin, ymax, xmax], 0) # Force the variable number of bounding boxes into the shape # [1, num_boxes, coords]. bbox = tf.expand_dims(bbox, 0) bbox = tf.transpose(a=bbox, perm=[0, 2, 1]) return features['image/encoded'], label, bbox def parse_record(raw_record, is_training, dtype): """Parses a record containing a training example of an image. The input record is parsed into a label and image, and the image is passed through preprocessing steps (cropping, flipping, and so on). Args: raw_record: scalar Tensor tf.string containing a serialized Example protocol buffer. is_training: A boolean denoting whether the input is for training. dtype: data type to use for images/features. Returns: Tuple with processed image tensor in a channel-last format and one-hot-encoded label tensor. """ image_buffer, label, bbox = parse_example_proto(raw_record) image = preprocess_image( image_buffer=image_buffer, bbox=bbox, output_height=DEFAULT_IMAGE_SIZE, output_width=DEFAULT_IMAGE_SIZE, num_channels=NUM_CHANNELS, is_training=is_training) image = tf.cast(image, dtype) # Subtract one so that labels are in [0, 1000), and cast to float32 for # Keras model. label = tf.cast( tf.cast(tf.reshape(label, shape=[1]), dtype=tf.int32) - 1, dtype=tf.float32) return image, label def get_parse_record_fn(use_keras_image_data_format=False): """Get a function for parsing the records, accounting for image format. This is useful by handling different types of Keras models. For instance, the current resnet_model.resnet50 input format is always channel-last, whereas the keras_applications mobilenet input format depends on tf.keras.backend.image_data_format(). We should set use_keras_image_data_format=False for the former and True for the latter. Args: use_keras_image_data_format: A boolean denoting whether data format is keras backend image data format. If False, the image format is channel-last. If True, the image format matches tf.keras.backend.image_data_format(). Returns: Function to use for parsing the records. """ def parse_record_fn(raw_record, is_training, dtype): image, label = parse_record(raw_record, is_training, dtype) if use_keras_image_data_format: if tf.keras.backend.image_data_format() == 'channels_first': image = tf.transpose(image, perm=[2, 0, 1]) return image, label return parse_record_fn def input_fn(is_training, data_dir, batch_size, dtype=tf.float32, datasets_num_private_threads=None, parse_record_fn=parse_record, input_context=None, drop_remainder=False, tf_data_experimental_slack=False, training_dataset_cache=False, filenames=None): """Input function which provides batches for train or eval. Args: is_training: A boolean denoting whether the input is for training. data_dir: The directory containing the input data. batch_size: The number of samples per batch. dtype: Data type to use for images/features datasets_num_private_threads: Number of private threads for tf.data. parse_record_fn: Function to use for parsing the records. input_context: A `tf.distribute.InputContext` object passed in by `tf.distribute.Strategy`. drop_remainder: A boolean indicates whether to drop the remainder of the batches. If True, the batch dimension will be static. tf_data_experimental_slack: Whether to enable tf.data's `experimental_slack` option. training_dataset_cache: Whether to cache the training dataset on workers. Typically used to improve training performance when training data is in remote storage and can fit into worker memory. filenames: Optional field for providing the file names of the TFRecords. Returns: A dataset that can be used for iteration. """ if filenames is None: filenames = get_filenames(is_training, data_dir) dataset = tf.data.Dataset.from_tensor_slices(filenames) if input_context: logging.info( 'Sharding the dataset: input_pipeline_id=%d num_input_pipelines=%d', input_context.input_pipeline_id, input_context.num_input_pipelines) dataset = dataset.shard(input_context.num_input_pipelines, input_context.input_pipeline_id) if is_training: # Shuffle the input files dataset = dataset.shuffle(buffer_size=_NUM_TRAIN_FILES) # Convert to individual records. # cycle_length = 10 means that up to 10 files will be read and deserialized in # parallel. You may want to increase this number if you have a large number of # CPU cores. dataset = dataset.interleave( tf.data.TFRecordDataset, cycle_length=10, num_parallel_calls=tf.data.experimental.AUTOTUNE) if is_training and training_dataset_cache: # Improve training performance when training data is in remote storage and # can fit into worker memory. dataset = dataset.cache() return process_record_dataset( dataset=dataset, is_training=is_training, batch_size=batch_size, shuffle_buffer=_SHUFFLE_BUFFER, parse_record_fn=parse_record_fn, dtype=dtype, datasets_num_private_threads=datasets_num_private_threads, drop_remainder=drop_remainder, tf_data_experimental_slack=tf_data_experimental_slack, ) def _decode_crop_and_flip(image_buffer, bbox, num_channels): """Crops the given image to a random part of the image, and randomly flips. We use the fused decode_and_crop op, which performs better than the two ops used separately in series, but note that this requires that the image be passed in as an un-decoded string Tensor. Args: image_buffer: scalar string Tensor representing the raw JPEG image buffer. bbox: 3-D float Tensor of bounding boxes arranged [1, num_boxes, coords] where each coordinate is [0, 1) and the coordinates are arranged as [ymin, xmin, ymax, xmax]. num_channels: Integer depth of the image buffer for decoding. Returns: 3-D tensor with cropped image. """ # A large fraction of image datasets contain a human-annotated bounding box # delineating the region of the image containing the object of interest. We # choose to create a new bounding box for the object which is a randomly # distorted version of the human-annotated bounding box that obeys an # allowed range of aspect ratios, sizes and overlap with the human-annotated # bounding box. If no box is supplied, then we assume the bounding box is # the entire image. sample_distorted_bounding_box = tf.image.sample_distorted_bounding_box( tf.image.extract_jpeg_shape(image_buffer), bounding_boxes=bbox, min_object_covered=0.1, aspect_ratio_range=[0.75, 1.33], area_range=[0.05, 1.0], max_attempts=100, use_image_if_no_bounding_boxes=True) bbox_begin, bbox_size, _ = sample_distorted_bounding_box # Reassemble the bounding box in the format the crop op requires. offset_y, offset_x, _ = tf.unstack(bbox_begin) target_height, target_width, _ = tf.unstack(bbox_size) crop_window = tf.stack([offset_y, offset_x, target_height, target_width]) # Use the fused decode and crop op here, which is faster than each in series. cropped = tf.image.decode_and_crop_jpeg( image_buffer, crop_window, channels=num_channels) # Flip to add a little more random distortion in. cropped = tf.image.random_flip_left_right(cropped) return cropped def _central_crop(image, crop_height, crop_width): """Performs central crops of the given image list. Args: image: a 3-D image tensor crop_height: the height of the image following the crop. crop_width: the width of the image following the crop. Returns: 3-D tensor with cropped image. """ shape = tf.shape(input=image) height, width = shape[0], shape[1] amount_to_be_cropped_h = (height - crop_height) crop_top = amount_to_be_cropped_h // 2 amount_to_be_cropped_w = (width - crop_width) crop_left = amount_to_be_cropped_w // 2 return tf.slice(image, [crop_top, crop_left, 0], [crop_height, crop_width, -1]) def _mean_image_subtraction(image, means, num_channels): """Subtracts the given means from each image channel. For example: means = [123.68, 116.779, 103.939] image = _mean_image_subtraction(image, means) Note that the rank of `image` must be known. Args: image: a tensor of size [height, width, C]. means: a C-vector of values to subtract from each channel. num_channels: number of color channels in the image that will be distorted. Returns: the centered image. Raises: ValueError: If the rank of `image` is unknown, if `image` has a rank other than three or if the number of channels in `image` doesn't match the number of values in `means`. """ if image.get_shape().ndims != 3: raise ValueError('Input must be of size [height, width, C>0]') if len(means) != num_channels: raise ValueError('len(means) must match the number of channels') # We have a 1-D tensor of means; convert to 3-D. # Note(b/130245863): we explicitly call `broadcast` instead of simply # expanding dimensions for better performance. means = tf.broadcast_to(means, tf.shape(image)) return image - means def _smallest_size_at_least(height, width, resize_min): """Computes new shape with the smallest side equal to `smallest_side`. Computes new shape with the smallest side equal to `smallest_side` while preserving the original aspect ratio. Args: height: an int32 scalar tensor indicating the current height. width: an int32 scalar tensor indicating the current width. resize_min: A python integer or scalar `Tensor` indicating the size of the smallest side after resize. Returns: new_height: an int32 scalar tensor indicating the new height. new_width: an int32 scalar tensor indicating the new width. """ resize_min = tf.cast(resize_min, tf.float32) # Convert to floats to make subsequent calculations go smoothly. height, width = tf.cast(height, tf.float32), tf.cast(width, tf.float32) smaller_dim = tf.minimum(height, width) scale_ratio = resize_min / smaller_dim # Convert back to ints to make heights and widths that TF ops will accept. new_height = tf.cast(height * scale_ratio, tf.int32) new_width = tf.cast(width * scale_ratio, tf.int32) return new_height, new_width def _aspect_preserving_resize(image, resize_min): """Resize images preserving the original aspect ratio. Args: image: A 3-D image `Tensor`. resize_min: A python integer or scalar `Tensor` indicating the size of the smallest side after resize. Returns: resized_image: A 3-D tensor containing the resized image. """ shape = tf.shape(input=image) height, width = shape[0], shape[1] new_height, new_width = _smallest_size_at_least(height, width, resize_min) return _resize_image(image, new_height, new_width) def _resize_image(image, height, width): """Simple wrapper around tf.resize_images. This is primarily to make sure we use the same `ResizeMethod` and other details each time. Args: image: A 3-D image `Tensor`. height: The target height for the resized image. width: The target width for the resized image. Returns: resized_image: A 3-D tensor containing the resized image. The first two dimensions have the shape [height, width]. """ return tf.compat.v1.image.resize( image, [height, width], method=tf.image.ResizeMethod.BILINEAR, align_corners=False) def preprocess_image(image_buffer, bbox, output_height, output_width, num_channels, is_training=False): """Preprocesses the given image. Preprocessing includes decoding, cropping, and resizing for both training and eval images. Training preprocessing, however, introduces some random distortion of the image to improve accuracy. Args: image_buffer: scalar string Tensor representing the raw JPEG image buffer. bbox: 3-D float Tensor of bounding boxes arranged [1, num_boxes, coords] where each coordinate is [0, 1) and the coordinates are arranged as [ymin, xmin, ymax, xmax]. output_height: The height of the image after preprocessing. output_width: The width of the image after preprocessing. num_channels: Integer depth of the image buffer for decoding. is_training: `True` if we're preprocessing the image for training and `False` otherwise. Returns: A preprocessed image. """ if is_training: # For training, we want to randomize some of the distortions. image = _decode_crop_and_flip(image_buffer, bbox, num_channels) image = _resize_image(image, output_height, output_width) else: # For validation, we want to decode, resize, then just crop the middle. image = tf.image.decode_jpeg(image_buffer, channels=num_channels) image = _aspect_preserving_resize(image, _RESIZE_MIN) image = _central_crop(image, output_height, output_width) image.set_shape([output_height, output_width, num_channels]) return _mean_image_subtraction(image, CHANNEL_MEANS, num_channels)
21,168
35.815652
80
py
models
models-master/official/legacy/image_classification/resnet/resnet_config.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Configuration definitions for ResNet losses, learning rates, and optimizers.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import dataclasses from official.legacy.image_classification.configs import base_configs from official.modeling.hyperparams import base_config @dataclasses.dataclass class ResNetModelConfig(base_configs.ModelConfig): """Configuration for the ResNet model.""" name: str = 'ResNet' num_classes: int = 1000 model_params: base_config.Config = dataclasses.field( # pylint: disable=g-long-lambda default_factory=lambda: { 'num_classes': 1000, 'batch_size': None, 'use_l2_regularizer': True, 'rescale_inputs': False, }) # pylint: enable=g-long-lambda loss: base_configs.LossConfig = dataclasses.field( default_factory=lambda: base_configs.LossConfig( # pylint: disable=g-long-lambda name='sparse_categorical_crossentropy' ) ) optimizer: base_configs.OptimizerConfig = dataclasses.field( default_factory=lambda: base_configs.OptimizerConfig( # pylint: disable=g-long-lambda name='momentum', decay=0.9, epsilon=0.001, momentum=0.9, moving_average_decay=None, ) ) learning_rate: base_configs.LearningRateConfig = dataclasses.field( default_factory=lambda: base_configs.LearningRateConfig( # pylint: disable=g-long-lambda name='stepwise', initial_lr=0.1, examples_per_epoch=1281167, boundaries=[30, 60, 80], warmup_epochs=5, scale_by_batch_size=1.0 / 256.0, multipliers=[0.1 / 256, 0.01 / 256, 0.001 / 256, 0.0001 / 256], ) )
2,368
36.015625
95
py
models
models-master/official/legacy/image_classification/resnet/tfhub_export.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """A script to export TF-Hub SavedModel.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import os # Import libraries from absl import app from absl import flags import tensorflow as tf from official.legacy.image_classification.resnet import imagenet_preprocessing from official.legacy.image_classification.resnet import resnet_model FLAGS = flags.FLAGS flags.DEFINE_string("model_path", None, "File path to TF model checkpoint or H5 file.") flags.DEFINE_string("export_path", None, "TF-Hub SavedModel destination path to export.") def export_tfhub(model_path, hub_destination): """Restores a tf.keras.Model and saves for TF-Hub.""" model = resnet_model.resnet50( num_classes=imagenet_preprocessing.NUM_CLASSES, rescale_inputs=True) model.load_weights(model_path) model.save( os.path.join(hub_destination, "classification"), include_optimizer=False) # Extracts a sub-model to use pooling feature vector as model output. image_input = model.get_layer(index=0).get_output_at(0) feature_vector_output = model.get_layer(name="reduce_mean").get_output_at(0) hub_model = tf.keras.Model(image_input, feature_vector_output) # Exports a SavedModel. hub_model.save( os.path.join(hub_destination, "feature-vector"), include_optimizer=False) def main(argv): if len(argv) > 1: raise app.UsageError("Too many command-line arguments.") export_tfhub(FLAGS.model_path, FLAGS.export_path) if __name__ == "__main__": app.run(main)
2,189
31.686567
79
py
models
models-master/official/legacy/image_classification/resnet/common.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Common util functions and classes used by both keras cifar and imagenet.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import os from absl import flags import tensorflow as tf import tensorflow_model_optimization as tfmot from official.utils.flags import core as flags_core from official.utils.misc import keras_utils FLAGS = flags.FLAGS BASE_LEARNING_RATE = 0.1 # This matches Jing's version. TRAIN_TOP_1 = 'training_accuracy_top_1' LR_SCHEDULE = [ # (multiplier, epoch to start) tuples (1.0, 5), (0.1, 30), (0.01, 60), (0.001, 80) ] class PiecewiseConstantDecayWithWarmup( tf.keras.optimizers.schedules.LearningRateSchedule): """Piecewise constant decay with warmup schedule.""" def __init__(self, batch_size, epoch_size, warmup_epochs, boundaries, multipliers, compute_lr_on_cpu=True, name=None): super(PiecewiseConstantDecayWithWarmup, self).__init__() if len(boundaries) != len(multipliers) - 1: raise ValueError('The length of boundaries must be 1 less than the ' 'length of multipliers') base_lr_batch_size = 256 steps_per_epoch = epoch_size // batch_size self.rescaled_lr = BASE_LEARNING_RATE * batch_size / base_lr_batch_size self.step_boundaries = [float(steps_per_epoch) * x for x in boundaries] self.lr_values = [self.rescaled_lr * m for m in multipliers] self.warmup_steps = warmup_epochs * steps_per_epoch self.compute_lr_on_cpu = compute_lr_on_cpu self.name = name self.learning_rate_ops_cache = {} def __call__(self, step): if tf.executing_eagerly(): return self._get_learning_rate(step) # In an eager function or graph, the current implementation of optimizer # repeatedly call and thus create ops for the learning rate schedule. To # avoid this, we cache the ops if not executing eagerly. graph = tf.compat.v1.get_default_graph() if graph not in self.learning_rate_ops_cache: if self.compute_lr_on_cpu: with tf.device('/device:CPU:0'): self.learning_rate_ops_cache[graph] = self._get_learning_rate(step) else: self.learning_rate_ops_cache[graph] = self._get_learning_rate(step) return self.learning_rate_ops_cache[graph] def _get_learning_rate(self, step): """Compute learning rate at given step.""" step = tf.cast(step, dtype=tf.float32) warmup_steps = tf.cast(self.warmup_steps, dtype=tf.float32) with tf.name_scope('PiecewiseConstantDecayWithWarmup'): def warmup_lr(step): return self.rescaled_lr * (step / warmup_steps) def piecewise_lr(step): return tf.compat.v1.train.piecewise_constant(step, self.step_boundaries, self.lr_values) return tf.cond(step < warmup_steps, lambda: warmup_lr(step), lambda: piecewise_lr(step)) def get_config(self): return { 'rescaled_lr': self.rescaled_lr, 'step_boundaries': self.step_boundaries, 'lr_values': self.lr_values, 'warmup_steps': self.warmup_steps, 'compute_lr_on_cpu': self.compute_lr_on_cpu, 'name': self.name } def get_optimizer(learning_rate=0.1, use_legacy_optimizer=True): """Returns optimizer to use.""" # The learning_rate is overwritten at the beginning of each step by callback. if use_legacy_optimizer: return tf.keras.optimizers.legacy.SGD( learning_rate=learning_rate, momentum=0.9) else: return tf.keras.optimizers.SGD(learning_rate=learning_rate, momentum=0.9) def get_callbacks(pruning_method=None, enable_checkpoint_and_export=False, model_dir=None): """Returns common callbacks.""" time_callback = keras_utils.TimeHistory( FLAGS.batch_size, FLAGS.log_steps, logdir=FLAGS.model_dir if FLAGS.enable_tensorboard else None) callbacks = [time_callback] if FLAGS.enable_tensorboard: tensorboard_callback = tf.keras.callbacks.TensorBoard( log_dir=FLAGS.model_dir, profile_batch=FLAGS.profile_steps) callbacks.append(tensorboard_callback) is_pruning_enabled = pruning_method is not None if is_pruning_enabled: callbacks.append(tfmot.sparsity.keras.UpdatePruningStep()) if model_dir is not None: callbacks.append( tfmot.sparsity.keras.PruningSummaries( log_dir=model_dir, profile_batch=0)) if enable_checkpoint_and_export: if model_dir is not None: ckpt_full_path = os.path.join(model_dir, 'model.ckpt-{epoch:04d}') callbacks.append( tf.keras.callbacks.ModelCheckpoint( ckpt_full_path, save_weights_only=True)) return callbacks def build_stats(history, eval_output, callbacks): """Normalizes and returns dictionary of stats. Args: history: Results of the training step. Supports both categorical_accuracy and sparse_categorical_accuracy. eval_output: Output of the eval step. Assumes first value is eval_loss and second value is accuracy_top_1. callbacks: a list of callbacks which might include a time history callback used during keras.fit. Returns: Dictionary of normalized results. """ stats = {} if eval_output: stats['accuracy_top_1'] = float(eval_output[1]) stats['eval_loss'] = float(eval_output[0]) if history and history.history: train_hist = history.history # Gets final loss from training. stats['loss'] = float(train_hist['loss'][-1]) # Gets top_1 training accuracy. if 'categorical_accuracy' in train_hist: stats[TRAIN_TOP_1] = float(train_hist['categorical_accuracy'][-1]) elif 'sparse_categorical_accuracy' in train_hist: stats[TRAIN_TOP_1] = float(train_hist['sparse_categorical_accuracy'][-1]) elif 'accuracy' in train_hist: stats[TRAIN_TOP_1] = float(train_hist['accuracy'][-1]) if not callbacks: return stats # Look for the time history callback which was used during keras.fit for callback in callbacks: if isinstance(callback, keras_utils.TimeHistory): timestamp_log = callback.timestamp_log stats['step_timestamp_log'] = timestamp_log stats['train_finish_time'] = callback.train_finish_time if callback.epoch_runtime_log: stats['avg_exp_per_second'] = callback.average_examples_per_second return stats def define_keras_flags(model=False, optimizer=False, pretrained_filepath=False): """Define flags for Keras models.""" flags_core.define_base( clean=True, num_gpu=True, run_eagerly=True, train_epochs=True, epochs_between_evals=True, distribution_strategy=True) flags_core.define_performance( num_parallel_calls=False, synthetic_data=True, dtype=True, all_reduce_alg=True, num_packs=True, tf_gpu_thread_mode=True, datasets_num_private_threads=True, loss_scale=True, fp16_implementation=True, tf_data_experimental_slack=True, enable_xla=True, training_dataset_cache=True) flags_core.define_image() flags_core.define_benchmark() flags_core.define_distribution() flags.adopt_module_key_flags(flags_core) flags.DEFINE_boolean(name='enable_eager', default=False, help='Enable eager?') flags.DEFINE_boolean(name='skip_eval', default=False, help='Skip evaluation?') # TODO(b/135607288): Remove this flag once we understand the root cause of # slowdown when setting the learning phase in Keras backend. flags.DEFINE_boolean( name='set_learning_phase_to_train', default=True, help='If skip eval, also set Keras learning phase to 1 (training).') flags.DEFINE_boolean( name='explicit_gpu_placement', default=False, help='If not using distribution strategy, explicitly set device scope ' 'for the Keras training loop.') flags.DEFINE_boolean( name='use_trivial_model', default=False, help='Whether to use a trivial Keras model.') flags.DEFINE_boolean( name='report_accuracy_metrics', default=True, help='Report metrics during training and evaluation.') flags.DEFINE_boolean( name='use_tensor_lr', default=True, help='Use learning rate tensor instead of a callback.') flags.DEFINE_boolean( name='enable_tensorboard', default=False, help='Whether to enable TensorBoard callback.') flags.DEFINE_string( name='profile_steps', default=None, help='Save profiling data to model dir at given range of global steps. The ' 'value must be a comma separated pair of positive integers, specifying ' 'the first and last step to profile. For example, "--profile_steps=2,4" ' 'triggers the profiler to process 3 steps, starting from the 2nd step. ' 'Note that profiler has a non-trivial performance overhead, and the ' 'output file can be gigantic if profiling many steps.') flags.DEFINE_integer( name='train_steps', default=None, help='The number of steps to run for training. If it is larger than ' '# batches per epoch, then use # batches per epoch. This flag will be ' 'ignored if train_epochs is set to be larger than 1. ') flags.DEFINE_boolean( name='batchnorm_spatial_persistent', default=True, help='Enable the spacial persistent mode for CuDNN batch norm kernel.') flags.DEFINE_boolean( name='enable_get_next_as_optional', default=False, help='Enable get_next_as_optional behavior in DistributedIterator.') flags.DEFINE_boolean( name='enable_checkpoint_and_export', default=False, help='Whether to enable a checkpoint callback and export the savedmodel.') flags.DEFINE_string(name='tpu', default='', help='TPU address to connect to.') flags.DEFINE_integer( name='steps_per_loop', default=None, help='Number of steps per training loop. Only training step happens ' 'inside the loop. Callbacks will not be called inside. Will be capped at ' 'steps per epoch.') flags.DEFINE_boolean( name='use_tf_while_loop', default=True, help='Whether to build a tf.while_loop inside the training loop on the ' 'host. Setting it to True is critical to have peak performance on ' 'TPU.') if model: flags.DEFINE_string('model', 'resnet50_v1.5', 'Name of model preset. (mobilenet, resnet50_v1.5)') if optimizer: flags.DEFINE_string( 'optimizer', 'resnet50_default', 'Name of optimizer preset. ' '(mobilenet_default, resnet50_default)') # TODO(kimjaehong): Replace as general hyper-params not only for mobilenet. flags.DEFINE_float( 'initial_learning_rate_per_sample', 0.00007, 'Initial value of learning rate per sample for ' 'mobilenet_default.') flags.DEFINE_float('lr_decay_factor', 0.94, 'Learning rate decay factor for mobilenet_default.') flags.DEFINE_float('num_epochs_per_decay', 2.5, 'Number of epochs per decay for mobilenet_default.') if pretrained_filepath: flags.DEFINE_string('pretrained_filepath', '', 'Pretrained file path.') def get_synth_data(height, width, num_channels, num_classes, dtype): """Creates a set of synthetic random data. Args: height: Integer height that will be used to create a fake image tensor. width: Integer width that will be used to create a fake image tensor. num_channels: Integer depth that will be used to create a fake image tensor. num_classes: Number of classes that should be represented in the fake labels tensor dtype: Data type for features/images. Returns: A tuple of tensors representing the inputs and labels. """ # Synthetic input should be within [0, 255]. inputs = tf.random.truncated_normal([height, width, num_channels], dtype=dtype, mean=127, stddev=60, name='synthetic_inputs') labels = tf.random.uniform([1], minval=0, maxval=num_classes - 1, dtype=tf.int32, name='synthetic_labels') return inputs, labels def define_pruning_flags(): """Define flags for pruning methods.""" flags.DEFINE_string( 'pruning_method', None, 'Pruning method.' 'None (no pruning) or polynomial_decay.') flags.DEFINE_float('pruning_initial_sparsity', 0.0, 'Initial sparsity for pruning.') flags.DEFINE_float('pruning_final_sparsity', 0.5, 'Final sparsity for pruning.') flags.DEFINE_integer('pruning_begin_step', 0, 'Begin step for pruning.') flags.DEFINE_integer('pruning_end_step', 100000, 'End step for pruning.') flags.DEFINE_integer('pruning_frequency', 100, 'Frequency for pruning.') def define_clustering_flags(): """Define flags for clustering methods.""" flags.DEFINE_string('clustering_method', None, 'None (no clustering) or selective_clustering ' '(cluster last three Conv2D layers of the model).') def get_synth_input_fn(height, width, num_channels, num_classes, dtype=tf.float32, drop_remainder=True): """Returns an input function that returns a dataset with random data. This input_fn returns a data set that iterates over a set of random data and bypasses all preprocessing, e.g. jpeg decode and copy. The host to device copy is still included. This used to find the upper throughput bound when tuning the full input pipeline. Args: height: Integer height that will be used to create a fake image tensor. width: Integer width that will be used to create a fake image tensor. num_channels: Integer depth that will be used to create a fake image tensor. num_classes: Number of classes that should be represented in the fake labels tensor dtype: Data type for features/images. drop_remainder: A boolean indicates whether to drop the remainder of the batches. If True, the batch dimension will be static. Returns: An input_fn that can be used in place of a real one to return a dataset that can be used for iteration. """ # pylint: disable=unused-argument def input_fn(is_training, data_dir, batch_size, *args, **kwargs): """Returns dataset filled with random data.""" inputs, labels = get_synth_data( height=height, width=width, num_channels=num_channels, num_classes=num_classes, dtype=dtype) # Cast to float32 for Keras model. labels = tf.cast(labels, dtype=tf.float32) data = tf.data.Dataset.from_tensors((inputs, labels)).repeat() # `drop_remainder` will make dataset produce outputs with known shapes. data = data.batch(batch_size, drop_remainder=drop_remainder) data = data.prefetch(buffer_size=tf.data.experimental.AUTOTUNE) return data return input_fn def set_cudnn_batchnorm_mode(): """Set CuDNN batchnorm mode for better performance. Note: Spatial Persistent mode may lead to accuracy losses for certain models. """ if FLAGS.batchnorm_spatial_persistent: os.environ['TF_USE_CUDNN_BATCHNORM_SPATIAL_PERSISTENT'] = '1' else: os.environ.pop('TF_USE_CUDNN_BATCHNORM_SPATIAL_PERSISTENT', None)
16,294
37.431604
82
py
models
models-master/official/legacy/image_classification/resnet/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/image_classification/resnet/resnet_runnable.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Runs a ResNet model on the ImageNet dataset using custom training loops.""" import orbit import tensorflow as tf from official.legacy.image_classification.resnet import common from official.legacy.image_classification.resnet import imagenet_preprocessing from official.legacy.image_classification.resnet import resnet_model from official.modeling import grad_utils from official.modeling import performance from official.utils.flags import core as flags_core class ResnetRunnable(orbit.StandardTrainer, orbit.StandardEvaluator): """Implements the training and evaluation APIs for Resnet model.""" def __init__(self, flags_obj, time_callback, epoch_steps): self.strategy = tf.distribute.get_strategy() self.flags_obj = flags_obj self.dtype = flags_core.get_tf_dtype(flags_obj) self.time_callback = time_callback # Input pipeline related batch_size = flags_obj.batch_size if batch_size % self.strategy.num_replicas_in_sync != 0: raise ValueError( 'Batch size must be divisible by number of replicas : {}'.format( self.strategy.num_replicas_in_sync)) # As auto rebatching is not supported in # `distribute_datasets_from_function()` API, which is # required when cloning dataset to multiple workers in eager mode, # we use per-replica batch size. self.batch_size = int(batch_size / self.strategy.num_replicas_in_sync) if self.flags_obj.use_synthetic_data: self.input_fn = common.get_synth_input_fn( height=imagenet_preprocessing.DEFAULT_IMAGE_SIZE, width=imagenet_preprocessing.DEFAULT_IMAGE_SIZE, num_channels=imagenet_preprocessing.NUM_CHANNELS, num_classes=imagenet_preprocessing.NUM_CLASSES, dtype=self.dtype, drop_remainder=True) else: self.input_fn = imagenet_preprocessing.input_fn self.model = resnet_model.resnet50( num_classes=imagenet_preprocessing.NUM_CLASSES, use_l2_regularizer=not flags_obj.single_l2_loss_op) lr_schedule = common.PiecewiseConstantDecayWithWarmup( batch_size=flags_obj.batch_size, epoch_size=imagenet_preprocessing.NUM_IMAGES['train'], warmup_epochs=common.LR_SCHEDULE[0][1], boundaries=list(p[1] for p in common.LR_SCHEDULE[1:]), multipliers=list(p[0] for p in common.LR_SCHEDULE), compute_lr_on_cpu=True) self.optimizer = common.get_optimizer(lr_schedule) # Make sure iterations variable is created inside scope. self.global_step = self.optimizer.iterations self.optimizer = performance.configure_optimizer( self.optimizer, use_float16=self.dtype == tf.float16, loss_scale=flags_core.get_loss_scale(flags_obj, default_for_fp16=128)) self.train_loss = tf.keras.metrics.Mean('train_loss', dtype=tf.float32) self.train_accuracy = tf.keras.metrics.SparseCategoricalAccuracy( 'train_accuracy', dtype=tf.float32) self.test_loss = tf.keras.metrics.Mean('test_loss', dtype=tf.float32) self.test_accuracy = tf.keras.metrics.SparseCategoricalAccuracy( 'test_accuracy', dtype=tf.float32) self.checkpoint = tf.train.Checkpoint( model=self.model, optimizer=self.optimizer) # Handling epochs. self.epoch_steps = epoch_steps self.epoch_helper = orbit.utils.EpochHelper(epoch_steps, self.global_step) train_dataset = orbit.utils.make_distributed_dataset( self.strategy, self.input_fn, is_training=True, data_dir=self.flags_obj.data_dir, batch_size=self.batch_size, parse_record_fn=imagenet_preprocessing.parse_record, datasets_num_private_threads=self.flags_obj .datasets_num_private_threads, dtype=self.dtype, drop_remainder=True, training_dataset_cache=self.flags_obj.training_dataset_cache) orbit.StandardTrainer.__init__( self, train_dataset, options=orbit.StandardTrainerOptions( use_tf_while_loop=flags_obj.use_tf_while_loop, use_tf_function=flags_obj.use_tf_function)) if not flags_obj.skip_eval: eval_dataset = orbit.utils.make_distributed_dataset( self.strategy, self.input_fn, is_training=False, data_dir=self.flags_obj.data_dir, batch_size=self.batch_size, parse_record_fn=imagenet_preprocessing.parse_record, dtype=self.dtype) orbit.StandardEvaluator.__init__( self, eval_dataset, options=orbit.StandardEvaluatorOptions( use_tf_function=flags_obj.use_tf_function)) def train_loop_begin(self): """See base class.""" # Reset all metrics self.train_loss.reset_states() self.train_accuracy.reset_states() self._epoch_begin() self.time_callback.on_batch_begin(self.epoch_helper.batch_index) def train_step(self, iterator): """See base class.""" def step_fn(inputs): """Function to run on the device.""" images, labels = inputs with tf.GradientTape() as tape: logits = self.model(images, training=True) prediction_loss = tf.keras.losses.sparse_categorical_crossentropy( labels, logits) loss = tf.reduce_sum(prediction_loss) * (1.0 / self.flags_obj.batch_size) num_replicas = self.strategy.num_replicas_in_sync l2_weight_decay = 1e-4 if self.flags_obj.single_l2_loss_op: l2_loss = l2_weight_decay * 2 * tf.add_n([ tf.nn.l2_loss(v) for v in self.model.trainable_variables if 'bn' not in v.name ]) loss += (l2_loss / num_replicas) else: loss += (tf.reduce_sum(self.model.losses) / num_replicas) grad_utils.minimize_using_explicit_allreduce( tape, self.optimizer, loss, self.model.trainable_variables) self.train_loss.update_state(loss) self.train_accuracy.update_state(labels, logits) if self.flags_obj.enable_xla: step_fn = tf.function(step_fn, jit_compile=True) self.strategy.run(step_fn, args=(next(iterator),)) def train_loop_end(self): """See base class.""" metrics = { 'train_loss': self.train_loss.result(), 'train_accuracy': self.train_accuracy.result(), } self.time_callback.on_batch_end(self.epoch_helper.batch_index - 1) self._epoch_end() return metrics def eval_begin(self): """See base class.""" self.test_loss.reset_states() self.test_accuracy.reset_states() def eval_step(self, iterator): """See base class.""" def step_fn(inputs): """Function to run on the device.""" images, labels = inputs logits = self.model(images, training=False) loss = tf.keras.losses.sparse_categorical_crossentropy(labels, logits) loss = tf.reduce_sum(loss) * (1.0 / self.flags_obj.batch_size) self.test_loss.update_state(loss) self.test_accuracy.update_state(labels, logits) self.strategy.run(step_fn, args=(next(iterator),)) def eval_end(self): """See base class.""" return { 'test_loss': self.test_loss.result(), 'test_accuracy': self.test_accuracy.result() } def _epoch_begin(self): if self.epoch_helper.epoch_begin(): self.time_callback.on_epoch_begin(self.epoch_helper.current_epoch) def _epoch_end(self): if self.epoch_helper.epoch_end(): self.time_callback.on_epoch_end(self.epoch_helper.current_epoch)
8,124
37.507109
78
py
models
models-master/official/legacy/image_classification/resnet/resnet_model.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ResNet50 model for Keras. Adapted from tf.keras.applications.resnet50.ResNet50(). This is ResNet model version 1.5. Related papers/blogs: - https://arxiv.org/abs/1512.03385 - https://arxiv.org/pdf/1603.05027v2.pdf - http://torch.ch/blog/2016/02/04/resnets.html """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from official.legacy.image_classification.resnet import imagenet_preprocessing layers = tf.keras.layers def _gen_l2_regularizer(use_l2_regularizer=True, l2_weight_decay=1e-4): return tf.keras.regularizers.L2( l2_weight_decay) if use_l2_regularizer else None def identity_block(input_tensor, kernel_size, filters, stage, block, use_l2_regularizer=True, batch_norm_decay=0.9, batch_norm_epsilon=1e-5): """The identity block is the block that has no conv layer at shortcut. Args: input_tensor: input tensor kernel_size: default 3, the kernel size of middle conv layer at main path filters: list of integers, the filters of 3 conv layer at main path stage: integer, current stage label, used for generating layer names block: 'a','b'..., current block label, used for generating layer names use_l2_regularizer: whether to use L2 regularizer on Conv layer. batch_norm_decay: Moment of batch norm layers. batch_norm_epsilon: Epsilon of batch borm layers. Returns: Output tensor for the block. """ filters1, filters2, filters3 = filters if tf.keras.backend.image_data_format() == 'channels_last': bn_axis = 3 else: bn_axis = 1 conv_name_base = 'res' + str(stage) + block + '_branch' bn_name_base = 'bn' + str(stage) + block + '_branch' x = layers.Conv2D( filters1, (1, 1), use_bias=False, kernel_initializer='he_normal', kernel_regularizer=_gen_l2_regularizer(use_l2_regularizer), name=conv_name_base + '2a')( input_tensor) x = layers.BatchNormalization( axis=bn_axis, momentum=batch_norm_decay, epsilon=batch_norm_epsilon, name=bn_name_base + '2a')( x) x = layers.Activation('relu')(x) x = layers.Conv2D( filters2, kernel_size, padding='same', use_bias=False, kernel_initializer='he_normal', kernel_regularizer=_gen_l2_regularizer(use_l2_regularizer), name=conv_name_base + '2b')( x) x = layers.BatchNormalization( axis=bn_axis, momentum=batch_norm_decay, epsilon=batch_norm_epsilon, name=bn_name_base + '2b')( x) x = layers.Activation('relu')(x) x = layers.Conv2D( filters3, (1, 1), use_bias=False, kernel_initializer='he_normal', kernel_regularizer=_gen_l2_regularizer(use_l2_regularizer), name=conv_name_base + '2c')( x) x = layers.BatchNormalization( axis=bn_axis, momentum=batch_norm_decay, epsilon=batch_norm_epsilon, name=bn_name_base + '2c')( x) x = layers.add([x, input_tensor]) x = layers.Activation('relu')(x) return x def conv_block(input_tensor, kernel_size, filters, stage, block, strides=(2, 2), use_l2_regularizer=True, batch_norm_decay=0.9, batch_norm_epsilon=1e-5): """A block that has a conv layer at shortcut. Note that from stage 3, the second conv layer at main path is with strides=(2, 2) And the shortcut should have strides=(2, 2) as well Args: input_tensor: input tensor kernel_size: default 3, the kernel size of middle conv layer at main path filters: list of integers, the filters of 3 conv layer at main path stage: integer, current stage label, used for generating layer names block: 'a','b'..., current block label, used for generating layer names strides: Strides for the second conv layer in the block. use_l2_regularizer: whether to use L2 regularizer on Conv layer. batch_norm_decay: Moment of batch norm layers. batch_norm_epsilon: Epsilon of batch borm layers. Returns: Output tensor for the block. """ filters1, filters2, filters3 = filters if tf.keras.backend.image_data_format() == 'channels_last': bn_axis = 3 else: bn_axis = 1 conv_name_base = 'res' + str(stage) + block + '_branch' bn_name_base = 'bn' + str(stage) + block + '_branch' x = layers.Conv2D( filters1, (1, 1), use_bias=False, kernel_initializer='he_normal', kernel_regularizer=_gen_l2_regularizer(use_l2_regularizer), name=conv_name_base + '2a')( input_tensor) x = layers.BatchNormalization( axis=bn_axis, momentum=batch_norm_decay, epsilon=batch_norm_epsilon, name=bn_name_base + '2a')( x) x = layers.Activation('relu')(x) x = layers.Conv2D( filters2, kernel_size, strides=strides, padding='same', use_bias=False, kernel_initializer='he_normal', kernel_regularizer=_gen_l2_regularizer(use_l2_regularizer), name=conv_name_base + '2b')( x) x = layers.BatchNormalization( axis=bn_axis, momentum=batch_norm_decay, epsilon=batch_norm_epsilon, name=bn_name_base + '2b')( x) x = layers.Activation('relu')(x) x = layers.Conv2D( filters3, (1, 1), use_bias=False, kernel_initializer='he_normal', kernel_regularizer=_gen_l2_regularizer(use_l2_regularizer), name=conv_name_base + '2c')( x) x = layers.BatchNormalization( axis=bn_axis, momentum=batch_norm_decay, epsilon=batch_norm_epsilon, name=bn_name_base + '2c')( x) shortcut = layers.Conv2D( filters3, (1, 1), strides=strides, use_bias=False, kernel_initializer='he_normal', kernel_regularizer=_gen_l2_regularizer(use_l2_regularizer), name=conv_name_base + '1')( input_tensor) shortcut = layers.BatchNormalization( axis=bn_axis, momentum=batch_norm_decay, epsilon=batch_norm_epsilon, name=bn_name_base + '1')( shortcut) x = layers.add([x, shortcut]) x = layers.Activation('relu')(x) return x def resnet50(num_classes, batch_size=None, use_l2_regularizer=True, rescale_inputs=False, batch_norm_decay=0.9, batch_norm_epsilon=1e-5): """Instantiates the ResNet50 architecture. Args: num_classes: `int` number of classes for image classification. batch_size: Size of the batches for each step. use_l2_regularizer: whether to use L2 regularizer on Conv/Dense layer. rescale_inputs: whether to rescale inputs from 0 to 1. batch_norm_decay: Moment of batch norm layers. batch_norm_epsilon: Epsilon of batch borm layers. Returns: A Keras model instance. """ input_shape = (224, 224, 3) img_input = layers.Input(shape=input_shape, batch_size=batch_size) if rescale_inputs: # Hub image modules expect inputs in the range [0, 1]. This rescales these # inputs to the range expected by the trained model. x = layers.Lambda( lambda x: x * 255.0 - tf.keras.backend.constant( # pylint: disable=g-long-lambda imagenet_preprocessing.CHANNEL_MEANS, shape=[1, 1, 3], dtype=x.dtype), name='rescale')( img_input) else: x = img_input if tf.keras.backend.image_data_format() == 'channels_first': x = layers.Permute((3, 1, 2))(x) bn_axis = 1 else: # channels_last bn_axis = 3 block_config = dict( use_l2_regularizer=use_l2_regularizer, batch_norm_decay=batch_norm_decay, batch_norm_epsilon=batch_norm_epsilon) x = layers.ZeroPadding2D(padding=(3, 3), name='conv1_pad')(x) x = layers.Conv2D( 64, (7, 7), strides=(2, 2), padding='valid', use_bias=False, kernel_initializer='he_normal', kernel_regularizer=_gen_l2_regularizer(use_l2_regularizer), name='conv1')( x) x = layers.BatchNormalization( axis=bn_axis, momentum=batch_norm_decay, epsilon=batch_norm_epsilon, name='bn_conv1')( x) x = layers.Activation('relu')(x) x = layers.MaxPooling2D((3, 3), strides=(2, 2), padding='same')(x) x = conv_block( x, 3, [64, 64, 256], stage=2, block='a', strides=(1, 1), **block_config) x = identity_block(x, 3, [64, 64, 256], stage=2, block='b', **block_config) x = identity_block(x, 3, [64, 64, 256], stage=2, block='c', **block_config) x = conv_block(x, 3, [128, 128, 512], stage=3, block='a', **block_config) x = identity_block(x, 3, [128, 128, 512], stage=3, block='b', **block_config) x = identity_block(x, 3, [128, 128, 512], stage=3, block='c', **block_config) x = identity_block(x, 3, [128, 128, 512], stage=3, block='d', **block_config) x = conv_block(x, 3, [256, 256, 1024], stage=4, block='a', **block_config) x = identity_block(x, 3, [256, 256, 1024], stage=4, block='b', **block_config) x = identity_block(x, 3, [256, 256, 1024], stage=4, block='c', **block_config) x = identity_block(x, 3, [256, 256, 1024], stage=4, block='d', **block_config) x = identity_block(x, 3, [256, 256, 1024], stage=4, block='e', **block_config) x = identity_block(x, 3, [256, 256, 1024], stage=4, block='f', **block_config) x = conv_block(x, 3, [512, 512, 2048], stage=5, block='a', **block_config) x = identity_block(x, 3, [512, 512, 2048], stage=5, block='b', **block_config) x = identity_block(x, 3, [512, 512, 2048], stage=5, block='c', **block_config) x = layers.GlobalAveragePooling2D()(x) x = layers.Dense( num_classes, kernel_initializer=tf.initializers.random_normal(stddev=0.01), kernel_regularizer=_gen_l2_regularizer(use_l2_regularizer), bias_regularizer=_gen_l2_regularizer(use_l2_regularizer), name='fc1000')( x) # A softmax that is followed by the model loss must be done cannot be done # in float16 due to numeric issues. So we pass dtype=float32. x = layers.Activation('softmax', dtype='float32')(x) # Create model. return tf.keras.Model(img_input, x, name='resnet50')
10,945
32.576687
91
py
models
models-master/official/legacy/transformer/model_utils_test.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Test Transformer model helper methods.""" import tensorflow as tf from official.legacy.transformer import model_utils NEG_INF = -1e9 class ModelUtilsTest(tf.test.TestCase): def test_get_padding(self): x = tf.constant([[1, 0, 0, 0, 2], [3, 4, 0, 0, 0], [0, 5, 6, 0, 7]]) padding = model_utils.get_padding(x, padding_value=0) self.assertAllEqual([[0, 1, 1, 1, 0], [0, 0, 1, 1, 1], [1, 0, 0, 1, 0]], padding) def test_get_padding_bias(self): x = tf.constant([[1, 0, 0, 0, 2], [3, 4, 0, 0, 0], [0, 5, 6, 0, 7]]) bias = model_utils.get_padding_bias(x) bias_shape = tf.shape(bias) flattened_bias = tf.reshape(bias, [3, 5]) self.assertAllEqual( [[0, NEG_INF, NEG_INF, NEG_INF, 0], [0, 0, NEG_INF, NEG_INF, NEG_INF], [NEG_INF, 0, 0, NEG_INF, 0]], flattened_bias) self.assertAllEqual([3, 1, 1, 5], bias_shape) def test_get_decoder_self_attention_bias(self): length = 5 bias = model_utils.get_decoder_self_attention_bias(length) self.assertAllEqual( [[[[0, NEG_INF, NEG_INF, NEG_INF, NEG_INF], [0, 0, NEG_INF, NEG_INF, NEG_INF], [0, 0, 0, NEG_INF, NEG_INF], [0, 0, 0, 0, NEG_INF], [0, 0, 0, 0, 0]]]], bias) if __name__ == "__main__": tf.test.main()
1,891
32.785714
78
py
models
models-master/official/legacy/transformer/transformer_test.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Test Transformer model.""" import tensorflow as tf from official.legacy.transformer import model_params from official.legacy.transformer import transformer class TransformerV2Test(tf.test.TestCase): def setUp(self): super().setUp() self.params = params = model_params.TINY_PARAMS params["batch_size"] = params["default_batch_size"] = 16 params["use_synthetic_data"] = True params["hidden_size"] = 12 params["num_hidden_layers"] = 2 params["filter_size"] = 14 params["num_heads"] = 2 params["vocab_size"] = 41 params["extra_decode_length"] = 2 params["beam_size"] = 3 params["dtype"] = tf.float32 def test_create_model_train(self): model = transformer.create_model(self.params, True) inputs, outputs = model.inputs, model.outputs self.assertEqual(len(inputs), 2) self.assertEqual(len(outputs), 1) self.assertEqual(inputs[0].shape.as_list(), [None, None]) self.assertEqual(inputs[0].dtype, tf.int64) self.assertEqual(inputs[1].shape.as_list(), [None, None]) self.assertEqual(inputs[1].dtype, tf.int64) self.assertEqual(outputs[0].shape.as_list(), [None, None, 41]) self.assertEqual(outputs[0].dtype, tf.float32) def test_create_model_not_train(self): model = transformer.create_model(self.params, False) inputs, outputs = model.inputs, model.outputs self.assertEqual(len(inputs), 1) self.assertEqual(len(outputs), 2) self.assertEqual(inputs[0].shape.as_list(), [None, None]) self.assertEqual(inputs[0].dtype, tf.int64) self.assertEqual(outputs[0].shape.as_list(), [None, None]) self.assertEqual(outputs[0].dtype, tf.int32) self.assertEqual(outputs[1].shape.as_list(), [None]) self.assertEqual(outputs[1].dtype, tf.float32) def test_export(self): model = transformer.Transformer(self.params, name="transformer_v2") export_dir = self.get_temp_dir() batch_size = 5 max_length = 6 class SaveModule(tf.Module): def __init__(self, model): super(SaveModule, self).__init__() self.model = model @tf.function def serve(self, x): return self.model.call([x], training=False) save_module = SaveModule(model) tensor_shape = (None, None) sample_input = tf.zeros((batch_size, max_length), dtype=tf.int64) _ = save_module.serve(sample_input) signatures = dict( serving_default=save_module.serve.get_concrete_function( tf.TensorSpec(shape=tensor_shape, dtype=tf.int64, name="x"))) tf.saved_model.save(save_module, export_dir, signatures=signatures) imported = tf.saved_model.load(export_dir) serving_fn = imported.signatures["serving_default"] all_outputs = serving_fn(sample_input) output = all_outputs["outputs"] output_shapes = output.shape.as_list() self.assertEqual(output_shapes[0], batch_size) self.assertEqual(output_shapes[1], max_length + model.params["extra_decode_length"]) if __name__ == "__main__": tf.test.main()
3,628
35.656566
74
py
models
models-master/official/legacy/transformer/model_utils.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Transformer model helper methods.""" import math import numpy as np import tensorflow as tf # Very low numbers to represent -infinity. We do not actually use -Inf, since we # want to be able to multiply these values by zero to get zero. (-Inf * 0 = NaN) _NEG_INF_FP32 = -1e9 _NEG_INF_FP16 = np.finfo(np.float16).min def get_position_encoding(length, hidden_size, min_timescale=1.0, max_timescale=1.0e4): """Return positional encoding. Calculates the position encoding as a mix of sine and cosine functions with geometrically increasing wavelengths. Defined and formulized in Attention is All You Need, section 3.5. Args: length: Sequence length. hidden_size: Size of the min_timescale: Minimum scale that will be applied at each position max_timescale: Maximum scale that will be applied at each position Returns: Tensor with shape [length, hidden_size] """ # We compute the positional encoding in float32 even if the model uses # float16, as many of the ops used, like log and exp, are numerically unstable # in float16. position = tf.cast(tf.range(length), tf.float32) num_timescales = hidden_size // 2 log_timescale_increment = ( math.log(float(max_timescale) / float(min_timescale)) / (tf.cast(num_timescales, tf.float32) - 1)) inv_timescales = min_timescale * tf.exp( tf.cast(tf.range(num_timescales), tf.float32) * -log_timescale_increment) scaled_time = tf.expand_dims(position, 1) * tf.expand_dims(inv_timescales, 0) signal = tf.concat([tf.sin(scaled_time), tf.cos(scaled_time)], axis=1) return signal def get_decoder_self_attention_bias(length, dtype=tf.float32): """Calculate bias for decoder that maintains model's autoregressive property. Creates a tensor that masks out locations that correspond to illegal connections, so prediction at position i cannot draw information from future positions. Args: length: int length of sequences in batch. dtype: The dtype of the return value. Returns: float tensor of shape [1, 1, length, length] """ neg_inf = _NEG_INF_FP16 if dtype == tf.float16 else _NEG_INF_FP32 with tf.name_scope("decoder_self_attention_bias"): valid_locs = tf.linalg.band_part( tf.ones([length, length], dtype=dtype), -1, 0) valid_locs = tf.reshape(valid_locs, [1, 1, length, length]) decoder_bias = neg_inf * (1.0 - valid_locs) return decoder_bias def get_padding(x, padding_value=0, dtype=tf.float32): """Return float tensor representing the padding values in x. Args: x: int tensor with any shape padding_value: int which represents padded values in input dtype: The dtype of the return value. Returns: float tensor with same shape as x containing values 0 or 1. 0 -> non-padding, 1 -> padding """ with tf.name_scope("padding"): return tf.cast(tf.equal(x, padding_value), dtype) def get_padding_bias(x, padding_value=0, dtype=tf.float32): """Calculate bias tensor from padding values in tensor. Bias tensor that is added to the pre-softmax multi-headed attention logits, which has shape [batch_size, num_heads, length, length]. The tensor is zero at non-padding locations, and -1e9 (negative infinity) at padding locations. Args: x: int tensor with shape [batch_size, length] padding_value: int which represents padded values in input dtype: The dtype of the return value Returns: Attention bias tensor of shape [batch_size, 1, 1, length]. """ with tf.name_scope("attention_bias"): padding = get_padding(x, padding_value, dtype) attention_bias = padding * _NEG_INF_FP32 attention_bias = tf.expand_dims( tf.expand_dims(attention_bias, axis=1), axis=1) return attention_bias
4,427
35.295082
80
py
models
models-master/official/legacy/transformer/transformer_main_test.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Test Transformer model.""" import os import re import sys import unittest from absl import flags from absl.testing import flagsaver import tensorflow as tf from tensorflow.python.eager import context # pylint: disable=ungrouped-imports from official.legacy.transformer import misc from official.legacy.transformer import transformer_main FLAGS = flags.FLAGS FIXED_TIMESTAMP = 'my_time_stamp' WEIGHT_PATTERN = re.compile(r'weights-epoch-.+\.hdf5') def _generate_file(filepath, lines): with open(filepath, 'w') as f: for l in lines: f.write('{}\n'.format(l)) class TransformerTaskTest(tf.test.TestCase): local_flags = None def setUp(self): # pylint: disable=g-missing-super-call temp_dir = self.get_temp_dir() if TransformerTaskTest.local_flags is None: misc.define_transformer_flags() # Loads flags, array cannot be blank. flags.FLAGS(['foo']) TransformerTaskTest.local_flags = flagsaver.save_flag_values() else: flagsaver.restore_flag_values(TransformerTaskTest.local_flags) FLAGS.model_dir = os.path.join(temp_dir, FIXED_TIMESTAMP) FLAGS.param_set = 'tiny' FLAGS.use_synthetic_data = True FLAGS.steps_between_evals = 1 FLAGS.train_steps = 1 FLAGS.validation_steps = 1 FLAGS.batch_size = 4 FLAGS.max_length = 1 FLAGS.num_gpus = 1 FLAGS.distribution_strategy = 'off' FLAGS.dtype = 'fp32' self.model_dir = FLAGS.model_dir self.temp_dir = temp_dir self.vocab_file = os.path.join(temp_dir, 'vocab') self.vocab_size = misc.get_model_params(FLAGS.param_set, 0)['vocab_size'] self.bleu_source = os.path.join(temp_dir, 'bleu_source') self.bleu_ref = os.path.join(temp_dir, 'bleu_ref') self.orig_policy = ( tf.compat.v2.keras.mixed_precision.global_policy()) def tearDown(self): # pylint: disable=g-missing-super-call tf.compat.v2.keras.mixed_precision.set_global_policy(self.orig_policy) def _assert_exists(self, filepath): self.assertTrue(os.path.exists(filepath)) def test_train_no_dist_strat(self): if context.num_gpus() >= 2: self.skipTest('No need to test 2+ GPUs without a distribution strategy.') t = transformer_main.TransformerTask(FLAGS) t.train() def test_train_save_full_model(self): if context.num_gpus() >= 2: self.skipTest('No need to test 2+ GPUs without a distribution strategy.') FLAGS.save_weights_only = False t = transformer_main.TransformerTask(FLAGS) t.train() def test_train_static_batch(self): if context.num_gpus() >= 2: self.skipTest('No need to test 2+ GPUs without a distribution strategy.') FLAGS.distribution_strategy = 'one_device' if tf.test.is_built_with_cuda(): FLAGS.num_gpus = 1 else: FLAGS.num_gpus = 0 FLAGS.static_batch = True t = transformer_main.TransformerTask(FLAGS) t.train() @unittest.skipUnless(tf.test.is_built_with_cuda(), 'requires GPU') def test_train_1_gpu_with_dist_strat(self): FLAGS.distribution_strategy = 'one_device' t = transformer_main.TransformerTask(FLAGS) t.train() @unittest.skipUnless(tf.test.is_built_with_cuda(), 'requires GPU') def test_train_fp16(self): FLAGS.distribution_strategy = 'one_device' FLAGS.dtype = 'fp16' t = transformer_main.TransformerTask(FLAGS) t.train() @unittest.skipUnless(tf.test.is_built_with_cuda(), 'requires GPU') def test_train_2_gpu(self): if context.num_gpus() < 2: self.skipTest( '{} GPUs are not available for this test. {} GPUs are available' .format(2, context.num_gpus())) FLAGS.distribution_strategy = 'mirrored' FLAGS.num_gpus = 2 FLAGS.param_set = 'base' t = transformer_main.TransformerTask(FLAGS) t.train() @unittest.skipUnless(tf.test.is_built_with_cuda(), 'requires GPU') def test_train_2_gpu_fp16(self): if context.num_gpus() < 2: self.skipTest( '{} GPUs are not available for this test. {} GPUs are available' .format(2, context.num_gpus())) FLAGS.distribution_strategy = 'mirrored' FLAGS.num_gpus = 2 FLAGS.param_set = 'base' FLAGS.dtype = 'fp16' t = transformer_main.TransformerTask(FLAGS) t.train() def _prepare_files_and_flags(self, *extra_flags): # Make log dir. if not os.path.exists(self.temp_dir): os.makedirs(self.temp_dir) # Fake vocab, bleu_source and bleu_ref. tokens = [ "'<pad>'", "'<EOS>'", "'_'", "'a'", "'b'", "'c'", "'d'", "'a_'", "'b_'", "'c_'", "'d_'" ] tokens += ["'{}'".format(i) for i in range(self.vocab_size - len(tokens))] _generate_file(self.vocab_file, tokens) _generate_file(self.bleu_source, ['a b', 'c d']) _generate_file(self.bleu_ref, ['a b', 'd c']) # Update flags. update_flags = [ 'ignored_program_name', '--vocab_file={}'.format(self.vocab_file), '--bleu_source={}'.format(self.bleu_source), '--bleu_ref={}'.format(self.bleu_ref), ] if extra_flags: update_flags.extend(extra_flags) FLAGS(update_flags) def test_predict(self): if context.num_gpus() >= 2: self.skipTest('No need to test 2+ GPUs without a distribution strategy.') self._prepare_files_and_flags() t = transformer_main.TransformerTask(FLAGS) t.predict() @unittest.skipUnless(tf.test.is_built_with_cuda(), 'requires GPU') def test_predict_fp16(self): if context.num_gpus() >= 2: self.skipTest('No need to test 2+ GPUs without a distribution strategy.') self._prepare_files_and_flags('--dtype=fp16') t = transformer_main.TransformerTask(FLAGS) t.predict() def test_eval(self): if context.num_gpus() >= 2: self.skipTest('No need to test 2+ GPUs without a distribution strategy.') if 'test_xla' in sys.argv[0]: self.skipTest('TODO(xla): Make this test faster under XLA.') self._prepare_files_and_flags() t = transformer_main.TransformerTask(FLAGS) t.eval() if __name__ == '__main__': tf.test.main()
6,631
33.185567
80
py
models
models-master/official/legacy/transformer/beam_search_v1.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Beam search to find the translated sequence with the highest probability.""" import tensorflow.compat.v1 as tf from official.nlp.modeling.ops import beam_search _StateKeys = beam_search._StateKeys # pylint: disable=protected-access class SequenceBeamSearch(beam_search.SequenceBeamSearch): """Implementation of beam search loop.""" def _process_finished_state(self, finished_state): alive_seq = finished_state[_StateKeys.ALIVE_SEQ] alive_log_probs = finished_state[_StateKeys.ALIVE_LOG_PROBS] finished_seq = finished_state[_StateKeys.FINISHED_SEQ] finished_scores = finished_state[_StateKeys.FINISHED_SCORES] finished_flags = finished_state[_StateKeys.FINISHED_FLAGS] # Account for corner case where there are no finished sequences for a # particular batch item. In that case, return alive sequences for that batch # item. finished_seq = tf.where( tf.reduce_any(finished_flags, 1), finished_seq, alive_seq) finished_scores = tf.where( tf.reduce_any(finished_flags, 1), finished_scores, alive_log_probs) return finished_seq, finished_scores def sequence_beam_search(symbols_to_logits_fn, initial_ids, initial_cache, vocab_size, beam_size, alpha, max_decode_length, eos_id, padded_decode=False): """Search for sequence of subtoken ids with the largest probability. Args: symbols_to_logits_fn: A function that takes in ids, index, and cache as arguments. The passed in arguments will have shape: ids -> A tensor with shape [batch_size * beam_size, index]. index -> A scalar. cache -> A nested dictionary of tensors [batch_size * beam_size, ...]. The function must return a tuple of logits and new cache: logits -> A tensor with shape [batch * beam_size, vocab_size]. new cache -> A nested dictionary with the same shape/structure as the inputted cache. initial_ids: An int32 tensor with shape [batch_size]. Starting ids for each batch item. initial_cache: A dictionary, containing starting decoder variables information. vocab_size: An integer, the size of the vocabulary, used for topk computation. beam_size: An integer, the number of beams. alpha: A float, defining the strength of length normalization. max_decode_length: An integer, the maximum length to decoded a sequence. eos_id: An integer, ID of eos token, used to determine when a sequence has finished. padded_decode: A bool, indicating if max_sequence_length padding is used for beam search. Returns: Top decoded sequences [batch_size, beam_size, max_decode_length] sequence scores [batch_size, beam_size] """ sbs = SequenceBeamSearch(symbols_to_logits_fn, vocab_size, beam_size, alpha, max_decode_length, eos_id, padded_decode) return sbs.search(initial_ids, initial_cache)
3,679
43.337349
80
py
models
models-master/official/legacy/transformer/transformer_forward_test.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Forward pass test for Transformer model refactoring.""" import numpy as np import tensorflow as tf from official.legacy.transformer import metrics from official.legacy.transformer import model_params from official.legacy.transformer import transformer from official.nlp.modeling import models def _count_params(layer, trainable_only=True): """Returns the count of all model parameters, or just trainable ones.""" if not trainable_only: return layer.count_params() else: return int( np.sum([ tf.keras.backend.count_params(p) for p in layer.trainable_weights ])) def _create_model(params, is_train): """Creates transformer model.""" encdec_kwargs = dict( num_layers=params["num_hidden_layers"], num_attention_heads=params["num_heads"], intermediate_size=params["filter_size"], activation="relu", dropout_rate=params["relu_dropout"], attention_dropout_rate=params["attention_dropout"], use_bias=False, norm_first=True, norm_epsilon=1e-6, intermediate_dropout=params["relu_dropout"]) encoder_layer = models.TransformerEncoder(**encdec_kwargs) decoder_layer = models.TransformerDecoder(**encdec_kwargs) model_kwargs = dict( vocab_size=params["vocab_size"], embedding_width=params["hidden_size"], dropout_rate=params["layer_postprocess_dropout"], padded_decode=params["padded_decode"], decode_max_length=params["decode_max_length"], dtype=params["dtype"], extra_decode_length=params["extra_decode_length"], beam_size=params["beam_size"], alpha=params["alpha"], encoder_layer=encoder_layer, decoder_layer=decoder_layer, name="transformer_v2") if is_train: inputs = tf.keras.layers.Input((None,), dtype="int64", name="inputs") targets = tf.keras.layers.Input((None,), dtype="int64", name="targets") internal_model = models.Seq2SeqTransformer(**model_kwargs) logits = internal_model( dict(inputs=inputs, targets=targets), training=is_train) vocab_size = params["vocab_size"] label_smoothing = params["label_smoothing"] if params["enable_metrics_in_training"]: logits = metrics.MetricLayer(vocab_size)([logits, targets]) logits = tf.keras.layers.Lambda( lambda x: x, name="logits", dtype=tf.float32)( logits) model = tf.keras.Model([inputs, targets], logits) loss = metrics.transformer_loss(logits, targets, label_smoothing, vocab_size) model.add_loss(loss) return model batch_size = params["decode_batch_size"] if params["padded_decode"] else None inputs = tf.keras.layers.Input((None,), batch_size=batch_size, dtype="int64", name="inputs") internal_model = models.Seq2SeqTransformer(**model_kwargs) ret = internal_model(dict(inputs=inputs), training=is_train) outputs, scores = ret["outputs"], ret["scores"] return tf.keras.Model(inputs, [outputs, scores]) class TransformerForwardTest(tf.test.TestCase): def setUp(self): super(TransformerForwardTest, self).setUp() self.params = params = model_params.TINY_PARAMS params["batch_size"] = params["default_batch_size"] = 16 params["hidden_size"] = 12 params["num_hidden_layers"] = 3 params["filter_size"] = 14 params["num_heads"] = 2 params["vocab_size"] = 41 params["extra_decode_length"] = 0 params["beam_size"] = 3 params["dtype"] = tf.float32 params["layer_postprocess_dropout"] = 0.0 params["attention_dropout"] = 0.0 params["relu_dropout"] = 0.0 def test_forward_pass_train(self): # Set input_len different from target_len inputs = np.asarray([[5, 2, 1], [7, 5, 0], [1, 4, 0], [7, 5, 11]]) targets = np.asarray([[4, 3, 4, 0], [13, 19, 17, 8], [20, 14, 1, 2], [5, 7, 3, 0]]) # src_model is the original model before refactored. src_model = transformer.create_model(self.params, True) src_num_weights = _count_params(src_model) src_weights = src_model.get_weights() src_model_output = src_model([inputs, targets], training=True) # dest_model is the refactored model. dest_model = _create_model(self.params, True) dest_num_weights = _count_params(dest_model) self.assertEqual(src_num_weights, dest_num_weights) dest_model.set_weights(src_weights) dest_model_output = dest_model([inputs, targets], training=True) self.assertAllEqual(src_model_output, dest_model_output) def test_forward_pass_not_train(self): inputs = np.asarray([[5, 2, 1], [7, 5, 0], [1, 4, 0], [7, 5, 11]]) # src_model is the original model before refactored. src_model = transformer.create_model(self.params, False) src_num_weights = _count_params(src_model) src_weights = src_model.get_weights() src_model_output = src_model([inputs], training=False) # dest_model is the refactored model. dest_model = _create_model(self.params, False) dest_num_weights = _count_params(dest_model) self.assertEqual(src_num_weights, dest_num_weights) dest_model.set_weights(src_weights) dest_model_output = dest_model([inputs], training=False) self.assertAllEqual(src_model_output[0], dest_model_output[0]) self.assertAllEqual(src_model_output[1], dest_model_output[1]) if __name__ == "__main__": tf.test.main()
6,060
37.605096
79
py
models
models-master/official/legacy/transformer/misc.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Misc for Transformer.""" # pylint: disable=g-bad-import-order from absl import flags import tensorflow as tf from official.legacy.transformer import model_params from official.utils.flags import core as flags_core from official.utils.misc import keras_utils FLAGS = flags.FLAGS PARAMS_MAP = { 'tiny': model_params.TINY_PARAMS, 'base': model_params.BASE_PARAMS, 'big': model_params.BIG_PARAMS, } def get_model_params(param_set, num_gpus): """Gets predefined model params.""" if num_gpus > 1: if param_set == 'big': return model_params.BIG_MULTI_GPU_PARAMS.copy() elif param_set == 'base': return model_params.BASE_MULTI_GPU_PARAMS.copy() else: raise ValueError('Not valid params: param_set={} num_gpus={}'.format( param_set, num_gpus)) return PARAMS_MAP[param_set].copy() def define_transformer_flags(): """Add flags and flag validators for running transformer_main.""" # Add common flags (data_dir, model_dir, etc.). flags_core.define_base(num_gpu=True, distribution_strategy=True) flags_core.define_performance( num_parallel_calls=True, inter_op=False, intra_op=False, synthetic_data=True, max_train_steps=False, dtype=True, loss_scale=True, all_reduce_alg=True, num_packs=True, tf_gpu_thread_mode=True, datasets_num_private_threads=True, enable_xla=True, fp16_implementation=True) flags_core.define_benchmark() flags_core.define_device(tpu=True) flags.DEFINE_integer( name='train_steps', short_name='ts', default=300000, help=flags_core.help_wrap('The number of steps used to train.')) flags.DEFINE_integer( name='steps_between_evals', short_name='sbe', default=5000, help=flags_core.help_wrap( 'The Number of training steps to run between evaluations. This is ' 'used if --train_steps is defined.')) flags.DEFINE_boolean( name='enable_time_history', default=True, help='Whether to enable TimeHistory callback.') flags.DEFINE_boolean( name='enable_tensorboard', default=False, help='Whether to enable Tensorboard callback.') flags.DEFINE_boolean( name='enable_metrics_in_training', default=False, help='Whether to enable metrics during training.') flags.DEFINE_boolean( name='enable_mlir_bridge', default=False, help='Whether to enable the TF to XLA bridge.') # Set flags from the flags_core module as 'key flags' so they're listed when # the '-h' flag is used. Without this line, the flags defined above are # only shown in the full `--helpful` help text. flags.adopt_module_key_flags(flags_core) # Add transformer-specific flags flags.DEFINE_enum( name='param_set', short_name='mp', default='big', enum_values=PARAMS_MAP.keys(), help=flags_core.help_wrap( 'Parameter set to use when creating and training the model. The ' 'parameters define the input shape (batch size and max length), ' 'model configuration (size of embedding, # of hidden layers, etc.), ' 'and various other settings. The big parameter set increases the ' 'default batch size, embedding/hidden size, and filter size. For a ' 'complete list of parameters, please see model/model_params.py.')) flags.DEFINE_bool( name='static_batch', short_name='sb', default=False, help=flags_core.help_wrap( 'Whether the batches in the dataset should have static shapes. In ' 'general, this setting should be False. Dynamic shapes allow the ' 'inputs to be grouped so that the number of padding tokens is ' 'minimized, and helps model training. In cases where the input shape ' 'must be static (e.g. running on TPU), this setting will be ignored ' 'and static batching will always be used.')) flags.DEFINE_integer( name='max_length', short_name='ml', default=256, help=flags_core.help_wrap( 'Max sentence length for Transformer. Default is 256. Note: Usually ' 'it is more effective to use a smaller max length if static_batch is ' 'enabled, e.g. 64.')) # Flags for training with steps (may be used for debugging) flags.DEFINE_integer( name='validation_steps', short_name='vs', default=64, help=flags_core.help_wrap('The number of steps used in validation.')) # BLEU score computation flags.DEFINE_string( name='bleu_source', short_name='bls', default=None, help=flags_core.help_wrap( 'Path to source file containing text translate when calculating the ' 'official BLEU score. Both --bleu_source and --bleu_ref must be set. ' )) flags.DEFINE_string( name='bleu_ref', short_name='blr', default=None, help=flags_core.help_wrap( 'Path to source file containing text translate when calculating the ' 'official BLEU score. Both --bleu_source and --bleu_ref must be set. ' )) flags.DEFINE_string( name='vocab_file', short_name='vf', default=None, help=flags_core.help_wrap( 'Path to subtoken vocabulary file. If data_download.py was used to ' 'download and encode the training data, look in the data_dir to find ' 'the vocab file.')) flags.DEFINE_string( name='mode', default='train', help=flags_core.help_wrap('mode: train, eval, or predict')) flags.DEFINE_bool( name='use_ctl', default=False, help=flags_core.help_wrap( 'Whether the model runs with custom training loop.')) flags.DEFINE_integer( name='decode_batch_size', default=32, help=flags_core.help_wrap( 'Global batch size used for Transformer autoregressive decoding on ' 'TPU.')) flags.DEFINE_integer( name='decode_max_length', default=97, help=flags_core.help_wrap( 'Max sequence length of the decode/eval data. This is used by ' 'Transformer autoregressive decoding on TPU to have minimum ' 'paddings.')) flags.DEFINE_bool( name='padded_decode', default=False, help=flags_core.help_wrap( 'Whether the autoregressive decoding runs with input data padded to ' 'the decode_max_length. For TPU/XLA-GPU runs, this flag has to be ' 'set due the static shape requirement. Although CPU/GPU could also ' 'use padded_decode, it has not been tested. In addition, this method ' 'will introduce unnecessary overheads which grow quadratically with ' 'the max sequence length.')) flags.DEFINE_bool( name='enable_checkpointing', default=True, help=flags_core.help_wrap( 'Whether to do checkpointing during training. When running under ' 'benchmark harness, we will avoid checkpointing.')) flags.DEFINE_bool( name='save_weights_only', default=True, help=flags_core.help_wrap( 'Only used when above `enable_checkpointing` is True. ' 'If True, then only the model\'s weights will be saved ' '(`model.save_weights(filepath)`), else the full model is saved ' '(`model.save(filepath)`)')) flags_core.set_defaults( data_dir='/tmp/translate_ende', model_dir='/tmp/transformer_model', batch_size=None) # pylint: disable=unused-variable @flags.multi_flags_validator( ['bleu_source', 'bleu_ref'], message='Both or neither --bleu_source and --bleu_ref must be defined.') def _check_bleu_files(flags_dict): return (flags_dict['bleu_source'] is None) == ( flags_dict['bleu_ref'] is None) @flags.multi_flags_validator( ['bleu_source', 'bleu_ref', 'vocab_file'], message='--vocab_file must be defined if --bleu_source and --bleu_ref ' 'are defined.') def _check_bleu_vocab_file(flags_dict): if flags_dict['bleu_source'] and flags_dict['bleu_ref']: return flags_dict['vocab_file'] is not None return True # pylint: enable=unused-variable def get_callbacks(): """Returns common callbacks.""" callbacks = [] if FLAGS.enable_time_history: time_callback = keras_utils.TimeHistory( FLAGS.batch_size, FLAGS.log_steps, logdir=FLAGS.model_dir if FLAGS.enable_tensorboard else None) callbacks.append(time_callback) if FLAGS.enable_tensorboard: tensorboard_callback = tf.keras.callbacks.TensorBoard( log_dir=FLAGS.model_dir) callbacks.append(tensorboard_callback) return callbacks def update_stats(history, stats, callbacks): """Normalizes and updates dictionary of stats. Args: history: Results of the training step. stats: Dict with pre-existing training stats. callbacks: a list of callbacks which might include a time history callback used during keras.fit. """ if history and history.history: train_hist = history.history # Gets final loss from training. stats['loss'] = float(train_hist['loss'][-1]) if not callbacks: return # Look for the time history callback which was used during keras.fit for callback in callbacks: if isinstance(callback, keras_utils.TimeHistory): timestamp_log = callback.timestamp_log stats['step_timestamp_log'] = timestamp_log stats['train_finish_time'] = callback.train_finish_time if len(timestamp_log) > 1: stats['avg_exp_per_second'] = ( callback.batch_size * callback.log_steps * (len(callback.timestamp_log) - 1) / (timestamp_log[-1].timestamp - timestamp_log[0].timestamp))
10,346
34.802768
80
py
models
models-master/official/legacy/transformer/translate.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Translate text or files using trained transformer model.""" # Import libraries from absl import logging import numpy as np import tensorflow as tf from official.legacy.transformer.utils import tokenizer _EXTRA_DECODE_LENGTH = 100 _BEAM_SIZE = 4 _ALPHA = 0.6 def _get_sorted_inputs(filename): """Read and sort lines from the file sorted by decreasing length. Args: filename: String name of file to read inputs from. Returns: Sorted list of inputs, and dictionary mapping original index->sorted index of each element. """ with tf.io.gfile.GFile(filename) as f: records = f.read().split("\n") inputs = [record.strip() for record in records] if not inputs[-1]: inputs.pop() input_lens = [(i, len(line.split())) for i, line in enumerate(inputs)] sorted_input_lens = sorted(input_lens, key=lambda x: x[1], reverse=True) sorted_inputs = [None] * len(sorted_input_lens) sorted_keys = [0] * len(sorted_input_lens) for i, (index, _) in enumerate(sorted_input_lens): sorted_inputs[i] = inputs[index] sorted_keys[index] = i return sorted_inputs, sorted_keys def _encode_and_add_eos(line, subtokenizer): """Encode line with subtokenizer, and add EOS id to the end.""" return subtokenizer.encode(line) + [tokenizer.EOS_ID] def _trim_and_decode(ids, subtokenizer): """Trim EOS and PAD tokens from ids, and decode to return a string.""" try: index = list(ids).index(tokenizer.EOS_ID) return subtokenizer.decode(ids[:index]) except ValueError: # No EOS found in sequence return subtokenizer.decode(ids) def translate_file(model, params, subtokenizer, input_file, output_file=None, print_all_translations=True, distribution_strategy=None): """Translate lines in file, and save to output file if specified. Args: model: A Keras model, used to generate the translations. params: A dictionary, containing the translation related parameters. subtokenizer: A subtokenizer object, used for encoding and decoding source and translated lines. input_file: A file containing lines to translate. output_file: A file that stores the generated translations. print_all_translations: A bool. If true, all translations are printed to stdout. distribution_strategy: A distribution strategy, used to perform inference directly with tf.function instead of Keras model.predict(). Raises: ValueError: if output file is invalid. """ batch_size = params["decode_batch_size"] # Read and sort inputs by length. Keep dictionary (original index-->new index # in sorted list) to write translations in the original order. sorted_inputs, sorted_keys = _get_sorted_inputs(input_file) total_samples = len(sorted_inputs) num_decode_batches = (total_samples - 1) // batch_size + 1 def input_generator(): """Yield encoded strings from sorted_inputs.""" for i in range(num_decode_batches): lines = [ sorted_inputs[j + i * batch_size] for j in range(batch_size) if j + i * batch_size < total_samples ] lines = [_encode_and_add_eos(l, subtokenizer) for l in lines] if distribution_strategy: for j in range(batch_size - len(lines)): lines.append([tokenizer.EOS_ID]) batch = tf.keras.preprocessing.sequence.pad_sequences( lines, maxlen=params["decode_max_length"], dtype="int32", padding="post") logging.info("Decoding batch %d out of %d.", i, num_decode_batches) yield batch @tf.function def predict_step(inputs): """Decoding step function for TPU runs.""" def _step_fn(inputs): """Per replica step function.""" tag = inputs[0] val_inputs = inputs[1] val_outputs, _ = model([val_inputs], training=False) return tag, val_outputs return distribution_strategy.run(_step_fn, args=(inputs,)) translations = [] if distribution_strategy: num_replicas = distribution_strategy.num_replicas_in_sync local_batch_size = params["decode_batch_size"] // num_replicas for i, text in enumerate(input_generator()): if distribution_strategy: text = np.reshape(text, [num_replicas, local_batch_size, -1]) # Add tag to the input of each replica with the reordering logic after # outputs, to ensure the output order matches the input order. text = tf.constant(text) @tf.function def text_as_per_replica(): replica_context = tf.distribute.get_replica_context() replica_id = replica_context.replica_id_in_sync_group return replica_id, text[replica_id] # pylint: disable=cell-var-from-loop text = distribution_strategy.run(text_as_per_replica) outputs = distribution_strategy.experimental_local_results( predict_step(text)) val_outputs = [output for _, output in outputs] val_outputs = np.reshape(val_outputs, [params["decode_batch_size"], -1]) else: val_outputs, _ = model.predict(text) length = len(val_outputs) for j in range(length): if j + i * batch_size < total_samples: translation = _trim_and_decode(val_outputs[j], subtokenizer) translations.append(translation) if print_all_translations: logging.info("Translating:\n\tInput: %s\n\tOutput: %s", sorted_inputs[j + i * batch_size], translation) # Write translations in the order they appeared in the original file. if output_file is not None: if tf.io.gfile.isdir(output_file): raise ValueError("File output is a directory, will not save outputs to " "file.") logging.info("Writing to file %s", output_file) with tf.io.gfile.GFile(output_file, "w") as f: for i in sorted_keys: f.write("%s\n" % translations[i]) def translate_from_text(model, subtokenizer, txt): encoded_txt = _encode_and_add_eos(txt, subtokenizer) result = model.predict(encoded_txt) outputs = result["outputs"] logging.info("Original: \"%s\"", txt) translate_from_input(outputs, subtokenizer) def translate_from_input(outputs, subtokenizer): translation = _trim_and_decode(outputs, subtokenizer) logging.info("Translation: \"%s\"", translation)
6,951
35.397906
81
py
models
models-master/official/legacy/transformer/compute_bleu.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Script to compute official BLEU score. Source: https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/bleu_hook.py """ import re import sys import unicodedata from absl import app from absl import flags from absl import logging import six from six.moves import range import tensorflow as tf from official.legacy.transformer.utils import metrics from official.legacy.transformer.utils import tokenizer from official.utils.flags import core as flags_core class UnicodeRegex(object): """Ad-hoc hack to recognize all punctuation and symbols.""" def __init__(self): punctuation = self.property_chars("P") self.nondigit_punct_re = re.compile(r"([^\d])([" + punctuation + r"])") self.punct_nondigit_re = re.compile(r"([" + punctuation + r"])([^\d])") self.symbol_re = re.compile("([" + self.property_chars("S") + "])") def property_chars(self, prefix): return "".join( six.unichr(x) for x in range(sys.maxunicode) if unicodedata.category(six.unichr(x)).startswith(prefix)) uregex = UnicodeRegex() def bleu_tokenize(string): r"""Tokenize a string following the official BLEU implementation. See https://github.com/moses-smt/mosesdecoder/' 'blob/master/scripts/generic/mteval-v14.pl#L954-L983 In our case, the input string is expected to be just one line and no HTML entities de-escaping is needed. So we just tokenize on punctuation and symbols, except when a punctuation is preceded and followed by a digit (e.g. a comma/dot as a thousand/decimal separator). Note that a numer (e.g. a year) followed by a dot at the end of sentence is NOT tokenized, i.e. the dot stays with the number because `s/(\p{P})(\P{N})/ $1 $2/g` does not match this case (unless we add a space after each sentence). However, this error is already in the original mteval-v14.pl and we want to be consistent with it. Args: string: the input string Returns: a list of tokens """ string = uregex.nondigit_punct_re.sub(r"\1 \2 ", string) string = uregex.punct_nondigit_re.sub(r" \1 \2", string) string = uregex.symbol_re.sub(r" \1 ", string) return string.split() def bleu_wrapper(ref_filename, hyp_filename, case_sensitive=False): """Compute BLEU for two files (reference and hypothesis translation).""" ref_lines = tokenizer.native_to_unicode( tf.io.gfile.GFile(ref_filename).read()).strip().splitlines() hyp_lines = tokenizer.native_to_unicode( tf.io.gfile.GFile(hyp_filename).read()).strip().splitlines() return bleu_on_list(ref_lines, hyp_lines, case_sensitive) def bleu_on_list(ref_lines, hyp_lines, case_sensitive=False): """Compute BLEU for two list of strings (reference and hypothesis).""" if len(ref_lines) != len(hyp_lines): raise ValueError( "Reference and translation files have different number of " "lines (%d VS %d). If training only a few steps (100-200), the " "translation may be empty." % (len(ref_lines), len(hyp_lines))) if not case_sensitive: ref_lines = [x.lower() for x in ref_lines] hyp_lines = [x.lower() for x in hyp_lines] ref_tokens = [bleu_tokenize(x) for x in ref_lines] hyp_tokens = [bleu_tokenize(x) for x in hyp_lines] return metrics.compute_bleu(ref_tokens, hyp_tokens) * 100 def main(unused_argv): if FLAGS.bleu_variant in ("both", "uncased"): score = bleu_wrapper(FLAGS.reference, FLAGS.translation, False) logging.info("Case-insensitive results: %f", score) if FLAGS.bleu_variant in ("both", "cased"): score = bleu_wrapper(FLAGS.reference, FLAGS.translation, True) logging.info("Case-sensitive results: %f", score) def define_compute_bleu_flags(): """Add flags for computing BLEU score.""" flags.DEFINE_string( name="translation", default=None, help=flags_core.help_wrap("File containing translated text.")) flags.mark_flag_as_required("translation") flags.DEFINE_string( name="reference", default=None, help=flags_core.help_wrap("File containing reference translation.")) flags.mark_flag_as_required("reference") flags.DEFINE_enum( name="bleu_variant", short_name="bv", default="both", enum_values=["both", "uncased", "cased"], case_sensitive=False, help=flags_core.help_wrap( "Specify one or more BLEU variants to calculate. Variants: \"cased\"" ", \"uncased\", or \"both\".")) if __name__ == "__main__": define_compute_bleu_flags() FLAGS = flags.FLAGS app.run(main)
5,141
33.510067
88
py
models
models-master/official/legacy/transformer/embedding_layer.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Implementation of embedding layer with shared weights.""" import tensorflow as tf class EmbeddingSharedWeights(tf.keras.layers.Layer): """Calculates input embeddings and pre-softmax linear with shared weights.""" def __init__(self, vocab_size, hidden_size): """Specify characteristic parameters of embedding layer. Args: vocab_size: Number of tokens in the embedding. (Typically ~32,000) hidden_size: Dimensionality of the embedding. (Typically 512 or 1024) """ super(EmbeddingSharedWeights, self).__init__() self.vocab_size = vocab_size self.hidden_size = hidden_size def build(self, input_shape): """Build embedding layer.""" with tf.name_scope("embedding_and_softmax"): # Create and initialize weights. The random normal initializer was chosen # arbitrarily, and works well. self.shared_weights = self.add_weight( "weights", shape=[self.vocab_size, self.hidden_size], dtype=tf.float32, initializer=tf.random_normal_initializer( mean=0., stddev=self.hidden_size**-0.5)) super(EmbeddingSharedWeights, self).build(input_shape) def get_config(self): return { "vocab_size": self.vocab_size, "hidden_size": self.hidden_size, } def call(self, inputs, mode="embedding"): """Get token embeddings of inputs. Args: inputs: An int64 tensor with shape [batch_size, length] mode: string, a valid value is one of "embedding" and "linear". Returns: outputs: (1) If mode == "embedding", output embedding tensor, float32 with shape [batch_size, length, embedding_size]; (2) mode == "linear", output linear tensor, float32 with shape [batch_size, length, vocab_size]. Raises: ValueError: if mode is not valid. """ if mode == "embedding": return self._embedding(inputs) elif mode == "linear": return self._linear(inputs) else: raise ValueError("mode {} is not valid.".format(mode)) def _embedding(self, inputs): """Applies embedding based on inputs tensor.""" with tf.name_scope("embedding"): # Create binary mask of size [batch_size, length] embeddings = tf.gather(self.shared_weights, inputs) # mask = tf.cast(tf.not_equal(inputs, 0), embeddings.dtype) # embeddings *= tf.expand_dims(mask, -1) # Scale embedding by the sqrt of the hidden size embeddings *= self.hidden_size**0.5 return embeddings def _linear(self, inputs): """Computes logits by running inputs through a linear layer. Args: inputs: A float32 tensor with shape [batch_size, length, hidden_size] Returns: float32 tensor with shape [batch_size, length, vocab_size]. """ with tf.name_scope("presoftmax_linear"): batch_size = tf.shape(inputs)[0] length = tf.shape(inputs)[1] x = tf.reshape(inputs, [-1, self.hidden_size]) logits = tf.matmul(x, self.shared_weights, transpose_b=True) return tf.reshape(logits, [batch_size, length, self.vocab_size])
3,684
34.776699
80
py
models
models-master/official/legacy/transformer/data_download.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Download and preprocess WMT17 ende training and evaluation datasets.""" import os import random import tarfile # pylint: disable=g-bad-import-order from absl import app from absl import flags from absl import logging import six from six.moves import range from six.moves import urllib from six.moves import zip import tensorflow.compat.v1 as tf from official.legacy.transformer.utils import tokenizer from official.utils.flags import core as flags_core # pylint: enable=g-bad-import-order # Data sources for training/evaluating the transformer translation model. # If any of the training sources are changed, then either: # 1) use the flag `--search` to find the best min count or # 2) update the _TRAIN_DATA_MIN_COUNT constant. # min_count is the minimum number of times a token must appear in the data # before it is added to the vocabulary. "Best min count" refers to the value # that generates a vocabulary set that is closest in size to _TARGET_VOCAB_SIZE. _TRAIN_DATA_SOURCES = [ { "url": "http://data.statmt.org/wmt17/translation-task/" "training-parallel-nc-v12.tgz", "input": "news-commentary-v12.de-en.en", "target": "news-commentary-v12.de-en.de", }, { "url": "http://www.statmt.org/wmt13/training-parallel-commoncrawl.tgz", "input": "commoncrawl.de-en.en", "target": "commoncrawl.de-en.de", }, { "url": "http://www.statmt.org/wmt13/training-parallel-europarl-v7.tgz", "input": "europarl-v7.de-en.en", "target": "europarl-v7.de-en.de", }, ] # Use pre-defined minimum count to generate subtoken vocabulary. _TRAIN_DATA_MIN_COUNT = 6 _EVAL_DATA_SOURCES = [{ "url": "http://data.statmt.org/wmt17/translation-task/dev.tgz", "input": "newstest2013.en", "target": "newstest2013.de", }] _TEST_DATA_SOURCES = [{ "url": ("https://storage.googleapis.com/cloud-tpu-test-datasets/" "transformer_data/newstest2014.tgz"), "input": "newstest2014.en", "target": "newstest2014.de", }] # Vocabulary constants _TARGET_VOCAB_SIZE = 32768 # Number of subtokens in the vocabulary list. _TARGET_THRESHOLD = 327 # Accept vocabulary if size is within this threshold VOCAB_FILE = "vocab.ende.%d" % _TARGET_VOCAB_SIZE # Strings to inclue in the generated files. _PREFIX = "wmt32k" _TRAIN_TAG = "train" _EVAL_TAG = "dev" # Following WMT and Tensor2Tensor conventions, in which the # evaluation datasets are tagged as "dev" for development. # Number of files to split train and evaluation data _TRAIN_SHARDS = 100 _EVAL_SHARDS = 1 def find_file(path, filename, max_depth=5): """Returns full filepath if the file is in path or a subdirectory.""" for root, dirs, files in os.walk(path): if filename in files: return os.path.join(root, filename) # Don't search past max_depth depth = root[len(path) + 1:].count(os.sep) if depth > max_depth: del dirs[:] # Clear dirs return None ############################################################################### # Download and extraction functions ############################################################################### def get_raw_files(raw_dir, data_source): """Return raw files from source. Downloads/extracts if needed. Args: raw_dir: string directory to store raw files data_source: dictionary with {"url": url of compressed dataset containing input and target files "input": file with data in input language "target": file with data in target language} Returns: dictionary with {"inputs": list of files containing data in input language "targets": list of files containing corresponding data in target language } """ raw_files = { "inputs": [], "targets": [], } # keys for d in data_source: input_file, target_file = download_and_extract(raw_dir, d["url"], d["input"], d["target"]) raw_files["inputs"].append(input_file) raw_files["targets"].append(target_file) return raw_files def download_report_hook(count, block_size, total_size): """Report hook for download progress. Args: count: current block number block_size: block size total_size: total size """ percent = int(count * block_size * 100 / total_size) print(six.ensure_str("\r%d%%" % percent) + " completed", end="\r") def download_from_url(path, url): """Download content from a url. Args: path: string directory where file will be downloaded url: string url Returns: Full path to downloaded file """ filename = six.ensure_str(url).split("/")[-1] found_file = find_file(path, filename, max_depth=0) if found_file is None: filename = os.path.join(path, filename) logging.info("Downloading from %s to %s.", url, filename) inprogress_filepath = six.ensure_str(filename) + ".incomplete" inprogress_filepath, _ = urllib.request.urlretrieve( url, inprogress_filepath, reporthook=download_report_hook) # Print newline to clear the carriage return from the download progress. print() tf.gfile.Rename(inprogress_filepath, filename) return filename else: logging.info("Already downloaded: %s (at %s).", url, found_file) return found_file def download_and_extract(path, url, input_filename, target_filename): """Extract files from downloaded compressed archive file. Args: path: string directory where the files will be downloaded url: url containing the compressed input and target files input_filename: name of file containing data in source language target_filename: name of file containing data in target language Returns: Full paths to extracted input and target files. Raises: OSError: if the download/extraction fails. """ # Check if extracted files already exist in path input_file = find_file(path, input_filename) target_file = find_file(path, target_filename) if input_file and target_file: logging.info("Already downloaded and extracted %s.", url) return input_file, target_file # Download archive file if it doesn't already exist. compressed_file = download_from_url(path, url) # Extract compressed files logging.info("Extracting %s.", compressed_file) with tarfile.open(compressed_file, "r:gz") as corpus_tar: corpus_tar.extractall(path) # Return file paths of the requested files. input_file = find_file(path, input_filename) target_file = find_file(path, target_filename) if input_file and target_file: return input_file, target_file raise OSError("Download/extraction failed for url %s to path %s" % (url, path)) def txt_line_iterator(path): """Iterate through lines of file.""" with tf.io.gfile.GFile(path) as f: for line in f: yield line.strip() def compile_files(raw_dir, raw_files, tag): """Compile raw files into a single file for each language. Args: raw_dir: Directory containing downloaded raw files. raw_files: Dict containing filenames of input and target data. {"inputs": list of files containing data in input language "targets": list of files containing corresponding data in target language } tag: String to append to the compiled filename. Returns: Full path of compiled input and target files. """ logging.info("Compiling files with tag %s.", tag) filename = "%s-%s" % (_PREFIX, tag) input_compiled_file = os.path.join(raw_dir, six.ensure_str(filename) + ".lang1") target_compiled_file = os.path.join(raw_dir, six.ensure_str(filename) + ".lang2") with tf.io.gfile.GFile(input_compiled_file, mode="w") as input_writer: with tf.io.gfile.GFile(target_compiled_file, mode="w") as target_writer: for i in range(len(raw_files["inputs"])): input_file = raw_files["inputs"][i] target_file = raw_files["targets"][i] logging.info("Reading files %s and %s.", input_file, target_file) write_file(input_writer, input_file) write_file(target_writer, target_file) return input_compiled_file, target_compiled_file def write_file(writer, filename): """Write all of lines from file using the writer.""" for line in txt_line_iterator(filename): writer.write(line) writer.write("\n") ############################################################################### # Data preprocessing ############################################################################### def encode_and_save_files(subtokenizer, data_dir, raw_files, tag, total_shards): """Save data from files as encoded Examples in TFrecord format. Args: subtokenizer: Subtokenizer object that will be used to encode the strings. data_dir: The directory in which to write the examples raw_files: A tuple of (input, target) data files. Each line in the input and the corresponding line in target file will be saved in a tf.Example. tag: String that will be added onto the file names. total_shards: Number of files to divide the data into. Returns: List of all files produced. """ # Create a file for each shard. filepaths = [ shard_filename(data_dir, tag, n + 1, total_shards) for n in range(total_shards) ] if all_exist(filepaths): logging.info("Files with tag %s already exist.", tag) return filepaths logging.info("Saving files with tag %s.", tag) input_file = raw_files[0] target_file = raw_files[1] # Write examples to each shard in round robin order. tmp_filepaths = [six.ensure_str(fname) + ".incomplete" for fname in filepaths] writers = [tf.python_io.TFRecordWriter(fname) for fname in tmp_filepaths] counter, shard = 0, 0 for counter, (input_line, target_line) in enumerate( zip(txt_line_iterator(input_file), txt_line_iterator(target_file))): if counter > 0 and counter % 100000 == 0: logging.info("\tSaving case %d.", counter) example = dict_to_example({ "inputs": subtokenizer.encode(input_line, add_eos=True), "targets": subtokenizer.encode(target_line, add_eos=True) }) writers[shard].write(example.SerializeToString()) shard = (shard + 1) % total_shards for writer in writers: writer.close() for tmp_name, final_name in zip(tmp_filepaths, filepaths): tf.gfile.Rename(tmp_name, final_name) logging.info("Saved %d Examples", counter + 1) return filepaths def shard_filename(path, tag, shard_num, total_shards): """Create filename for data shard.""" return os.path.join( path, "%s-%s-%.5d-of-%.5d" % (_PREFIX, tag, shard_num, total_shards)) def shuffle_records(fname): """Shuffle records in a single file.""" logging.info("Shuffling records in file %s", fname) # Rename file prior to shuffling tmp_fname = six.ensure_str(fname) + ".unshuffled" tf.gfile.Rename(fname, tmp_fname) reader = tf.io.tf_record_iterator(tmp_fname) records = [] for record in reader: records.append(record) if len(records) % 100000 == 0: logging.info("\tRead: %d", len(records)) random.shuffle(records) # Write shuffled records to original file name with tf.python_io.TFRecordWriter(fname) as w: for count, record in enumerate(records): w.write(record) if count > 0 and count % 100000 == 0: logging.info("\tWriting record: %d", count) tf.gfile.Remove(tmp_fname) def dict_to_example(dictionary): """Converts a dictionary of string->int to a tf.Example.""" features = {} for k, v in six.iteritems(dictionary): features[k] = tf.train.Feature(int64_list=tf.train.Int64List(value=v)) return tf.train.Example(features=tf.train.Features(feature=features)) def all_exist(filepaths): """Returns true if all files in the list exist.""" for fname in filepaths: if not tf.gfile.Exists(fname): return False return True def make_dir(path): if not tf.gfile.Exists(path): logging.info("Creating directory %s", path) tf.gfile.MakeDirs(path) def main(unused_argv): """Obtain training and evaluation data for the Transformer model.""" make_dir(FLAGS.raw_dir) make_dir(FLAGS.data_dir) # Download test_data logging.info("Step 1/5: Downloading test data") get_raw_files(FLAGS.data_dir, _TEST_DATA_SOURCES) # Get paths of download/extracted training and evaluation files. logging.info("Step 2/5: Downloading data from source") train_files = get_raw_files(FLAGS.raw_dir, _TRAIN_DATA_SOURCES) eval_files = get_raw_files(FLAGS.raw_dir, _EVAL_DATA_SOURCES) # Create subtokenizer based on the training files. logging.info("Step 3/5: Creating subtokenizer and building vocabulary") train_files_flat = train_files["inputs"] + train_files["targets"] vocab_file = os.path.join(FLAGS.data_dir, VOCAB_FILE) subtokenizer = tokenizer.Subtokenizer.init_from_files( vocab_file, train_files_flat, _TARGET_VOCAB_SIZE, _TARGET_THRESHOLD, min_count=None if FLAGS.search else _TRAIN_DATA_MIN_COUNT) logging.info("Step 4/5: Compiling training and evaluation data") compiled_train_files = compile_files(FLAGS.raw_dir, train_files, _TRAIN_TAG) compiled_eval_files = compile_files(FLAGS.raw_dir, eval_files, _EVAL_TAG) # Tokenize and save data as Examples in the TFRecord format. logging.info("Step 5/5: Preprocessing and saving data") train_tfrecord_files = encode_and_save_files(subtokenizer, FLAGS.data_dir, compiled_train_files, _TRAIN_TAG, _TRAIN_SHARDS) encode_and_save_files(subtokenizer, FLAGS.data_dir, compiled_eval_files, _EVAL_TAG, _EVAL_SHARDS) for fname in train_tfrecord_files: shuffle_records(fname) def define_data_download_flags(): """Add flags specifying data download arguments.""" flags.DEFINE_string( name="data_dir", short_name="dd", default="/tmp/translate_ende", help=flags_core.help_wrap( "Directory for where the translate_ende_wmt32k dataset is saved.")) flags.DEFINE_string( name="raw_dir", short_name="rd", default="/tmp/translate_ende_raw", help=flags_core.help_wrap( "Path where the raw data will be downloaded and extracted.")) flags.DEFINE_bool( name="search", default=False, help=flags_core.help_wrap( "If set, use binary search to find the vocabulary set with size" "closest to the target size (%d)." % _TARGET_VOCAB_SIZE)) if __name__ == "__main__": logging.set_verbosity(logging.INFO) define_data_download_flags() FLAGS = flags.FLAGS app.run(main)
15,291
33.441441
80
py
models
models-master/official/legacy/transformer/ffn_layer.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Implementation of fully connected network.""" import tensorflow as tf class FeedForwardNetwork(tf.keras.layers.Layer): """Fully connected feedforward network.""" def __init__(self, hidden_size, filter_size, relu_dropout): """Initialize FeedForwardNetwork. Args: hidden_size: int, output dim of hidden layer. filter_size: int, filter size for the inner (first) dense layer. relu_dropout: float, dropout rate for training. """ super(FeedForwardNetwork, self).__init__() self.hidden_size = hidden_size self.filter_size = filter_size self.relu_dropout = relu_dropout def build(self, input_shape): self.filter_dense_layer = tf.keras.layers.Dense( self.filter_size, use_bias=True, activation=tf.nn.relu, name="filter_layer") self.output_dense_layer = tf.keras.layers.Dense( self.hidden_size, use_bias=True, name="output_layer") super(FeedForwardNetwork, self).build(input_shape) def get_config(self): return { "hidden_size": self.hidden_size, "filter_size": self.filter_size, "relu_dropout": self.relu_dropout, } def call(self, x, training): """Return outputs of the feedforward network. Args: x: tensor with shape [batch_size, length, hidden_size] training: boolean, whether in training mode or not. Returns: Output of the feedforward network. tensor with shape [batch_size, length, hidden_size] """ # Retrieve dynamically known shapes output = self.filter_dense_layer(x) if training: output = tf.nn.dropout(output, rate=self.relu_dropout) output = self.output_dense_layer(output) return output
2,320
31.236111
74
py
models
models-master/official/legacy/transformer/compute_bleu_test.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Test functions in compute_blue.py.""" import tempfile import tensorflow as tf from official.legacy.transformer import compute_bleu class ComputeBleuTest(tf.test.TestCase): def _create_temp_file(self, text): temp_file = tempfile.NamedTemporaryFile(delete=False) with tf.io.gfile.GFile(temp_file.name, "w") as w: w.write(text) return temp_file.name def test_bleu_same(self): ref = self._create_temp_file("test 1 two 3\nmore tests!") hyp = self._create_temp_file("test 1 two 3\nmore tests!") uncased_score = compute_bleu.bleu_wrapper(ref, hyp, False) cased_score = compute_bleu.bleu_wrapper(ref, hyp, True) self.assertEqual(100, uncased_score) self.assertEqual(100, cased_score) def test_bleu_same_different_case(self): ref = self._create_temp_file("Test 1 two 3\nmore tests!") hyp = self._create_temp_file("test 1 two 3\nMore tests!") uncased_score = compute_bleu.bleu_wrapper(ref, hyp, False) cased_score = compute_bleu.bleu_wrapper(ref, hyp, True) self.assertEqual(100, uncased_score) self.assertLess(cased_score, 100) def test_bleu_different(self): ref = self._create_temp_file("Testing\nmore tests!") hyp = self._create_temp_file("Dog\nCat") uncased_score = compute_bleu.bleu_wrapper(ref, hyp, False) cased_score = compute_bleu.bleu_wrapper(ref, hyp, True) self.assertLess(uncased_score, 100) self.assertLess(cased_score, 100) def test_bleu_tokenize(self): s = "Test0, 1 two, 3" tokenized = compute_bleu.bleu_tokenize(s) self.assertEqual(["Test0", ",", "1", "two", ",", "3"], tokenized) def test_bleu_list(self): ref = ["test 1 two 3", "more tests!"] hyp = ["test 1 two 3", "More tests!"] uncased_score = compute_bleu.bleu_on_list(ref, hyp, False) cased_score = compute_bleu.bleu_on_list(ref, hyp, True) self.assertEqual(uncased_score, 100) self.assertLess(cased_score, 100) if __name__ == "__main__": tf.test.main()
2,585
34.424658
74
py
models
models-master/official/legacy/transformer/transformer_layers_test.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for layers in Transformer.""" import tensorflow as tf from official.legacy.transformer import attention_layer from official.legacy.transformer import embedding_layer from official.legacy.transformer import ffn_layer from official.legacy.transformer import metrics class TransformerLayersTest(tf.test.TestCase): def test_attention_layer(self): hidden_size = 64 num_heads = 4 dropout = 0.5 dim_per_head = hidden_size // num_heads layer = attention_layer.SelfAttention(hidden_size, num_heads, dropout) self.assertDictEqual( layer.get_config(), { "hidden_size": hidden_size, "num_heads": num_heads, "attention_dropout": dropout, }) length = 2 x = tf.ones([1, length, hidden_size]) bias = tf.ones([1]) cache = { "k": tf.zeros([1, 0, num_heads, dim_per_head]), "v": tf.zeros([1, 0, num_heads, dim_per_head]), } y = layer(x, bias, training=True, cache=cache) self.assertEqual(y.shape, ( 1, length, 64, )) self.assertEqual(cache["k"].shape, ( 1, length, num_heads, dim_per_head, )) self.assertEqual(cache["v"].shape, ( 1, length, num_heads, dim_per_head, )) def test_embedding_shared_weights(self): vocab_size = 50 hidden_size = 64 length = 2 layer = embedding_layer.EmbeddingSharedWeights(vocab_size, hidden_size) self.assertDictEqual(layer.get_config(), { "vocab_size": 50, "hidden_size": 64, }) idx = tf.ones([1, length], dtype="int32") y = layer(idx) self.assertEqual(y.shape, ( 1, length, hidden_size, )) x = tf.ones([1, length, hidden_size]) output = layer(x, "linear") self.assertEqual(output.shape, ( 1, length, vocab_size, )) def test_feed_forward_network(self): hidden_size = 64 filter_size = 32 relu_dropout = 0.5 layer = ffn_layer.FeedForwardNetwork(hidden_size, filter_size, relu_dropout) self.assertDictEqual( layer.get_config(), { "hidden_size": hidden_size, "filter_size": filter_size, "relu_dropout": relu_dropout, }) length = 2 x = tf.ones([1, length, hidden_size]) y = layer(x, training=True) self.assertEqual(y.shape, ( 1, length, hidden_size, )) def test_metric_layer(self): vocab_size = 50 logits = tf.keras.layers.Input((None, vocab_size), dtype="float32", name="logits") targets = tf.keras.layers.Input((None,), dtype="int64", name="targets") output_logits = metrics.MetricLayer(vocab_size)([logits, targets]) self.assertEqual(output_logits.shape.as_list(), [ None, None, vocab_size, ]) if __name__ == "__main__": tf.test.main()
3,566
27.309524
80
py
models
models-master/official/legacy/transformer/model_params.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Defines Transformer model parameters.""" import collections BASE_PARAMS = collections.defaultdict( lambda: None, # Set default value to None. # Input params default_batch_size=2048, # Maximum number of tokens per batch of examples. default_batch_size_tpu=32768, max_length=256, # Maximum number of tokens per example. # Model params initializer_gain=1.0, # Used in trainable variable initialization. vocab_size=33708, # Number of tokens defined in the vocabulary file. hidden_size=512, # Model dimension in the hidden layers. num_hidden_layers=6, # Number of layers in the encoder and decoder stacks. num_heads=8, # Number of heads to use in multi-headed attention. filter_size=2048, # Inner layer dimension in the feedforward network. # Dropout values (only used when training) layer_postprocess_dropout=0.1, attention_dropout=0.1, relu_dropout=0.1, # Training params label_smoothing=0.1, learning_rate=2.0, learning_rate_decay_rate=1.0, learning_rate_warmup_steps=16000, # Optimizer params optimizer_adam_beta1=0.9, optimizer_adam_beta2=0.997, optimizer_adam_epsilon=1e-09, # Default prediction params extra_decode_length=50, beam_size=4, alpha=0.6, # used to calculate length normalization in beam search # TPU specific parameters use_tpu=False, static_batch=False, allow_ffn_pad=True, ) BIG_PARAMS = BASE_PARAMS.copy() BIG_PARAMS.update( default_batch_size=4096, # default batch size is smaller than for BASE_PARAMS due to memory limits. default_batch_size_tpu=16384, hidden_size=1024, filter_size=4096, num_heads=16, ) # Parameters for running the model in multi gpu. These should not change the # params that modify the model shape (such as the hidden_size or num_heads). BASE_MULTI_GPU_PARAMS = BASE_PARAMS.copy() BASE_MULTI_GPU_PARAMS.update( learning_rate_warmup_steps=8000 ) BIG_MULTI_GPU_PARAMS = BIG_PARAMS.copy() BIG_MULTI_GPU_PARAMS.update( layer_postprocess_dropout=0.3, learning_rate_warmup_steps=8000 ) # Parameters for testing the model TINY_PARAMS = BASE_PARAMS.copy() TINY_PARAMS.update( default_batch_size=1024, default_batch_size_tpu=1024, hidden_size=32, num_heads=4, filter_size=256, )
2,936
29.278351
79
py
models
models-master/official/legacy/transformer/transformer.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Defines the Transformer model in TF 2.0. Model paper: https://arxiv.org/pdf/1706.03762.pdf Transformer model code source: https://github.com/tensorflow/tensor2tensor """ import tensorflow as tf from official.legacy.transformer import attention_layer from official.legacy.transformer import embedding_layer from official.legacy.transformer import ffn_layer from official.legacy.transformer import metrics from official.legacy.transformer import model_utils from official.legacy.transformer.utils.tokenizer import EOS_ID from official.nlp.modeling.layers import position_embedding from official.nlp.modeling.ops import beam_search # Disable the not-callable lint error, since it claims many objects are not # callable when they actually are. # pylint: disable=not-callable def create_model(params, is_train): """Creates transformer model.""" with tf.name_scope("model"): if is_train: inputs = tf.keras.layers.Input((None,), dtype="int64", name="inputs") targets = tf.keras.layers.Input((None,), dtype="int64", name="targets") internal_model = Transformer(params, name="transformer_v2") logits = internal_model([inputs, targets], training=is_train) vocab_size = params["vocab_size"] label_smoothing = params["label_smoothing"] if params["enable_metrics_in_training"]: logits = metrics.MetricLayer(vocab_size)([logits, targets]) logits = tf.keras.layers.Lambda( lambda x: x, name="logits", dtype=tf.float32)( logits) model = tf.keras.Model([inputs, targets], logits) loss = metrics.transformer_loss(logits, targets, label_smoothing, vocab_size) model.add_loss(loss) return model else: inputs = tf.keras.layers.Input((None,), dtype="int64", name="inputs") internal_model = Transformer(params, name="transformer_v2") ret = internal_model([inputs], training=is_train) outputs, scores = ret["outputs"], ret["scores"] return tf.keras.Model(inputs, [outputs, scores]) class Transformer(tf.keras.Model): """Transformer model with Keras. Implemented as described in: https://arxiv.org/pdf/1706.03762.pdf The Transformer model consists of an encoder and decoder. The input is an int sequence (or a batch of sequences). The encoder produces a continuous representation, and the decoder uses the encoder output to generate probabilities for the output sequence. """ def __init__(self, params, name=None): """Initialize layers to build Transformer model. Args: params: hyperparameter object defining layer sizes, dropout values, etc. name: name of the model. """ super(Transformer, self).__init__(name=name) self.params = params self.embedding_softmax_layer = embedding_layer.EmbeddingSharedWeights( params["vocab_size"], params["hidden_size"]) self.encoder_stack = EncoderStack(params) self.decoder_stack = DecoderStack(params) self.position_embedding = position_embedding.RelativePositionEmbedding( hidden_size=self.params["hidden_size"]) def get_config(self): return { "params": self.params, } def call(self, inputs, training): """Calculate target logits or inferred target sequences. Args: inputs: input tensor list of size 1 or 2. First item, inputs: int tensor with shape [batch_size, input_length]. Second item (optional), targets: None or int tensor with shape [batch_size, target_length]. training: boolean, whether in training mode or not. Returns: If targets is defined, then return logits for each word in the target sequence. float tensor with shape [batch_size, target_length, vocab_size] If target is none, then generate output sequence one token at a time. returns a dictionary { outputs: int tensor with shape [batch_size, decoded_length] scores: float tensor with shape [batch_size]} Even when float16 is used, the output tensor(s) are always float32. Raises: NotImplementedError: If try to use padded decode method on CPU/GPUs. """ inputs = inputs if isinstance(inputs, list) else [inputs] if len(inputs) == 2: inputs, targets = inputs[0], inputs[1] else: # Decoding path. inputs, targets = inputs[0], None if self.params["padded_decode"]: if not self.params["num_replicas"]: raise NotImplementedError( "Padded decoding on CPU/GPUs is not supported.") decode_batch_size = int(self.params["decode_batch_size"] / self.params["num_replicas"]) inputs.set_shape([decode_batch_size, self.params["decode_max_length"]]) # Variance scaling is used here because it seems to work in many problems. # Other reasonable initializers may also work just as well. with tf.name_scope("Transformer"): # Calculate attention bias for encoder self-attention and decoder # multi-headed attention layers. attention_bias = model_utils.get_padding_bias(inputs) # Run the inputs through the encoder layer to map the symbol # representations to continuous representations. encoder_outputs = self.encode(inputs, attention_bias, training) # Generate output sequence if targets is None, or return logits if target # sequence is known. if targets is None: return self.predict(encoder_outputs, attention_bias, training) else: logits = self.decode(targets, encoder_outputs, attention_bias, training) return logits def encode(self, inputs, attention_bias, training): """Generate continuous representation for inputs. Args: inputs: int tensor with shape [batch_size, input_length]. attention_bias: float tensor with shape [batch_size, 1, 1, input_length]. training: boolean, whether in training mode or not. Returns: float tensor with shape [batch_size, input_length, hidden_size] """ with tf.name_scope("encode"): # Prepare inputs to the layer stack by adding positional encodings and # applying dropout. embedded_inputs = self.embedding_softmax_layer(inputs) embedded_inputs = tf.cast(embedded_inputs, self.params["dtype"]) inputs_padding = model_utils.get_padding(inputs) attention_bias = tf.cast(attention_bias, self.params["dtype"]) with tf.name_scope("add_pos_encoding"): pos_encoding = self.position_embedding(inputs=embedded_inputs) pos_encoding = tf.cast(pos_encoding, self.params["dtype"]) encoder_inputs = embedded_inputs + pos_encoding if training: encoder_inputs = tf.nn.dropout( encoder_inputs, rate=self.params["layer_postprocess_dropout"]) return self.encoder_stack( encoder_inputs, attention_bias, inputs_padding, training=training) def decode(self, targets, encoder_outputs, attention_bias, training): """Generate logits for each value in the target sequence. Args: targets: target values for the output sequence. int tensor with shape [batch_size, target_length] encoder_outputs: continuous representation of input sequence. float tensor with shape [batch_size, input_length, hidden_size] attention_bias: float tensor with shape [batch_size, 1, 1, input_length] training: boolean, whether in training mode or not. Returns: float32 tensor with shape [batch_size, target_length, vocab_size] """ with tf.name_scope("decode"): # Prepare inputs to decoder layers by shifting targets, adding positional # encoding and applying dropout. with tf.name_scope("shift_targets"): # Shift targets to the right, and remove the last element targets = tf.pad(targets, [[0, 0], [1, 0]])[:, :-1] decoder_inputs = self.embedding_softmax_layer(targets) decoder_inputs = tf.cast(decoder_inputs, self.params["dtype"]) attention_bias = tf.cast(attention_bias, self.params["dtype"]) with tf.name_scope("add_pos_encoding"): length = tf.shape(decoder_inputs)[1] pos_encoding = self.position_embedding(decoder_inputs) pos_encoding = tf.cast(pos_encoding, self.params["dtype"]) decoder_inputs += pos_encoding if training: decoder_inputs = tf.nn.dropout( decoder_inputs, rate=self.params["layer_postprocess_dropout"]) # Run values decoder_self_attention_bias = model_utils.get_decoder_self_attention_bias( length, dtype=self.params["dtype"]) outputs = self.decoder_stack( decoder_inputs, encoder_outputs, decoder_self_attention_bias, attention_bias, training=training) logits = self.embedding_softmax_layer(outputs, mode="linear") logits = tf.cast(logits, tf.float32) return logits def _get_symbols_to_logits_fn(self, max_decode_length, training): """Returns a decoding function that calculates logits of the next tokens.""" timing_signal = self.position_embedding( inputs=None, length=max_decode_length + 1) timing_signal = tf.cast(timing_signal, self.params["dtype"]) decoder_self_attention_bias = model_utils.get_decoder_self_attention_bias( max_decode_length, dtype=self.params["dtype"]) def symbols_to_logits_fn(ids, i, cache): """Generate logits for next potential IDs. Args: ids: Current decoded sequences. int tensor with shape [batch_size * beam_size, i + 1]. i: Loop index. cache: dictionary of values storing the encoder output, encoder-decoder attention bias, and previous decoder attention values. Returns: Tuple of (logits with shape [batch_size * beam_size, vocab_size], updated cache values) """ # Set decoder input to the last generated IDs decoder_input = ids[:, -1:] # Preprocess decoder input by getting embeddings and adding timing signal. decoder_input = self.embedding_softmax_layer(decoder_input) decoder_input += timing_signal[i] if self.params["padded_decode"]: bias_shape = decoder_self_attention_bias.shape.as_list() self_attention_bias = tf.slice( decoder_self_attention_bias, [0, 0, i, 0], [bias_shape[0], bias_shape[1], 1, bias_shape[3]]) else: self_attention_bias = decoder_self_attention_bias[:, :, i:i + 1, :i + 1] decoder_outputs = self.decoder_stack( decoder_input, cache.get("encoder_outputs"), self_attention_bias, cache.get("encoder_decoder_attention_bias"), training=training, cache=cache, decode_loop_step=i if self.params["padded_decode"] else None) logits = self.embedding_softmax_layer(decoder_outputs, mode="linear") logits = tf.squeeze(logits, axis=[1]) return logits, cache return symbols_to_logits_fn def predict(self, encoder_outputs, encoder_decoder_attention_bias, training): """Return predicted sequence.""" encoder_outputs = tf.cast(encoder_outputs, self.params["dtype"]) if self.params["padded_decode"]: batch_size = encoder_outputs.shape.as_list()[0] input_length = encoder_outputs.shape.as_list()[1] else: batch_size = tf.shape(encoder_outputs)[0] input_length = tf.shape(encoder_outputs)[1] max_decode_length = input_length + self.params["extra_decode_length"] encoder_decoder_attention_bias = tf.cast(encoder_decoder_attention_bias, self.params["dtype"]) symbols_to_logits_fn = self._get_symbols_to_logits_fn( max_decode_length, training) # Create initial set of IDs that will be passed into symbols_to_logits_fn. initial_ids = tf.zeros([batch_size], dtype=tf.int32) # Create cache storing decoder attention values for each layer. # pylint: disable=g-complex-comprehension init_decode_length = ( max_decode_length if self.params["padded_decode"] else 0) num_heads = self.params["num_heads"] dim_per_head = self.params["hidden_size"] // num_heads cache = { "layer_%d" % layer: { "k": tf.zeros( [batch_size, init_decode_length, num_heads, dim_per_head], dtype=self.params["dtype"]), "v": tf.zeros( [batch_size, init_decode_length, num_heads, dim_per_head], dtype=self.params["dtype"]) } for layer in range(self.params["num_hidden_layers"]) } # pylint: enable=g-complex-comprehension # Add encoder output and attention bias to the cache. cache["encoder_outputs"] = encoder_outputs cache["encoder_decoder_attention_bias"] = encoder_decoder_attention_bias # Use beam search to find the top beam_size sequences and scores. decoded_ids, scores = beam_search.sequence_beam_search( symbols_to_logits_fn=symbols_to_logits_fn, initial_ids=initial_ids, initial_cache=cache, vocab_size=self.params["vocab_size"], beam_size=self.params["beam_size"], alpha=self.params["alpha"], max_decode_length=max_decode_length, eos_id=EOS_ID, padded_decode=self.params["padded_decode"], dtype=self.params["dtype"]) # Get the top sequence for each batch element top_decoded_ids = decoded_ids[:, 0, 1:] top_scores = scores[:, 0] return {"outputs": top_decoded_ids, "scores": top_scores} class PrePostProcessingWrapper(tf.keras.layers.Layer): """Wrapper class that applies layer pre-processing and post-processing.""" def __init__(self, layer, params): super(PrePostProcessingWrapper, self).__init__() self.layer = layer self.params = params self.postprocess_dropout = params["layer_postprocess_dropout"] def build(self, input_shape): # Create normalization layer self.layer_norm = tf.keras.layers.LayerNormalization( epsilon=1e-6, dtype="float32") super(PrePostProcessingWrapper, self).build(input_shape) def get_config(self): return { "params": self.params, } def call(self, x, *args, **kwargs): """Calls wrapped layer with same parameters.""" # Preprocessing: apply layer normalization training = kwargs["training"] y = self.layer_norm(x) # Get layer output y = self.layer(y, *args, **kwargs) # Postprocessing: apply dropout and residual connection if training: y = tf.nn.dropout(y, rate=self.postprocess_dropout) return x + y class EncoderStack(tf.keras.layers.Layer): """Transformer encoder stack. The encoder stack is made up of N identical layers. Each layer is composed of the sublayers: 1. Self-attention layer 2. Feedforward network (which is 2 fully-connected layers) """ def __init__(self, params): super(EncoderStack, self).__init__() self.params = params self.layers = [] def build(self, input_shape): """Builds the encoder stack.""" params = self.params for _ in range(params["num_hidden_layers"]): # Create sublayers for each layer. self_attention_layer = attention_layer.SelfAttention( params["hidden_size"], params["num_heads"], params["attention_dropout"]) feed_forward_network = ffn_layer.FeedForwardNetwork( params["hidden_size"], params["filter_size"], params["relu_dropout"]) self.layers.append([ PrePostProcessingWrapper(self_attention_layer, params), PrePostProcessingWrapper(feed_forward_network, params) ]) # Create final layer normalization layer. self.output_normalization = tf.keras.layers.LayerNormalization( epsilon=1e-6, dtype="float32") super(EncoderStack, self).build(input_shape) def get_config(self): return { "params": self.params, } def call(self, encoder_inputs, attention_bias, inputs_padding, training): """Return the output of the encoder layer stacks. Args: encoder_inputs: tensor with shape [batch_size, input_length, hidden_size] attention_bias: bias for the encoder self-attention layer. [batch_size, 1, 1, input_length] inputs_padding: tensor with shape [batch_size, input_length], inputs with zero paddings. training: boolean, whether in training mode or not. Returns: Output of encoder layer stack. float32 tensor with shape [batch_size, input_length, hidden_size] """ for n, layer in enumerate(self.layers): # Run inputs through the sublayers. self_attention_layer = layer[0] feed_forward_network = layer[1] with tf.name_scope("layer_%d" % n): with tf.name_scope("self_attention"): encoder_inputs = self_attention_layer( encoder_inputs, attention_bias, training=training) with tf.name_scope("ffn"): encoder_inputs = feed_forward_network( encoder_inputs, training=training) return self.output_normalization(encoder_inputs) class DecoderStack(tf.keras.layers.Layer): """Transformer decoder stack. Like the encoder stack, the decoder stack is made up of N identical layers. Each layer is composed of the sublayers: 1. Self-attention layer 2. Multi-headed attention layer combining encoder outputs with results from the previous self-attention layer. 3. Feedforward network (2 fully-connected layers) """ def __init__(self, params): super(DecoderStack, self).__init__() self.params = params self.layers = [] def build(self, input_shape): """Builds the decoder stack.""" params = self.params for _ in range(params["num_hidden_layers"]): self_attention_layer = attention_layer.SelfAttention( params["hidden_size"], params["num_heads"], params["attention_dropout"]) enc_dec_attention_layer = attention_layer.Attention( params["hidden_size"], params["num_heads"], params["attention_dropout"]) feed_forward_network = ffn_layer.FeedForwardNetwork( params["hidden_size"], params["filter_size"], params["relu_dropout"]) self.layers.append([ PrePostProcessingWrapper(self_attention_layer, params), PrePostProcessingWrapper(enc_dec_attention_layer, params), PrePostProcessingWrapper(feed_forward_network, params) ]) self.output_normalization = tf.keras.layers.LayerNormalization( epsilon=1e-6, dtype="float32") super(DecoderStack, self).build(input_shape) def get_config(self): return { "params": self.params, } def call(self, decoder_inputs, encoder_outputs, decoder_self_attention_bias, attention_bias, training, cache=None, decode_loop_step=None): """Return the output of the decoder layer stacks. Args: decoder_inputs: A tensor with shape [batch_size, target_length, hidden_size]. encoder_outputs: A tensor with shape [batch_size, input_length, hidden_size] decoder_self_attention_bias: A tensor with shape [1, 1, target_len, target_length], the bias for decoder self-attention layer. attention_bias: A tensor with shape [batch_size, 1, 1, input_length], the bias for encoder-decoder attention layer. training: A bool, whether in training mode or not. cache: (Used for fast decoding) A nested dictionary storing previous decoder self-attention values. The items are: {layer_n: {"k": A tensor with shape [batch_size, i, key_channels], "v": A tensor with shape [batch_size, i, value_channels]}, ...} decode_loop_step: An integer, the step number of the decoding loop. Used only for autoregressive inference on TPU. Returns: Output of decoder layer stack. float32 tensor with shape [batch_size, target_length, hidden_size] """ for n, layer in enumerate(self.layers): self_attention_layer = layer[0] enc_dec_attention_layer = layer[1] feed_forward_network = layer[2] # Run inputs through the sublayers. layer_name = "layer_%d" % n layer_cache = cache[layer_name] if cache is not None else None with tf.name_scope(layer_name): with tf.name_scope("self_attention"): decoder_inputs = self_attention_layer( decoder_inputs, decoder_self_attention_bias, training=training, cache=layer_cache, decode_loop_step=decode_loop_step) with tf.name_scope("encdec_attention"): decoder_inputs = enc_dec_attention_layer( decoder_inputs, encoder_outputs, attention_bias, training=training) with tf.name_scope("ffn"): decoder_inputs = feed_forward_network( decoder_inputs, training=training) return self.output_normalization(decoder_inputs)
21,751
38.477314
80
py
models
models-master/official/legacy/transformer/data_pipeline.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Input pipeline for the transformer model to read, filter, and batch examples. Two things to note in the pipeline: 1. Batching scheme The examples encoded in the TFRecord files contain data in the format: {"inputs": [variable length array of integers], "targets": [variable length array of integers]} Where integers in the arrays refer to tokens in the English and German vocab file (named `vocab.ende.32768`). Prior to batching, elements in the dataset are grouped by length (max between "inputs" and "targets" length). Each group is then batched such that: group_batch_size * length <= batch_size. Another way to view batch_size is the maximum number of tokens in each batch. Once batched, each element in the dataset will have the shape: {"inputs": [group_batch_size, padded_input_length], "targets": [group_batch_size, padded_target_length]} Lengths are padded to the longest "inputs" or "targets" sequence in the batch (padded_input_length and padded_target_length can be different). This batching scheme decreases the fraction of padding tokens per training batch, thus improving the training speed significantly. 2. Shuffling While training, the dataset is shuffled in two places in the code. The first is the list of training files. Second, while reading records using `parallel_interleave`, the `sloppy` argument is used to generate randomness in the order of the examples. """ import os from absl import logging import tensorflow as tf from official.utils.misc import model_helpers # Buffer size for reading records from a TFRecord file. Each training file is # 7.2 MB, so 8 MB allows an entire file to be kept in memory. _READ_RECORD_BUFFER = 8 * 1000 * 1000 # Example grouping constants. Defines length boundaries for each group. # These values are the defaults used in Tensor2Tensor. _MIN_BOUNDARY = 8 _BOUNDARY_SCALE = 1.1 def _load_records(filename): """Read file and return a dataset of tf.Examples.""" return tf.data.TFRecordDataset(filename, buffer_size=_READ_RECORD_BUFFER) def _parse_example(serialized_example): """Return inputs and targets Tensors from a serialized tf.Example.""" data_fields = { "inputs": tf.io.VarLenFeature(tf.int64), "targets": tf.io.VarLenFeature(tf.int64) } parsed = tf.io.parse_single_example(serialized_example, data_fields) inputs = tf.sparse.to_dense(parsed["inputs"]) targets = tf.sparse.to_dense(parsed["targets"]) return inputs, targets def _filter_max_length(example, max_length=256): """Indicates whether the example's length is lower than the maximum length.""" return tf.logical_and( tf.size(example[0]) <= max_length, tf.size(example[1]) <= max_length) def _get_example_length(example): """Returns the maximum length between the example inputs and targets.""" length = tf.maximum(tf.shape(example[0])[0], tf.shape(example[1])[0]) return length def _create_min_max_boundaries(max_length, min_boundary=_MIN_BOUNDARY, boundary_scale=_BOUNDARY_SCALE): """Create min and max boundary lists up to max_length. For example, when max_length=24, min_boundary=4 and boundary_scale=2, the returned values will be: buckets_min = [0, 4, 8, 16, 24] buckets_max = [4, 8, 16, 24, 25] Args: max_length: The maximum length of example in dataset. min_boundary: Minimum length in boundary. boundary_scale: Amount to scale consecutive boundaries in the list. Returns: min and max boundary lists """ # Create bucket boundaries list by scaling the previous boundary or adding 1 # (to ensure increasing boundary sizes). bucket_boundaries = [] x = min_boundary while x < max_length: bucket_boundaries.append(x) x = max(x + 1, int(x * boundary_scale)) # Create min and max boundary lists from the initial list. buckets_min = [0] + bucket_boundaries buckets_max = bucket_boundaries + [max_length + 1] return buckets_min, buckets_max def _batch_examples(dataset, batch_size, max_length): """Group examples by similar lengths, and return batched dataset. Each batch of similar-length examples are padded to the same length, and may have different number of elements in each batch, such that: group_batch_size * padded_length <= batch_size. This decreases the number of padding tokens per batch, which improves the training speed. Args: dataset: Dataset of unbatched examples. batch_size: Max number of tokens per batch of examples. max_length: Max number of tokens in an example input or target sequence. Returns: Dataset of batched examples with similar lengths. """ # Get min and max boundary lists for each example. These are used to calculate # the `bucket_id`, which is the index at which: # buckets_min[bucket_id] <= len(example) < buckets_max[bucket_id] # Note that using both min and max lists improves the performance. buckets_min, buckets_max = _create_min_max_boundaries(max_length) # Create list of batch sizes for each bucket_id, so that # bucket_batch_size[bucket_id] * buckets_max[bucket_id] <= batch_size bucket_batch_sizes = [int(batch_size) // x for x in buckets_max] # bucket_id will be a tensor, so convert this list to a tensor as well. bucket_batch_sizes = tf.constant(bucket_batch_sizes, dtype=tf.int64) def example_to_bucket_id(example_input, example_target): """Return int64 bucket id for this example, calculated based on length.""" seq_length = _get_example_length((example_input, example_target)) # TODO(xunkai): investigate if removing code branching improves performance. conditions_c = tf.logical_and( tf.less_equal(buckets_min, seq_length), tf.less(seq_length, buckets_max)) bucket_id = tf.reduce_min(tf.where(conditions_c)) return bucket_id def window_size_fn(bucket_id): """Return number of examples to be grouped when given a bucket id.""" return bucket_batch_sizes[bucket_id] def batching_fn(bucket_id, grouped_dataset): """Batch and add padding to a dataset of elements with similar lengths.""" bucket_batch_size = window_size_fn(bucket_id) # Batch the dataset and add padding so that all input sequences in the # examples have the same length, and all target sequences have the same # lengths as well. Resulting lengths of inputs and targets can differ. return grouped_dataset.padded_batch(bucket_batch_size, ([None], [None])) return dataset.apply( tf.data.experimental.group_by_window( key_func=example_to_bucket_id, reduce_func=batching_fn, window_size=None, window_size_func=window_size_fn)) def _read_and_batch_from_files(file_pattern, batch_size, max_length, max_io_parallelism, shuffle, repeat, static_batch=False, num_replicas=1, ctx=None): """Create dataset where each item is a dict of "inputs" and "targets". Args: file_pattern: String used to match the input TFRecord files. batch_size: Maximum number of tokens per global batch of examples. max_length: Maximum number of tokens per example max_io_parallelism: Max number of cpu cores for parallel input processing. shuffle: If true, randomizes order of elements. repeat: Number of times to repeat the dataset. If None, the dataset is repeated forever. static_batch: Whether the batches in the dataset should have static shapes. If True, the input is batched so that every batch has the shape [batch_size // max_length, max_length]. If False, the input is grouped by length, and batched so that batches may have different shapes [N, M], where: N * M <= batch_size M <= max_length In general, this setting should be False. Dynamic shapes allow the inputs to be grouped so that the number of padding tokens is minimized, and helps model training. In cases where the input shape must be static (e.g. running on TPU), this setting should be set to True. num_replicas: Number of GPUs or other workers. We will generate global batches, and each global batch is equally divisible by number of replicas. Currently it is only effective when static_batch==True. TODO: make it effective when static_batch=False. ctx: Input context. Returns: tf.data.Dataset object containing examples loaded from the files. """ dataset = tf.data.Dataset.list_files(file_pattern, shuffle=shuffle) if ctx and ctx.num_input_pipelines > 1: logging.info("Shard %d of the dataset.", ctx.input_pipeline_id) dataset = dataset.shard(ctx.num_input_pipelines, ctx.input_pipeline_id) # Read files and interleave results. When training, the order of the examples # will be non-deterministic. options = tf.data.Options() options.experimental_deterministic = False dataset = dataset.interleave( _load_records, cycle_length=max_io_parallelism, num_parallel_calls=tf.data.experimental.AUTOTUNE).with_options(options) # Parse each tf.Example into a dictionary # TODO: Look into prefetch_input_elements for performance optimization. # pylint: disable=g-bad-todo dataset = dataset.map( _parse_example, num_parallel_calls=tf.data.experimental.AUTOTUNE) # Remove examples where the input or target length exceeds the maximum length, dataset = dataset.filter(lambda x, y: _filter_max_length((x, y), max_length)) if static_batch: dataset = dataset.padded_batch( # First calculate batch size (token number) per worker, then divide it # into sentences, and finally expand to a global batch. It could prove # the global batch divisble for distribution strategy. int(batch_size // num_replicas // max_length * num_replicas), ([max_length], [max_length]), drop_remainder=True) else: # Group and batch such that each batch has examples of similar length. # TODO(xunkai): _batch_examples might need to do something special for # num_replicas. dataset = _batch_examples(dataset, batch_size, max_length) dataset = dataset.repeat(repeat) # Prefetch the next element to improve speed of input pipeline. dataset = dataset.prefetch(buffer_size=tf.data.experimental.AUTOTUNE) return dataset def _generate_synthetic_data(params): """Create synthetic data based on the parameter batch size.""" batch_size = int(params["batch_size"] // params["max_length"]) length = params["max_length"] dataset = model_helpers.generate_synthetic_data( input_shape=tf.TensorShape([length]), input_value=1, input_dtype=tf.int64, label_shape=tf.TensorShape([length]), label_value=1, label_dtype=tf.int64, ) if params["static_batch"]: dataset = dataset.batch(batch_size, drop_remainder=True) else: dataset = dataset.padded_batch(batch_size, ([None], [None])) return dataset def train_input_fn(params, ctx=None): """Load and return dataset of batched examples for use during training.""" file_pattern = os.path.join(params["data_dir"] or "", "*train*") if params["use_synthetic_data"]: return _generate_synthetic_data(params) return _read_and_batch_from_files( file_pattern, params["batch_size"], params["max_length"], params["max_io_parallelism"], shuffle=True, repeat=params["repeat_dataset"], static_batch=params["static_batch"], num_replicas=params["num_gpus"], ctx=ctx) def eval_input_fn(params, ctx=None): """Load and return dataset of batched examples for use during evaluation.""" file_pattern = os.path.join(params["data_dir"] or "", "*dev*") if params["use_synthetic_data"]: return _generate_synthetic_data(params) return _read_and_batch_from_files( file_pattern, params["batch_size"], params["max_length"], params["max_io_parallelism"], shuffle=False, repeat=1, static_batch=params["static_batch"], num_replicas=params["num_gpus"], ctx=ctx) def map_data_for_transformer_fn(x, y): """Maps data for training, and handles weried behaviors for different vers.""" # Will transform input x and targets y into tuple(x, y) as new model inputs. # For TF v2, the 2nd parameter is omitted to make Keras training work. return ((x, y),)
13,237
38.993958
103
py
models
models-master/official/legacy/transformer/metrics.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Functions for calculating loss, accuracy, and other model metrics. Metrics: - Padded loss, accuracy, and negative log perplexity. Source: https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/metrics.py - BLEU approximation. Source: https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/bleu_hook.py - ROUGE score. Source: https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/rouge.py """ import functools import tensorflow as tf def _pad_tensors_to_same_length(x, y): """Pad x and y so that the results have the same length (second dimension).""" with tf.name_scope("pad_to_same_length"): x_length = tf.shape(x)[1] y_length = tf.shape(y)[1] max_length = tf.maximum(x_length, y_length) x = tf.pad(x, [[0, 0], [0, max_length - x_length], [0, 0]]) y = tf.pad(y, [[0, 0], [0, max_length - y_length]]) return x, y def padded_cross_entropy_loss(logits, labels, smoothing, vocab_size): """Calculate cross entropy loss while ignoring padding. Args: logits: Tensor of size [batch_size, length_logits, vocab_size] labels: Tensor of size [batch_size, length_labels] smoothing: Label smoothing constant, used to determine the on and off values vocab_size: int size of the vocabulary Returns: Returns the cross entropy loss and weight tensors: float32 tensors with shape [batch_size, max(length_logits, length_labels)] """ with tf.name_scope("loss"): logits, labels = _pad_tensors_to_same_length(logits, labels) # Calculate smoothing cross entropy with tf.name_scope("smoothing_cross_entropy"): confidence = 1.0 - smoothing low_confidence = (1.0 - confidence) / tf.cast(vocab_size - 1, tf.float32) soft_targets = tf.one_hot( tf.cast(labels, tf.int32), depth=vocab_size, on_value=confidence, off_value=low_confidence) xentropy = tf.nn.softmax_cross_entropy_with_logits( logits=logits, labels=soft_targets) # Calculate the best (lowest) possible value of cross entropy, and # subtract from the cross entropy loss. normalizing_constant = -( confidence * tf.math.log(confidence) + tf.cast(vocab_size - 1, tf.float32) * low_confidence * tf.math.log(low_confidence + 1e-20)) xentropy -= normalizing_constant weights = tf.cast(tf.not_equal(labels, 0), tf.float32) return xentropy * weights, weights def padded_accuracy(logits, labels): """Percentage of times that predictions matches labels on non-0s.""" with tf.name_scope("padded_accuracy"): logits, labels = _pad_tensors_to_same_length(logits, labels) weights = tf.cast(tf.not_equal(labels, 0), tf.float32) outputs = tf.cast(tf.argmax(logits, axis=-1), tf.int32) padded_labels = tf.cast(labels, tf.int32) return tf.cast(tf.equal(outputs, padded_labels), tf.float32), weights def padded_accuracy_topk(logits, labels, k): """Percentage of times that top-k predictions matches labels on non-0s.""" with tf.name_scope("padded_accuracy_topk"): logits, labels = _pad_tensors_to_same_length(logits, labels) weights = tf.cast(tf.not_equal(labels, 0), tf.float32) effective_k = tf.minimum(k, tf.shape(logits)[-1]) _, outputs = tf.nn.top_k(logits, k=effective_k) outputs = tf.cast(outputs, tf.int32) padded_labels = tf.cast(labels, tf.int32) padded_labels = tf.expand_dims(padded_labels, axis=-1) padded_labels += tf.zeros_like(outputs) # Pad to same shape. same = tf.cast(tf.equal(outputs, padded_labels), tf.float32) same_topk = tf.reduce_sum(same, axis=-1) return same_topk, weights def padded_accuracy_top5(logits, labels): return padded_accuracy_topk(logits, labels, 5) def padded_sequence_accuracy(logits, labels): """Percentage of times that predictions matches labels everywhere (non-0).""" with tf.name_scope("padded_sequence_accuracy"): logits, labels = _pad_tensors_to_same_length(logits, labels) weights = tf.cast(tf.not_equal(labels, 0), tf.float32) outputs = tf.cast(tf.argmax(logits, axis=-1), tf.int32) padded_labels = tf.cast(labels, tf.int32) not_correct = tf.cast(tf.not_equal(outputs, padded_labels), tf.float32) * weights axis = list(range(1, len(outputs.get_shape()))) correct_seq = 1.0 - tf.minimum(1.0, tf.reduce_sum(not_correct, axis=axis)) return correct_seq, tf.constant(1.0) def padded_neg_log_perplexity(logits, labels, vocab_size): """Average log-perplexity excluding padding 0s. No smoothing.""" num, den = padded_cross_entropy_loss(logits, labels, 0, vocab_size) return -num, den class MetricLayer(tf.keras.layers.Layer): """Custom a layer of metrics for Transformer model.""" def __init__(self, vocab_size): super(MetricLayer, self).__init__() self.vocab_size = vocab_size self.metric_mean_fns = [] def build(self, input_shape): """"Builds metric layer.""" neg_log_perplexity = functools.partial( padded_neg_log_perplexity, vocab_size=self.vocab_size) self.metric_mean_fns = [ (tf.keras.metrics.Mean("accuracy"), padded_accuracy), (tf.keras.metrics.Mean("accuracy_top5"), padded_accuracy_top5), (tf.keras.metrics.Mean("accuracy_per_sequence"), padded_sequence_accuracy), (tf.keras.metrics.Mean("neg_log_perplexity"), neg_log_perplexity), ] super(MetricLayer, self).build(input_shape) def get_config(self): return {"vocab_size": self.vocab_size} def call(self, inputs): logits, targets = inputs[0], inputs[1] for mean, fn in self.metric_mean_fns: m = mean(*fn(logits, targets)) self.add_metric(m) return logits def transformer_loss(logits, labels, smoothing, vocab_size): """Calculates total loss containing cross entropy with padding ignored. Args: logits: Tensor of size [batch_size, length_logits, vocab_size] labels: Tensor of size [batch_size, length_labels] smoothing: Label smoothing constant, used to determine the on and off values vocab_size: int size of the vocabulary Returns: A scalar float tensor for loss. """ xentropy, weights = padded_cross_entropy_loss(logits, labels, smoothing, vocab_size) return tf.reduce_sum(xentropy) / tf.reduce_sum(weights)
7,007
37.718232
93
py
models
models-master/official/legacy/transformer/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/transformer/transformer_main.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Train and evaluate the Transformer model. See README for description of setting the training schedule and evaluating the BLEU score. """ import os import tempfile # Import libraries from absl import app from absl import flags from absl import logging import tensorflow as tf from official.common import distribute_utils from official.legacy.transformer import compute_bleu from official.legacy.transformer import data_pipeline from official.legacy.transformer import metrics from official.legacy.transformer import misc from official.legacy.transformer import optimizer from official.legacy.transformer import transformer from official.legacy.transformer import translate from official.legacy.transformer.utils import tokenizer from official.modeling import performance from official.utils.flags import core as flags_core from official.utils.misc import keras_utils # pylint:disable=logging-format-interpolation INF = int(1e9) BLEU_DIR = "bleu" _SINGLE_SAMPLE = 1 def translate_and_compute_bleu(model, params, subtokenizer, bleu_source, bleu_ref, distribution_strategy=None): """Translate file and report the cased and uncased bleu scores. Args: model: A Keras model, used to generate the translations. params: A dictionary, containing the translation related parameters. subtokenizer: A subtokenizer object, used for encoding and decoding source and translated lines. bleu_source: A file containing source sentences for translation. bleu_ref: A file containing the reference for the translated sentences. distribution_strategy: A platform distribution strategy, used for TPU based translation. Returns: uncased_score: A float, the case insensitive BLEU score. cased_score: A float, the case sensitive BLEU score. """ # Create temporary file to store translation. tmp = tempfile.NamedTemporaryFile(delete=False) tmp_filename = tmp.name translate.translate_file( model, params, subtokenizer, bleu_source, output_file=tmp_filename, print_all_translations=False, distribution_strategy=distribution_strategy) # Compute uncased and cased bleu scores. uncased_score = compute_bleu.bleu_wrapper(bleu_ref, tmp_filename, False) cased_score = compute_bleu.bleu_wrapper(bleu_ref, tmp_filename, True) os.remove(tmp_filename) return uncased_score, cased_score def evaluate_and_log_bleu(model, params, bleu_source, bleu_ref, vocab_file, distribution_strategy=None): """Calculate and record the BLEU score. Args: model: A Keras model, used to generate the translations. params: A dictionary, containing the translation related parameters. bleu_source: A file containing source sentences for translation. bleu_ref: A file containing the reference for the translated sentences. vocab_file: A file containing the vocabulary for translation. distribution_strategy: A platform distribution strategy, used for TPU based translation. Returns: uncased_score: A float, the case insensitive BLEU score. cased_score: A float, the case sensitive BLEU score. """ subtokenizer = tokenizer.Subtokenizer(vocab_file) uncased_score, cased_score = translate_and_compute_bleu( model, params, subtokenizer, bleu_source, bleu_ref, distribution_strategy) logging.info("Bleu score (uncased): %s", uncased_score) logging.info("Bleu score (cased): %s", cased_score) return uncased_score, cased_score class TransformerTask(object): """Main entry of Transformer model.""" def __init__(self, flags_obj): """Init function of TransformerMain. Args: flags_obj: Object containing parsed flag values, i.e., FLAGS. Raises: ValueError: if not using static batch for input data on TPU. """ self.flags_obj = flags_obj self.predict_model = None # Add flag-defined parameters to params object num_gpus = flags_core.get_num_gpus(flags_obj) self.params = params = misc.get_model_params(flags_obj.param_set, num_gpus) params["num_gpus"] = num_gpus params["use_ctl"] = flags_obj.use_ctl params["data_dir"] = flags_obj.data_dir params["model_dir"] = flags_obj.model_dir params["static_batch"] = flags_obj.static_batch params["max_length"] = flags_obj.max_length params["decode_batch_size"] = flags_obj.decode_batch_size params["decode_max_length"] = flags_obj.decode_max_length params["padded_decode"] = flags_obj.padded_decode params["max_io_parallelism"] = ( flags_obj.num_parallel_calls or tf.data.experimental.AUTOTUNE) params["use_synthetic_data"] = flags_obj.use_synthetic_data params["batch_size"] = flags_obj.batch_size or params["default_batch_size"] params["repeat_dataset"] = None params["dtype"] = flags_core.get_tf_dtype(flags_obj) params["enable_tensorboard"] = flags_obj.enable_tensorboard params["enable_metrics_in_training"] = flags_obj.enable_metrics_in_training params["steps_between_evals"] = flags_obj.steps_between_evals params["enable_checkpointing"] = flags_obj.enable_checkpointing params["save_weights_only"] = flags_obj.save_weights_only self.distribution_strategy = distribute_utils.get_distribution_strategy( distribution_strategy=flags_obj.distribution_strategy, num_gpus=num_gpus, all_reduce_alg=flags_obj.all_reduce_alg, num_packs=flags_obj.num_packs, tpu_address=flags_obj.tpu or "") if self.use_tpu: params["num_replicas"] = self.distribution_strategy.num_replicas_in_sync else: logging.info("Running transformer with num_gpus = %d", num_gpus) if self.distribution_strategy: logging.info("For training, using distribution strategy: %s", self.distribution_strategy) else: logging.info("Not using any distribution strategy.") performance.set_mixed_precision_policy(params["dtype"]) @property def use_tpu(self): if self.distribution_strategy: return isinstance(self.distribution_strategy, tf.distribute.TPUStrategy) return False def train(self): """Trains the model.""" params = self.params flags_obj = self.flags_obj # Sets config options. keras_utils.set_session_config(enable_xla=flags_obj.enable_xla) _ensure_dir(flags_obj.model_dir) with distribute_utils.get_strategy_scope(self.distribution_strategy): model = transformer.create_model(params, is_train=True) opt = self._create_optimizer() current_step = 0 checkpoint = tf.train.Checkpoint(model=model, optimizer=opt) latest_checkpoint = tf.train.latest_checkpoint(flags_obj.model_dir) if latest_checkpoint: checkpoint.restore(latest_checkpoint) logging.info("Loaded checkpoint %s", latest_checkpoint) current_step = opt.iterations.numpy() if params["use_ctl"]: train_loss_metric = tf.keras.metrics.Mean( "training_loss", dtype=tf.float32) if params["enable_tensorboard"]: summary_writer = tf.summary.create_file_writer( os.path.join(flags_obj.model_dir, "summary")) else: summary_writer = tf.summary.create_noop_writer() train_metrics = [train_loss_metric] if params["enable_metrics_in_training"]: train_metrics = train_metrics + model.metrics else: model.compile(opt) model.summary() if self.use_tpu: # Different from experimental_distribute_dataset, # distribute_datasets_from_function requires # per-replica/local batch size. params["batch_size"] /= self.distribution_strategy.num_replicas_in_sync train_ds = ( self.distribution_strategy.distribute_datasets_from_function( lambda ctx: data_pipeline.train_input_fn(params, ctx))) else: train_ds = data_pipeline.train_input_fn(params) map_data_fn = data_pipeline.map_data_for_transformer_fn train_ds = train_ds.map( map_data_fn, num_parallel_calls=tf.data.experimental.AUTOTUNE) if params["use_ctl"]: train_ds_iterator = iter(train_ds) callbacks = self._create_callbacks(flags_obj.model_dir, params) # Only TimeHistory callback is supported for CTL if params["use_ctl"]: callbacks = [cb for cb in callbacks if isinstance(cb, keras_utils.TimeHistory)] @tf.function def train_steps(iterator, steps): """Training steps function for TPU runs. Args: iterator: The input iterator of the training dataset. steps: An integer, the number of training steps. Returns: A float, the loss value. """ def _step_fn(inputs): """Per-replica step function.""" inputs, targets = inputs with tf.GradientTape() as tape: logits = model([inputs, targets], training=True) loss = metrics.transformer_loss(logits, targets, params["label_smoothing"], params["vocab_size"]) # Scales the loss, which results in using the average loss across all # of the replicas for backprop. scaled_loss = loss / self.distribution_strategy.num_replicas_in_sync # De-dupes variables due to keras tracking issues. tvars = list({id(v): v for v in model.trainable_variables}.values()) grads = tape.gradient(scaled_loss, tvars) opt.apply_gradients(zip(grads, tvars)) # For reporting, the metric takes the mean of losses. train_loss_metric.update_state(loss) for _ in tf.range(steps): train_loss_metric.reset_states() self.distribution_strategy.run( _step_fn, args=(next(iterator),)) cased_score, uncased_score = None, None cased_score_history, uncased_score_history = [], [] while current_step < flags_obj.train_steps: remaining_steps = flags_obj.train_steps - current_step train_steps_per_eval = ( remaining_steps if remaining_steps < flags_obj.steps_between_evals else flags_obj.steps_between_evals) current_iteration = current_step // flags_obj.steps_between_evals logging.info( "Start train iteration at global step:{}".format(current_step)) history = None if params["use_ctl"]: if not self.use_tpu: raise NotImplementedError( "Custom training loop on GPUs is not implemented.") # Runs training steps. with summary_writer.as_default(): for cb in callbacks: cb.on_epoch_begin(current_iteration) cb.on_batch_begin(0) train_steps( train_ds_iterator, tf.convert_to_tensor(train_steps_per_eval, dtype=tf.int32)) current_step += train_steps_per_eval train_loss = train_loss_metric.result().numpy().astype(float) logging.info("Train Step: %d/%d / loss = %s", current_step, flags_obj.train_steps, train_loss) for cb in callbacks: cb.on_batch_end(train_steps_per_eval - 1) cb.on_epoch_end(current_iteration) if params["enable_tensorboard"]: for metric_obj in train_metrics: tf.summary.scalar(metric_obj.name, metric_obj.result(), current_step) summary_writer.flush() for cb in callbacks: cb.on_train_end() if flags_obj.enable_checkpointing: # avoid check-pointing when running for benchmarking. checkpoint_name = checkpoint.save( os.path.join(flags_obj.model_dir, "ctl_step_{}.ckpt".format(current_step))) logging.info("Saved checkpoint to %s", checkpoint_name) else: if self.use_tpu: raise NotImplementedError( "Keras model.fit on TPUs is not implemented.") history = model.fit( train_ds, initial_epoch=current_iteration, epochs=current_iteration + 1, steps_per_epoch=train_steps_per_eval, callbacks=callbacks, # If TimeHistory is enabled, progress bar would be messy. Increase # the verbose level to get rid of it. verbose=(2 if flags_obj.enable_time_history else 1)) current_step += train_steps_per_eval logging.info("Train history: {}".format(history.history)) logging.info("End train iteration at global step:{}".format(current_step)) if (flags_obj.bleu_source and flags_obj.bleu_ref): uncased_score, cased_score = self.eval() cased_score_history.append([current_iteration + 1, cased_score]) uncased_score_history.append([current_iteration + 1, uncased_score]) stats = ({ "loss": train_loss } if history is None else {}) misc.update_stats(history, stats, callbacks) if uncased_score and cased_score: stats["bleu_uncased"] = uncased_score stats["bleu_cased"] = cased_score stats["bleu_uncased_history"] = uncased_score_history stats["bleu_cased_history"] = cased_score_history return stats def eval(self): """Evaluates the model.""" distribution_strategy = self.distribution_strategy if self.use_tpu else None # We only want to create the model under DS scope for TPU case. # When 'distribution_strategy' is None, a no-op DummyContextManager will # be used. with distribute_utils.get_strategy_scope(distribution_strategy): if not self.predict_model: self.predict_model = transformer.create_model(self.params, False) self._load_weights_if_possible( self.predict_model, tf.train.latest_checkpoint(self.flags_obj.model_dir)) self.predict_model.summary() return evaluate_and_log_bleu( self.predict_model, self.params, self.flags_obj.bleu_source, self.flags_obj.bleu_ref, self.flags_obj.vocab_file, distribution_strategy) def predict(self): """Predicts result from the model.""" params = self.params flags_obj = self.flags_obj with tf.name_scope("model"): model = transformer.create_model(params, is_train=False) self._load_weights_if_possible( model, tf.train.latest_checkpoint(self.flags_obj.model_dir)) model.summary() subtokenizer = tokenizer.Subtokenizer(flags_obj.vocab_file) ds = data_pipeline.eval_input_fn(params) ds = ds.map(lambda x, y: x).take(_SINGLE_SAMPLE) ret = model.predict(ds) val_outputs, _ = ret length = len(val_outputs) for i in range(length): translate.translate_from_input(val_outputs[i], subtokenizer) def _create_callbacks(self, cur_log_dir, params): """Creates a list of callbacks.""" callbacks = misc.get_callbacks() if params["enable_checkpointing"]: ckpt_full_path = os.path.join(cur_log_dir, "cp-{epoch:04d}.ckpt") callbacks.append( tf.keras.callbacks.ModelCheckpoint( ckpt_full_path, save_weights_only=params["save_weights_only"])) return callbacks def _load_weights_if_possible(self, model, init_weight_path=None): """Loads model weights when it is provided.""" if init_weight_path: logging.info("Load weights: {}".format(init_weight_path)) if self.use_tpu: checkpoint = tf.train.Checkpoint( model=model, optimizer=self._create_optimizer()) checkpoint.restore(init_weight_path) else: model.load_weights(init_weight_path) else: logging.info("Weights not loaded from path:{}".format(init_weight_path)) def _create_optimizer(self): """Creates optimizer.""" params = self.params lr_schedule = optimizer.LearningRateSchedule( params["learning_rate"], params["hidden_size"], params["learning_rate_warmup_steps"]) opt = tf.keras.optimizers.Adam( lr_schedule, params["optimizer_adam_beta1"], params["optimizer_adam_beta2"], epsilon=params["optimizer_adam_epsilon"]) opt = performance.configure_optimizer( opt, use_float16=params["dtype"] == tf.float16, loss_scale=flags_core.get_loss_scale( self.flags_obj, default_for_fp16="dynamic")) return opt def _ensure_dir(log_dir): """Makes log dir if not existed.""" if not tf.io.gfile.exists(log_dir): tf.io.gfile.makedirs(log_dir) def main(_): flags_obj = flags.FLAGS if flags_obj.enable_mlir_bridge: tf.config.experimental.enable_mlir_bridge() task = TransformerTask(flags_obj) # Execute flag override logic for better model performance if flags_obj.tf_gpu_thread_mode: keras_utils.set_gpu_thread_mode_and_count( per_gpu_thread_count=flags_obj.per_gpu_thread_count, gpu_thread_mode=flags_obj.tf_gpu_thread_mode, num_gpus=flags_obj.num_gpus, datasets_num_private_threads=flags_obj.datasets_num_private_threads) if flags_obj.mode == "train": task.train() elif flags_obj.mode == "predict": task.predict() elif flags_obj.mode == "eval": task.eval() else: raise ValueError("Invalid mode {}".format(flags_obj.mode)) if __name__ == "__main__": logging.set_verbosity(logging.INFO) misc.define_transformer_flags() app.run(main)
18,183
36.415638
80
py
models
models-master/official/legacy/transformer/attention_layer.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Implementation of multiheaded attention and self-attention layers.""" import math import tensorflow as tf from official.modeling import tf_utils class Attention(tf.keras.layers.Layer): """Multi-headed attention layer.""" def __init__(self, hidden_size, num_heads, attention_dropout): """Initialize Attention. Args: hidden_size: int, output dim of hidden layer. num_heads: int, number of heads to repeat the same attention structure. attention_dropout: float, dropout rate inside attention for training. """ if hidden_size % num_heads: raise ValueError( "Hidden size ({}) must be divisible by the number of heads ({})." .format(hidden_size, num_heads)) super(Attention, self).__init__() self.hidden_size = hidden_size self.num_heads = num_heads self.attention_dropout = attention_dropout def build(self, input_shape): """Builds the layer.""" # Layers for linearly projecting the queries, keys, and values. size_per_head = self.hidden_size // self.num_heads def _glorot_initializer(fan_in, fan_out): limit = math.sqrt(6.0 / (fan_in + fan_out)) return tf.keras.initializers.RandomUniform(minval=-limit, maxval=limit) attention_initializer = _glorot_initializer(input_shape.as_list()[-1], self.hidden_size) self.query_dense_layer = tf.keras.layers.EinsumDense( "BTE,ENH->BTNH", output_shape=(None, self.num_heads, size_per_head), kernel_initializer=tf_utils.clone_initializer(attention_initializer), bias_axes=None, name="query") self.key_dense_layer = tf.keras.layers.EinsumDense( "BTE,ENH->BTNH", output_shape=(None, self.num_heads, size_per_head), kernel_initializer=tf_utils.clone_initializer(attention_initializer), bias_axes=None, name="key") self.value_dense_layer = tf.keras.layers.EinsumDense( "BTE,ENH->BTNH", output_shape=(None, self.num_heads, size_per_head), kernel_initializer=tf_utils.clone_initializer(attention_initializer), bias_axes=None, name="value") output_initializer = _glorot_initializer(self.hidden_size, self.hidden_size) self.output_dense_layer = tf.keras.layers.EinsumDense( "BTNH,NHE->BTE", output_shape=(None, self.hidden_size), kernel_initializer=output_initializer, bias_axes=None, name="output_transform") super(Attention, self).build(input_shape) def get_config(self): return { "hidden_size": self.hidden_size, "num_heads": self.num_heads, "attention_dropout": self.attention_dropout, } def call(self, query_input, source_input, bias, training, cache=None, decode_loop_step=None): """Apply attention mechanism to query_input and source_input. Args: query_input: A tensor with shape [batch_size, length_query, hidden_size]. source_input: A tensor with shape [batch_size, length_source, hidden_size]. bias: A tensor with shape [batch_size, 1, length_query, length_source], the attention bias that will be added to the result of the dot product. training: A bool, whether in training mode or not. cache: (Used during prediction) A dictionary with tensors containing results of previous attentions. The dictionary must have the items: {"k": tensor with shape [batch_size, i, heads, dim_per_head], "v": tensor with shape [batch_size, i, heads, dim_per_head]} where i is the current decoded length for non-padded decode, or max sequence length for padded decode. decode_loop_step: An integer, step number of the decoding loop. Used only for autoregressive inference on TPU. Returns: Attention layer output with shape [batch_size, length_query, hidden_size] """ # Linearly project the query, key and value using different learned # projections. Splitting heads is automatically done during the linear # projections --> [batch_size, length, num_heads, dim_per_head]. query = self.query_dense_layer(query_input) key = self.key_dense_layer(source_input) value = self.value_dense_layer(source_input) if cache is not None: # Combine cached keys and values with new keys and values. if decode_loop_step is not None: cache_k_shape = cache["k"].shape.as_list() indices = tf.reshape( tf.one_hot(decode_loop_step, cache_k_shape[1], dtype=key.dtype), [1, cache_k_shape[1], 1, 1]) key = cache["k"] + key * indices cache_v_shape = cache["v"].shape.as_list() indices = tf.reshape( tf.one_hot(decode_loop_step, cache_v_shape[1], dtype=value.dtype), [1, cache_v_shape[1], 1, 1]) value = cache["v"] + value * indices else: key = tf.concat([tf.cast(cache["k"], key.dtype), key], axis=1) value = tf.concat([tf.cast(cache["v"], value.dtype), value], axis=1) # Update cache cache["k"] = key cache["v"] = value # Scale query to prevent the dot product between query and key from growing # too large. depth = (self.hidden_size // self.num_heads) query *= depth**-0.5 # Calculate dot product attention logits = tf.einsum("BTNH,BFNH->BNFT", key, query) logits += bias # Note that softmax internally performs math operations using float32 # for numeric stability. When training with float16, we keep the input # and output in float16 for better performance. weights = tf.nn.softmax(logits, name="attention_weights") if training: weights = tf.nn.dropout(weights, rate=self.attention_dropout) attention_output = tf.einsum("BNFT,BTNH->BFNH", weights, value) # Run the outputs through another linear projection layer. Recombining heads # is automatically done --> [batch_size, length, hidden_size] attention_output = self.output_dense_layer(attention_output) return attention_output class SelfAttention(Attention): """Multiheaded self-attention layer.""" def call(self, query_input, bias, training, cache=None, decode_loop_step=None): return super(SelfAttention, self).call(query_input, query_input, bias, training, cache, decode_loop_step)
7,119
38.776536
80
py
models
models-master/official/legacy/transformer/optimizer.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Optimizer from addons and learning rate scheduler.""" import tensorflow as tf class LearningRateSchedule(tf.keras.optimizers.schedules.LearningRateSchedule): """Learning rate schedule.""" def __init__(self, initial_learning_rate, hidden_size, warmup_steps): """Initialize configuration of the learning rate schedule. Args: initial_learning_rate: A float, the initial learning rate. hidden_size: An integer, the model dimension in the hidden layers. warmup_steps: An integer, the number of steps required for linear warmup. """ super(LearningRateSchedule, self).__init__() self.initial_learning_rate = initial_learning_rate self.hidden_size = hidden_size self.warmup_steps = warmup_steps self.warmup_steps_tensor = tf.cast(warmup_steps, tf.float32) def __call__(self, global_step): """Calculate learning rate with linear warmup and rsqrt decay. Args: global_step: An integer, the current global step used for learning rate calculation. Returns: A float, the learning rate needs to be used for current global step. """ with tf.name_scope('learning_rate_schedule'): global_step = tf.cast(global_step, tf.float32) learning_rate = self.initial_learning_rate learning_rate *= (self.hidden_size**-0.5) # Apply linear warmup learning_rate *= tf.minimum(1.0, global_step / self.warmup_steps_tensor) # Apply rsqrt decay learning_rate /= tf.sqrt( tf.maximum(global_step, self.warmup_steps_tensor)) return learning_rate def get_config(self): """Get the configuration of the learning rate schedule.""" return { 'initial_learning_rate': self.initial_learning_rate, 'hidden_size': self.hidden_size, 'warmup_steps': self.warmup_steps, }
2,434
36.461538
79
py
models
models-master/official/legacy/transformer/utils/tokenizer.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Defines Subtokenizer class to encode and decode strings.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import collections import re import sys import unicodedata from absl import logging import numpy as np import six from six.moves import xrange # pylint: disable=redefined-builtin import tensorflow as tf # pylint: disable=g-complex-comprehension PAD = "<pad>" PAD_ID = 0 EOS = "<EOS>" EOS_ID = 1 RESERVED_TOKENS = [PAD, EOS] # Set of characters that will be used in the function _escape_token() (see func # docstring for more details). # This set is added to the alphabet list to ensure that all escaped tokens can # be encoded. _ESCAPE_CHARS = set(u"\\_u;0123456789") # Regex for the function _unescape_token(), the inverse of _escape_token(). # This is used to find "\u", "\\", and "\###;" substrings in the token. _UNESCAPE_REGEX = re.compile(r"\\u|\\\\|\\([0-9]+);") _UNDEFINED_UNICODE = u"\u3013" def alphanumeric_char_set(): return set( six.unichr(i) for i in xrange(sys.maxunicode) if (unicodedata.category(six.unichr(i)).startswith("L") or unicodedata.category(six.unichr(i)).startswith("N"))) # Set contains all letter and number characters. _ALPHANUMERIC_CHAR_SET = alphanumeric_char_set() # min_count is the minimum number of times a subtoken must appear in the data # before before it is added to the vocabulary. The value is found using binary # search to obtain the target vocabulary size. _MIN_MIN_COUNT = 1 # min value to use when binary searching for min_count _MAX_MIN_COUNT = 1000 # max value to use when binary searching for min_count class Subtokenizer(object): """Encodes and decodes strings to/from integer IDs.""" def __init__(self, vocab_file, reserved_tokens=None, master_char_set=None): """Initializes class, creating a vocab file if data_files is provided.""" logging.info("Initializing Subtokenizer from file %s.", vocab_file) if master_char_set is None: master_char_set = _ALPHANUMERIC_CHAR_SET if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS self.subtoken_list = _load_vocab_file(vocab_file, reserved_tokens) self.alphabet = _generate_alphabet_dict(self.subtoken_list) self.subtoken_to_id_dict = _list_to_index_dict(self.subtoken_list) self.max_subtoken_length = 0 for subtoken in self.subtoken_list: self.max_subtoken_length = max(self.max_subtoken_length, len(subtoken)) # Create cache to speed up subtokenization self._cache_size = 2**20 self._cache = [(None, None)] * self._cache_size self._master_char_set = master_char_set @staticmethod def init_from_files(vocab_file, files, target_vocab_size, threshold, min_count=None, file_byte_limit=1e6, reserved_tokens=None, correct_strip=True, master_char_set=None): """Create subtoken vocabulary based on files, and save vocab to file. Args: vocab_file: String name of vocab file to store subtoken vocabulary. files: List of file paths that will be used to generate vocabulary. target_vocab_size: target vocabulary size to generate. threshold: int threshold of vocabulary size to accept. min_count: int minimum count to use for generating the vocabulary. The min count is the minimum number of times a subtoken should appear in the files before it is added to the vocabulary. If set to none, this value is found using binary search. file_byte_limit: (Default 1e6) Maximum number of bytes of sample text that will be drawn from the files. reserved_tokens: List of string tokens that are guaranteed to be at the beginning of the subtoken vocabulary list. correct_strip: Whether to convert text to unicode before strip. master_char_set: the char set. Returns: Subtokenizer object """ if master_char_set is None: master_char_set = _ALPHANUMERIC_CHAR_SET if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS if tf.io.gfile.exists(vocab_file): logging.info("Vocab file already exists (%s)", vocab_file) else: logging.info("Begin steps to create subtoken vocabulary...") token_counts = _count_tokens(files, file_byte_limit, correct_strip, master_char_set) alphabet = _generate_alphabet_dict(token_counts) subtoken_list = _generate_subtokens_with_target_vocab_size( token_counts, alphabet, target_vocab_size, threshold, min_count, reserved_tokens) logging.info("Generated vocabulary with %d subtokens.", len(subtoken_list)) _save_vocab_file(vocab_file, subtoken_list) return Subtokenizer(vocab_file, master_char_set=master_char_set) def encode(self, raw_string, add_eos=False): """Encodes a string into a list of int subtoken ids.""" ret = [] tokens = _split_string_to_tokens( native_to_unicode(raw_string), self._master_char_set) for token in tokens: ret.extend(self._token_to_subtoken_ids(token)) if add_eos: assert EOS in self.subtoken_list, \ "Can't append 'EOS' because it is not in list of known subtokens." ret.append(EOS_ID) return ret def _token_to_subtoken_ids(self, token): """Encode a single token into a list of subtoken ids.""" cache_location = hash(token) % self._cache_size cache_key, cache_value = self._cache[cache_location] if cache_key == token: return cache_value ret = _split_token_to_subtokens( _escape_token(token, self.alphabet), self.subtoken_to_id_dict, self.max_subtoken_length) ret = [self.subtoken_to_id_dict[subtoken_id] for subtoken_id in ret] self._cache[cache_location] = (token, ret) return ret def decode(self, subtokens): """Converts list of int subtokens ids into a string.""" if isinstance(subtokens, np.ndarray): # Note that list(subtokens) converts subtokens to a python list, but the # items remain as np.int32. This converts both the array and its items. subtokens = subtokens.tolist() if not subtokens: return "" assert isinstance(subtokens, list) and isinstance(subtokens[0], int), ( "Subtokens argument passed into decode() must be a list of integers.") return _unicode_to_native( _join_tokens_to_string( self._subtoken_ids_to_tokens(subtokens), self._master_char_set)) def _subtoken_ids_to_tokens(self, subtokens): """Convert list of int subtoken ids to a list of string tokens.""" escaped_tokens = "".join([ self.subtoken_list[s] for s in subtokens if s < len(self.subtoken_list) ]) escaped_tokens = escaped_tokens.split("_") # All tokens in the vocabulary list have been escaped (see _escape_token()) # so each token must be unescaped when decoding. ret = [] for token in escaped_tokens: if token: ret.append(_unescape_token(token)) return ret def _save_vocab_file(vocab_file, subtoken_list): """Save subtokens to file.""" with tf.io.gfile.GFile(vocab_file, mode="w") as f: for subtoken in subtoken_list: f.write("'%s'\n" % _unicode_to_native(subtoken)) def _load_vocab_file(vocab_file, reserved_tokens=None): """Load vocabulary while ensuring reserved tokens are at the top.""" if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS subtoken_list = [] with tf.io.gfile.GFile(vocab_file, mode="r") as f: for line in f: subtoken = native_to_unicode(line.strip()) subtoken = subtoken[1:-1] # Remove surrounding single-quotes if subtoken in reserved_tokens: continue subtoken_list.append(native_to_unicode(subtoken)) return reserved_tokens + subtoken_list def native_to_unicode(s): """Convert string to unicode (required in Python 2).""" try: # Python 2 return s if isinstance(s, unicode) else s.decode("utf-8") except NameError: # Python 3 return s def _unicode_to_native(s): """Convert string from unicode to native format (required in Python 2).""" try: # Python 2 return s.encode("utf-8") if isinstance(s, unicode) else s except NameError: # Python 3 return s def _split_string_to_tokens(text, master_char_set): """Splits text to a list of string tokens.""" if not text: return [] ret = [] token_start = 0 # Classify each character in the input string is_master = [c in master_char_set for c in text] for pos in xrange(1, len(text)): if is_master[pos] != is_master[pos - 1]: token = text[token_start:pos] if token != u" " or token_start == 0: ret.append(token) token_start = pos final_token = text[token_start:] ret.append(final_token) return ret def _join_tokens_to_string(tokens, master_char_set): """Join a list of string tokens into a single string.""" token_is_master = [t[0] in master_char_set for t in tokens] ret = [] for i, token in enumerate(tokens): if i > 0 and token_is_master[i - 1] and token_is_master[i]: ret.append(u" ") ret.append(token) return "".join(ret) def _escape_token(token, alphabet): r"""Replace characters that aren't in the alphabet and append "_" to token. Apply three transformations to the token: 1. Replace underline character "_" with "\u", and backslash "\" with "\\". 2. Replace characters outside of the alphabet with "\###;", where ### is the character's Unicode code point. 3. Appends "_" to mark the end of a token. Args: token: unicode string to be escaped alphabet: list of all known characters Returns: escaped string """ token = token.replace(u"\\", u"\\\\").replace(u"_", u"\\u") ret = [c if c in alphabet and c != u"\n" else r"\%d;" % ord(c) for c in token] return u"".join(ret) + "_" def _unescape_token(token): r"""Replaces escaped characters in the token with their unescaped versions. Applies inverse transformations as _escape_token(): 1. Replace "\u" with "_", and "\\" with "\". 2. Replace "\###;" with the unicode character the ### refers to. Args: token: escaped string Returns: unescaped string """ def match(m): r"""Returns replacement string for matched object. Matched objects contain one of the strings that matches the regex pattern: r"\\u|\\\\|\\([0-9]+);" The strings can be '\u', '\\', or '\###;' (### is any digit number). m.group(0) refers to the entire matched string ('\u', '\\', or '\###;'). m.group(1) refers to the first parenthesized subgroup ('###'). m.group(0) exists for all match objects, while m.group(1) exists only for the string '\###;'. This function looks to see if m.group(1) exists. If it doesn't, then the matched string must be '\u' or '\\' . In this case, the corresponding replacement ('_' and '\') are returned. Note that in python, a single backslash is written as '\\', and double backslash as '\\\\'. If m.goup(1) exists, then use the integer in m.group(1) to return a unicode character. Args: m: match object Returns: String to replace matched object with. """ # Check if the matched strings are '\u' or '\\'. if m.group(1) is None: return u"_" if m.group(0) == u"\\u" else u"\\" # If m.group(1) exists, try and return unicode character. try: return six.unichr(int(m.group(1))) except (ValueError, OverflowError) as _: return _UNDEFINED_UNICODE # Use match function to replace escaped substrings in the token. return _UNESCAPE_REGEX.sub(match, token) def _count_tokens(files, file_byte_limit=1e6, correct_strip=True, master_char_set=None): """Return token counts of words in the files. Samples file_byte_limit bytes from each file, and counts the words that appear in the samples. The samples are semi-evenly distributed across the file. Args: files: List of filepaths file_byte_limit: Max number of bytes that will be read from each file. correct_strip: Whether to convert text to unicode before strip. This affects vocabulary generation for PY2. Sets correct_strip to False in PY2 to reproduce previous common public result. Sets correct_strip to True will let PY2 and PY3 get a consistent vocabulary. master_char_set: the char set. Returns: Dictionary mapping tokens to the number of times they appear in the sampled lines from the files. """ if master_char_set is None: master_char_set = _ALPHANUMERIC_CHAR_SET token_counts = collections.defaultdict(int) for filepath in files: with tf.io.gfile.GFile(filepath, mode="r") as reader: file_byte_budget = file_byte_limit counter = 0 lines_to_skip = int(reader.size() / (file_byte_budget * 2)) for line in reader: if counter < lines_to_skip: counter += 1 else: if file_byte_budget < 0: break if correct_strip: line = native_to_unicode(line) line = line.strip() file_byte_budget -= len(line) counter = 0 # Add words to token counts for token in _split_string_to_tokens( native_to_unicode(line), master_char_set): token_counts[token] += 1 return token_counts def _list_to_index_dict(lst): """Create dictionary mapping list items to their indices in the list.""" return {item: n for n, item in enumerate(lst)} def _split_token_to_subtokens(token, subtoken_dict, max_subtoken_length): """Splits a token into subtokens defined in the subtoken dict.""" ret = [] start = 0 token_len = len(token) while start < token_len: # Find the longest subtoken, so iterate backwards. for end in xrange(min(token_len, start + max_subtoken_length), start, -1): subtoken = token[start:end] if subtoken in subtoken_dict: ret.append(subtoken) start = end break else: # Did not break # If there is no possible encoding of the escaped token then one of the # characters in the token is not in the alphabet. This should be # impossible and would be indicative of a bug. raise ValueError("Was unable to split token \"%s\" into subtokens." % token) return ret def _generate_subtokens_with_target_vocab_size(token_counts, alphabet, target_size, threshold, min_count=None, reserved_tokens=None): """Generate subtoken vocabulary close to the target size.""" if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS if min_count is not None: logging.info("Using min_count=%d to generate vocab with target size %d", min_count, target_size) return _generate_subtokens( token_counts, alphabet, min_count, reserved_tokens=reserved_tokens) def bisect(min_val, max_val): """Recursive function to binary search for subtoken vocabulary.""" cur_count = (min_val + max_val) // 2 logging.info("Binary search: trying min_count=%d (%d %d)", cur_count, min_val, max_val) subtoken_list = _generate_subtokens( token_counts, alphabet, cur_count, reserved_tokens=reserved_tokens) val = len(subtoken_list) logging.info("Binary search: min_count=%d resulted in %d tokens", cur_count, val) within_threshold = abs(val - target_size) < threshold if within_threshold or min_val >= max_val or cur_count < 2: return subtoken_list if val > target_size: other_subtoken_list = bisect(cur_count + 1, max_val) else: other_subtoken_list = bisect(min_val, cur_count - 1) # Return vocabulary dictionary with the closest number of tokens. other_val = len(other_subtoken_list) if abs(other_val - target_size) < abs(val - target_size): return other_subtoken_list return subtoken_list logging.info("Finding best min_count to get target size of %d", target_size) return bisect(_MIN_MIN_COUNT, _MAX_MIN_COUNT) def _generate_alphabet_dict(iterable, reserved_tokens=None): """Create set of characters that appear in any element in the iterable.""" if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS alphabet = {c for token in iterable for c in token} alphabet |= {c for token in reserved_tokens for c in token} alphabet |= _ESCAPE_CHARS # Add escape characters to alphabet set. return alphabet def _count_and_gen_subtokens(token_counts, alphabet, subtoken_dict, max_subtoken_length): """Count number of times subtokens appear, and generate new subtokens. Args: token_counts: dict mapping tokens to the number of times they appear in the original files. alphabet: list of allowed characters. Used to escape the tokens, which guarantees that all tokens can be split into subtokens. subtoken_dict: dict mapping subtokens to ids. max_subtoken_length: maximum length of subtoken in subtoken_dict. Returns: A defaultdict mapping subtokens to the number of times they appear in the tokens. The dict may contain new subtokens. """ subtoken_counts = collections.defaultdict(int) for token, count in six.iteritems(token_counts): token = _escape_token(token, alphabet) subtokens = _split_token_to_subtokens(token, subtoken_dict, max_subtoken_length) # Generate new subtokens by taking substrings from token. start = 0 for subtoken in subtokens: for end in xrange(start + 1, len(token) + 1): new_subtoken = token[start:end] subtoken_counts[new_subtoken] += count start += len(subtoken) return subtoken_counts def _filter_and_bucket_subtokens(subtoken_counts, min_count): """Return a bucketed list of subtokens that are filtered by count. Args: subtoken_counts: defaultdict mapping subtokens to their counts min_count: int count used to filter subtokens Returns: List of subtoken sets, where subtokens in set i have the same length=i. """ # Create list of buckets, where subtokens in bucket i have length i. subtoken_buckets = [] for subtoken, count in six.iteritems(subtoken_counts): if count < min_count: # Filter out subtokens that don't appear enough continue while len(subtoken_buckets) <= len(subtoken): subtoken_buckets.append(set()) subtoken_buckets[len(subtoken)].add(subtoken) return subtoken_buckets def _gen_new_subtoken_list(subtoken_counts, min_count, alphabet, reserved_tokens=None): """Generate candidate subtokens ordered by count, and new max subtoken length. Add subtokens to the candiate list in order of length (longest subtokens first). When a subtoken is added, the counts of each of its prefixes are decreased. Prefixes that don't appear much outside the subtoken are not added to the candidate list. For example: subtoken being added to candidate list: 'translate' subtoken_counts: {'translate':10, 't':40, 'tr':16, 'tra':12, ...} min_count: 5 When 'translate' is added, subtoken_counts is updated to: {'translate':0, 't':30, 'tr':6, 'tra': 2, ...} The subtoken 'tra' will not be added to the candidate list, because it appears twice (less than min_count) outside of 'translate'. Args: subtoken_counts: defaultdict mapping str subtokens to int counts min_count: int minumum count requirement for subtokens alphabet: set of characters. Each character is added to the subtoken list to guarantee that all tokens can be encoded. reserved_tokens: list of tokens that will be added to the beginning of the returned subtoken list. Returns: List of candidate subtokens in decreasing count order, and maximum subtoken length """ if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS # Create a list of (count, subtoken) for each candidate subtoken. subtoken_candidates = [] # Use bucketted list to iterate through subtokens in order of length. # subtoken_buckets[i] = set(subtokens), where each subtoken has length i. subtoken_buckets = _filter_and_bucket_subtokens(subtoken_counts, min_count) max_subtoken_length = len(subtoken_buckets) - 1 # Go through the list in reverse order to consider longer subtokens first. for subtoken_len in xrange(max_subtoken_length, 0, -1): for subtoken in subtoken_buckets[subtoken_len]: count = subtoken_counts[subtoken] # Possible if this subtoken is a prefix of another token. if count < min_count: continue # Ignore alphabet/reserved tokens, which will be added manually later. if subtoken not in alphabet and subtoken not in reserved_tokens: subtoken_candidates.append((count, subtoken)) # Decrement count of the subtoken's prefixes (if a longer subtoken is # added, its prefixes lose priority to be added). for end in xrange(1, subtoken_len): subtoken_counts[subtoken[:end]] -= count # Add alphabet subtokens (guarantees that all strings are encodable). subtoken_candidates.extend((subtoken_counts.get(a, 0), a) for a in alphabet) # Order subtoken candidates by decreasing count. subtoken_list = [t for _, t in sorted(subtoken_candidates, reverse=True)] # Add reserved tokens to beginning of the list. subtoken_list = reserved_tokens + subtoken_list return subtoken_list, max_subtoken_length def _generate_subtokens(token_counts, alphabet, min_count, num_iterations=4, reserved_tokens=None): """Create a list of subtokens in decreasing order of frequency. Args: token_counts: dict mapping str tokens -> int count alphabet: set of characters min_count: int minimum number of times a subtoken must appear before it is added to the vocabulary. num_iterations: int number of iterations to generate new tokens. reserved_tokens: list of tokens that will be added to the beginning to the returned subtoken list. Returns: Sorted list of subtokens (most frequent first) """ if reserved_tokens is None: reserved_tokens = RESERVED_TOKENS # Use alphabet set to create initial list of subtokens subtoken_list = reserved_tokens + list(alphabet) max_subtoken_length = 1 # On each iteration, segment all words using the subtokens defined in # subtoken_dict, count how often the resulting subtokens appear, and update # the dictionary with subtokens w/ high enough counts. for i in xrange(num_iterations): logging.info("\tGenerating subtokens: iteration %d", i) # Generate new subtoken->id dictionary using the new subtoken list. subtoken_dict = _list_to_index_dict(subtoken_list) # Create dict mapping subtoken->count, with additional subtokens created # from substrings taken from the tokens. subtoken_counts = _count_and_gen_subtokens(token_counts, alphabet, subtoken_dict, max_subtoken_length) # Generate new list of subtokens sorted by subtoken count. subtoken_list, max_subtoken_length = _gen_new_subtoken_list( subtoken_counts, min_count, alphabet, reserved_tokens) logging.info("\tVocab size: %d", len(subtoken_list)) return subtoken_list
24,517
36.092284
80
py
models
models-master/official/legacy/transformer/utils/metrics.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Functions for calculating loss, accuracy, and other model metrics. Metrics: - Padded loss, accuracy, and negative log perplexity. Source: https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/metrics.py - BLEU approximation. Source: https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/bleu_hook.py - ROUGE score. Source: https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/rouge.py """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import collections import math import numpy as np import six from six.moves import xrange # pylint: disable=redefined-builtin import tensorflow.compat.v1 as tf def _pad_tensors_to_same_length(x, y): """Pad x and y so that the results have the same length (second dimension).""" with tf.name_scope("pad_to_same_length"): x_length = tf.shape(x)[1] y_length = tf.shape(y)[1] max_length = tf.maximum(x_length, y_length) x = tf.pad(x, [[0, 0], [0, max_length - x_length], [0, 0]]) y = tf.pad(y, [[0, 0], [0, max_length - y_length]]) return x, y def padded_cross_entropy_loss(logits, labels, smoothing, vocab_size): """Calculate cross entropy loss while ignoring padding. Args: logits: Tensor of size [batch_size, length_logits, vocab_size] labels: Tensor of size [batch_size, length_labels] smoothing: Label smoothing constant, used to determine the on and off values vocab_size: int size of the vocabulary Returns: Returns the cross entropy loss and weight tensors: float32 tensors with shape [batch_size, max(length_logits, length_labels)] """ with tf.name_scope("loss", values=[logits, labels]): logits, labels = _pad_tensors_to_same_length(logits, labels) # Calculate smoothing cross entropy with tf.name_scope("smoothing_cross_entropy", values=[logits, labels]): confidence = 1.0 - smoothing low_confidence = (1.0 - confidence) / tf.cast(vocab_size - 1, tf.float32) soft_targets = tf.one_hot( tf.cast(labels, tf.int32), depth=vocab_size, on_value=confidence, off_value=low_confidence) xentropy = tf.nn.softmax_cross_entropy_with_logits_v2( logits=logits, labels=soft_targets) # Calculate the best (lowest) possible value of cross entropy, and # subtract from the cross entropy loss. normalizing_constant = -( confidence * tf.log(confidence) + tf.cast(vocab_size - 1, tf.float32) * low_confidence * tf.log(low_confidence + 1e-20)) xentropy -= normalizing_constant weights = tf.cast(tf.not_equal(labels, 0), tf.float32) return xentropy * weights, weights def _convert_to_eval_metric(metric_fn): """Wrap a metric fn that returns scores and weights as an eval metric fn. The input metric_fn returns values for the current batch. The wrapper aggregates the return values collected over all of the batches evaluated. Args: metric_fn: function that returns scores and weights for the current batch's logits and predicted labels. Returns: function that aggregates the scores and weights from metric_fn. """ def problem_metric_fn(*args): """Returns an aggregation of the metric_fn's returned values.""" (scores, weights) = metric_fn(*args) # The tf.metrics.mean function assures correct aggregation. return tf.metrics.mean(scores, weights) return problem_metric_fn def get_eval_metrics(logits, labels, params): """Return dictionary of model evaluation metrics.""" metrics = { "accuracy": _convert_to_eval_metric(padded_accuracy)(logits, labels), "accuracy_top5": _convert_to_eval_metric(padded_accuracy_top5)( logits, labels), "accuracy_per_sequence": _convert_to_eval_metric( padded_sequence_accuracy)(logits, labels), "neg_log_perplexity": _convert_to_eval_metric(padded_neg_log_perplexity)( logits, labels, params["vocab_size"]), } if not params["use_tpu"]: # TPU does not support tf.py_func metrics.update({ "approx_bleu_score": _convert_to_eval_metric( bleu_score)(logits, labels), "rouge_2_fscore": _convert_to_eval_metric( rouge_2_fscore)(logits, labels), "rouge_L_fscore": _convert_to_eval_metric( rouge_l_fscore)(logits, labels), }) # Prefix each of the metric names with "metrics/". This allows the metric # graphs to display under the "metrics" category in TensorBoard. metrics = {"metrics/%s" % k: v for k, v in six.iteritems(metrics)} return metrics def padded_accuracy(logits, labels): """Percentage of times that predictions matches labels on non-0s.""" with tf.variable_scope("padded_accuracy", values=[logits, labels]): logits, labels = _pad_tensors_to_same_length(logits, labels) weights = tf.cast(tf.not_equal(labels, 0), tf.float32) outputs = tf.cast(tf.argmax(logits, axis=-1), tf.int32) padded_labels = tf.cast(labels, tf.int32) return tf.cast(tf.equal(outputs, padded_labels), tf.float32), weights def padded_accuracy_topk(logits, labels, k): """Percentage of times that top-k predictions matches labels on non-0s.""" with tf.variable_scope("padded_accuracy_topk", values=[logits, labels]): logits, labels = _pad_tensors_to_same_length(logits, labels) weights = tf.cast(tf.not_equal(labels, 0), tf.float32) effective_k = tf.minimum(k, tf.shape(logits)[-1]) _, outputs = tf.nn.top_k(logits, k=effective_k) outputs = tf.cast(outputs, tf.int32) padded_labels = tf.cast(labels, tf.int32) padded_labels = tf.expand_dims(padded_labels, axis=-1) padded_labels += tf.zeros_like(outputs) # Pad to same shape. same = tf.cast(tf.equal(outputs, padded_labels), tf.float32) same_topk = tf.reduce_sum(same, axis=-1) return same_topk, weights def padded_accuracy_top5(logits, labels): return padded_accuracy_topk(logits, labels, 5) def padded_sequence_accuracy(logits, labels): """Percentage of times that predictions matches labels everywhere (non-0).""" with tf.variable_scope("padded_sequence_accuracy", values=[logits, labels]): logits, labels = _pad_tensors_to_same_length(logits, labels) weights = tf.cast(tf.not_equal(labels, 0), tf.float32) outputs = tf.cast(tf.argmax(logits, axis=-1), tf.int32) padded_labels = tf.cast(labels, tf.int32) not_correct = (tf.cast(tf.not_equal(outputs, padded_labels), tf.float32) * weights) axis = list(range(1, len(outputs.get_shape()))) correct_seq = 1.0 - tf.minimum(1.0, tf.reduce_sum(not_correct, axis=axis)) return correct_seq, tf.constant(1.0) def padded_neg_log_perplexity(logits, labels, vocab_size): """Average log-perplexity excluding padding 0s. No smoothing.""" num, den = padded_cross_entropy_loss(logits, labels, 0, vocab_size) return -num, den def bleu_score(logits, labels): """Approximate BLEU score computation between labels and predictions. An approximate BLEU scoring method since we do not glue word pieces or decode the ids and tokenize the output. By default, we use ngram order of 4 and use brevity penalty. Also, this does not have beam search. Args: logits: Tensor of size [batch_size, length_logits, vocab_size] labels: Tensor of size [batch-size, length_labels] Returns: bleu: int, approx bleu score """ predictions = tf.cast(tf.argmax(logits, axis=-1), tf.int32) # TODO: Look into removing use of py_func # pylint: disable=g-bad-todo bleu = tf.py_func(compute_bleu, (labels, predictions), tf.float32) return bleu, tf.constant(1.0) def _get_ngrams_with_counter(segment, max_order): """Extracts all n-grams up to a given maximum order from an input segment. Args: segment: text segment from which n-grams will be extracted. max_order: maximum length in tokens of the n-grams returned by this methods. Returns: The Counter containing all n-grams upto max_order in segment with a count of how many times each n-gram occurred. """ ngram_counts = collections.Counter() for order in xrange(1, max_order + 1): for i in xrange(0, len(segment) - order + 1): ngram = tuple(segment[i:i + order]) ngram_counts[ngram] += 1 return ngram_counts def compute_bleu(reference_corpus, translation_corpus, max_order=4, use_bp=True): """Computes BLEU score of translated segments against one or more references. Args: reference_corpus: list of references for each translation. Each reference should be tokenized into a list of tokens. translation_corpus: list of translations to score. Each translation should be tokenized into a list of tokens. max_order: Maximum n-gram order to use when computing BLEU score. use_bp: boolean, whether to apply brevity penalty. Returns: BLEU score. """ reference_length = 0 translation_length = 0 bp = 1.0 geo_mean = 0 matches_by_order = [0] * max_order possible_matches_by_order = [0] * max_order precisions = [] for (references, translations) in zip(reference_corpus, translation_corpus): reference_length += len(references) translation_length += len(translations) ref_ngram_counts = _get_ngrams_with_counter(references, max_order) translation_ngram_counts = _get_ngrams_with_counter(translations, max_order) overlap = dict((ngram, min(count, translation_ngram_counts[ngram])) for ngram, count in ref_ngram_counts.items()) for ngram in overlap: matches_by_order[len(ngram) - 1] += overlap[ngram] for ngram in translation_ngram_counts: possible_matches_by_order[len(ngram) - 1] += translation_ngram_counts[ ngram] precisions = [0] * max_order smooth = 1.0 for i in xrange(0, max_order): if possible_matches_by_order[i] > 0: precisions[i] = float(matches_by_order[i]) / possible_matches_by_order[i] if matches_by_order[i] > 0: precisions[i] = float(matches_by_order[i]) / possible_matches_by_order[ i] else: smooth *= 2 precisions[i] = 1.0 / (smooth * possible_matches_by_order[i]) else: precisions[i] = 0.0 if max(precisions) > 0: p_log_sum = sum(math.log(p) for p in precisions if p) geo_mean = math.exp(p_log_sum / max_order) if use_bp: ratio = translation_length / reference_length bp = math.exp(1 - 1. / ratio) if ratio < 1.0 else 1.0 bleu = geo_mean * bp return np.float32(bleu) def rouge_2_fscore(logits, labels): """ROUGE-2 F1 score computation between labels and predictions. This is an approximate ROUGE scoring method since we do not glue word pieces or decode the ids and tokenize the output. Args: logits: tensor, model predictions labels: tensor, gold output. Returns: rouge2_fscore: approx rouge-2 f1 score. """ predictions = tf.cast(tf.argmax(logits, axis=-1), tf.int32) # TODO: Look into removing use of py_func # pylint: disable=g-bad-todo rouge_2_f_score = tf.py_func(rouge_n, (predictions, labels), tf.float32) return rouge_2_f_score, tf.constant(1.0) def _get_ngrams(n, text): """Calculates n-grams. Args: n: which n-grams to calculate text: An array of tokens Returns: A set of n-grams """ ngram_set = set() text_length = len(text) max_index_ngram_start = text_length - n for i in range(max_index_ngram_start + 1): ngram_set.add(tuple(text[i:i + n])) return ngram_set def rouge_n(eval_sentences, ref_sentences, n=2): """Computes ROUGE-N f1 score of two text collections of sentences. Source: https://www.microsoft.com/en-us/research/publication/ rouge-a-package-for-automatic-evaluation-of-summaries/ Args: eval_sentences: Predicted sentences. ref_sentences: Sentences from the reference set n: Size of ngram. Defaults to 2. Returns: f1 score for ROUGE-N """ f1_scores = [] for eval_sentence, ref_sentence in zip(eval_sentences, ref_sentences): eval_ngrams = _get_ngrams(n, eval_sentence) ref_ngrams = _get_ngrams(n, ref_sentence) ref_count = len(ref_ngrams) eval_count = len(eval_ngrams) # Count the overlapping ngrams between evaluated and reference overlapping_ngrams = eval_ngrams.intersection(ref_ngrams) overlapping_count = len(overlapping_ngrams) # Handle edge case. This isn't mathematically correct, but it's good enough if eval_count == 0: precision = 0.0 else: precision = float(overlapping_count) / eval_count if ref_count == 0: recall = 0.0 else: recall = float(overlapping_count) / ref_count f1_scores.append(2.0 * ((precision * recall) / (precision + recall + 1e-8))) # return overlapping_count / reference_count return np.mean(f1_scores, dtype=np.float32) def rouge_l_fscore(predictions, labels): """ROUGE scores computation between labels and predictions. This is an approximate ROUGE scoring method since we do not glue word pieces or decode the ids and tokenize the output. Args: predictions: tensor, model predictions labels: tensor, gold output. Returns: rouge_l_fscore: approx rouge-l f1 score. """ outputs = tf.cast(tf.argmax(predictions, axis=-1), tf.int32) rouge_l_f_score = tf.py_func(rouge_l_sentence_level, (outputs, labels), tf.float32) return rouge_l_f_score, tf.constant(1.0) def rouge_l_sentence_level(eval_sentences, ref_sentences): """Computes ROUGE-L (sentence level) of two collections of sentences. Source: https://www.microsoft.com/en-us/research/publication/ rouge-a-package-for-automatic-evaluation-of-summaries/ Calculated according to: R_lcs = LCS(X,Y)/m P_lcs = LCS(X,Y)/n F_lcs = ((1 + beta^2)*R_lcs*P_lcs) / (R_lcs + (beta^2) * P_lcs) where: X = reference summary Y = Candidate summary m = length of reference summary n = length of candidate summary Args: eval_sentences: The sentences that have been picked by the summarizer ref_sentences: The sentences from the reference set Returns: A float: F_lcs """ f1_scores = [] for eval_sentence, ref_sentence in zip(eval_sentences, ref_sentences): m = float(len(ref_sentence)) n = float(len(eval_sentence)) lcs = _len_lcs(eval_sentence, ref_sentence) f1_scores.append(_f_lcs(lcs, m, n)) return np.mean(f1_scores, dtype=np.float32) def _len_lcs(x, y): """Returns the length of the Longest Common Subsequence between two seqs. Source: http://www.algorithmist.com/index.php/Longest_Common_Subsequence Args: x: sequence of words y: sequence of words Returns integer: Length of LCS between x and y """ table = _lcs(x, y) n, m = len(x), len(y) return table[n, m] def _lcs(x, y): """Computes the length of the LCS between two seqs. The implementation below uses a DP programming algorithm and runs in O(nm) time where n = len(x) and m = len(y). Source: http://www.algorithmist.com/index.php/Longest_Common_Subsequence Args: x: collection of words y: collection of words Returns: Table of dictionary of coord and len lcs """ n, m = len(x), len(y) table = dict() for i in range(n + 1): for j in range(m + 1): if i == 0 or j == 0: table[i, j] = 0 elif x[i - 1] == y[j - 1]: table[i, j] = table[i - 1, j - 1] + 1 else: table[i, j] = max(table[i - 1, j], table[i, j - 1]) return table def _f_lcs(llcs, m, n): """Computes the LCS-based F-measure score. Source: http://research.microsoft.com/en-us/um/people/cyl/download/papers/ rouge-working-note-v1.3.1.pdf Args: llcs: Length of LCS m: number of words in reference summary n: number of words in candidate summary Returns: Float. LCS-based F-measure score """ r_lcs = llcs / m p_lcs = llcs / n beta = p_lcs / (r_lcs + 1e-12) num = (1 + (beta ** 2)) * r_lcs * p_lcs denom = r_lcs + ((beta ** 2) * p_lcs) f_lcs = num / (denom + 1e-12) return f_lcs
16,706
32.957317
93
py
models
models-master/official/legacy/transformer/utils/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/transformer/utils/tokenizer_test.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Test Subtokenizer and string helper methods.""" import collections import tempfile import tensorflow as tf from official.legacy.transformer.utils import tokenizer class SubtokenizerTest(tf.test.TestCase): def _init_subtokenizer(self, vocab_list): temp_file = tempfile.NamedTemporaryFile(delete=False) with tf.io.gfile.GFile(temp_file.name, "w") as w: for subtoken in vocab_list: w.write("'%s'" % subtoken) w.write("\n") return tokenizer.Subtokenizer(temp_file.name, reserved_tokens=[]) def test_encode(self): vocab_list = ["123_", "test", "ing_"] subtokenizer = self._init_subtokenizer(vocab_list) s = "testing 123" encoded_list = subtokenizer.encode(s) self.assertEqual([1, 2, 0], encoded_list) def test_decode(self): vocab_list = ["123_", "test", "ing_"] subtokenizer = self._init_subtokenizer(vocab_list) encoded_list = [1, 2, 0] # testing 123 decoded_str = subtokenizer.decode(encoded_list) self.assertEqual("testing 123", decoded_str) def test_subtoken_ids_to_tokens(self): vocab_list = ["123_", "test", "ing_"] subtokenizer = self._init_subtokenizer(vocab_list) encoded_list = [1, 2, 0] # testing 123 token_list = subtokenizer._subtoken_ids_to_tokens(encoded_list) self.assertEqual([u"testing", u"123"], token_list) class StringHelperTest(tf.test.TestCase): def test_split_string_to_tokens(self): text = "test? testing 123." tokens = tokenizer._split_string_to_tokens(text, tokenizer._ALPHANUMERIC_CHAR_SET) self.assertEqual(["test", "? ", "testing", "123", "."], tokens) def test_join_tokens_to_string(self): tokens = ["test", "? ", "testing", "123", "."] s = tokenizer._join_tokens_to_string(tokens, tokenizer._ALPHANUMERIC_CHAR_SET) self.assertEqual("test? testing 123.", s) def test_escape_token(self): token = u"abc_\\4" alphabet = set("abc_\\u;") escaped_token = tokenizer._escape_token(token, alphabet) self.assertEqual("abc\\u\\\\\\52;_", escaped_token) def test_unescape_token(self): escaped_token = u"Underline: \\u, Backslash: \\\\, Unicode: \\52;" unescaped_token = tokenizer._unescape_token(escaped_token) self.assertEqual("Underline: _, Backslash: \\, Unicode: 4", unescaped_token) def test_list_to_index_dict(self): lst = ["test", "strings"] d = tokenizer._list_to_index_dict(lst) self.assertDictEqual({"test": 0, "strings": 1}, d) def test_split_token_to_subtokens(self): token = "abc" subtoken_dict = {"a": 0, "b": 1, "c": 2, "ab": 3} max_subtoken_length = 2 subtokens = tokenizer._split_token_to_subtokens(token, subtoken_dict, max_subtoken_length) self.assertEqual(["ab", "c"], subtokens) def test_generate_alphabet_dict(self): s = ["testing", "123"] reserved_tokens = ["???"] alphabet = tokenizer._generate_alphabet_dict(s, reserved_tokens) self.assertIn("?", alphabet) self.assertIn("t", alphabet) self.assertIn("e", alphabet) self.assertIn("s", alphabet) self.assertIn("i", alphabet) self.assertIn("n", alphabet) self.assertIn("g", alphabet) self.assertIn("1", alphabet) self.assertIn("2", alphabet) self.assertIn("3", alphabet) def test_count_and_gen_subtokens(self): token_counts = {"abc": 5} alphabet = set("abc_") subtoken_dict = {"a": 0, "b": 1, "c": 2, "_": 3} max_subtoken_length = 2 subtoken_counts = tokenizer._count_and_gen_subtokens( token_counts, alphabet, subtoken_dict, max_subtoken_length) self.assertIsInstance(subtoken_counts, collections.defaultdict) self.assertDictEqual( { "a": 5, "b": 5, "c": 5, "_": 5, "ab": 5, "bc": 5, "c_": 5, "abc": 5, "bc_": 5, "abc_": 5 }, subtoken_counts) def test_filter_and_bucket_subtokens(self): subtoken_counts = collections.defaultdict(int, { "a": 2, "b": 4, "c": 1, "ab": 6, "ac": 3, "abbc": 5 }) min_count = 3 subtoken_buckets = tokenizer._filter_and_bucket_subtokens( subtoken_counts, min_count) self.assertEqual(len(subtoken_buckets[0]), 0) self.assertEqual(set("b"), subtoken_buckets[1]) self.assertEqual(set(["ab", "ac"]), subtoken_buckets[2]) self.assertEqual(len(subtoken_buckets[3]), 0) self.assertEqual(set(["abbc"]), subtoken_buckets[4]) def test_gen_new_subtoken_list(self): subtoken_counts = collections.defaultdict(int, { "translate": 10, "t": 40, "tr": 16, "tra": 12 }) min_count = 5 alphabet = set("translate") reserved_tokens = ["reserved", "tokens"] subtoken_list, max_token_length = tokenizer._gen_new_subtoken_list( subtoken_counts, min_count, alphabet, reserved_tokens) # Check that "tra" isn"t in the list (its count should be decremented to 2, # so it should not be added to the canddiate list). self.assertNotIn("tra", subtoken_list) self.assertIn("tr", subtoken_list) self.assertIn("t", subtoken_list) self.assertEqual(len("translate"), max_token_length) def test_generate_subtokens(self): token_counts = {"ab": 1, "bc": 3, "abc": 5} alphabet = set("abc_") min_count = 100 num_iterations = 1 reserved_tokens = ["reserved", "tokens"] vocab_list = tokenizer._generate_subtokens(token_counts, alphabet, min_count, num_iterations, reserved_tokens) # Check that reserved tokens are at the front of the list self.assertEqual(vocab_list[:2], reserved_tokens) # Check that each character in alphabet is in the vocab list for c in alphabet: self.assertIn(c, vocab_list) if __name__ == "__main__": tf.test.main()
6,658
31.482927
80
py
models
models-master/official/legacy/detection/main.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Main function to train various object detection models.""" import functools import pprint from absl import app from absl import flags from absl import logging import tensorflow as tf from official.common import distribute_utils from official.legacy.detection.configs import factory as config_factory from official.legacy.detection.dataloader import input_reader from official.legacy.detection.dataloader import mode_keys as ModeKeys from official.legacy.detection.executor import distributed_executor as executor from official.legacy.detection.executor.detection_executor import DetectionDistributedExecutor from official.legacy.detection.modeling import factory as model_factory from official.modeling.hyperparams import params_dict from official.utils import hyperparams_flags from official.utils.flags import core as flags_core from official.utils.misc import keras_utils hyperparams_flags.initialize_common_flags() flags_core.define_log_steps() flags.DEFINE_bool('enable_xla', default=False, help='Enable XLA for GPU') flags.DEFINE_string( 'mode', default='train', help='Mode to run: `train`, `eval` or `eval_once`.') flags.DEFINE_string( 'model', default='retinanet', help='Model to run: `retinanet`, `mask_rcnn` or `shapemask`.') flags.DEFINE_string('training_file_pattern', None, 'Location of the train data.') flags.DEFINE_string('eval_file_pattern', None, 'Location of ther eval data') flags.DEFINE_string( 'checkpoint_path', None, 'The checkpoint path to eval. Only used in eval_once mode.') FLAGS = flags.FLAGS def run_executor(params, mode, checkpoint_path=None, train_input_fn=None, eval_input_fn=None, callbacks=None, prebuilt_strategy=None): """Runs the object detection model on distribution strategy defined by the user.""" if params.architecture.use_bfloat16: tf.compat.v2.keras.mixed_precision.set_global_policy('mixed_bfloat16') model_builder = model_factory.model_generator(params) if prebuilt_strategy is not None: strategy = prebuilt_strategy else: strategy_config = params.strategy_config distribute_utils.configure_cluster(strategy_config.worker_hosts, strategy_config.task_index) strategy = distribute_utils.get_distribution_strategy( distribution_strategy=params.strategy_type, num_gpus=strategy_config.num_gpus, all_reduce_alg=strategy_config.all_reduce_alg, num_packs=strategy_config.num_packs, tpu_address=strategy_config.tpu) num_workers = int(strategy.num_replicas_in_sync + 7) // 8 is_multi_host = (int(num_workers) >= 2) if mode == 'train': def _model_fn(params): return model_builder.build_model(params, mode=ModeKeys.TRAIN) logging.info( 'Train num_replicas_in_sync %d num_workers %d is_multi_host %s', strategy.num_replicas_in_sync, num_workers, is_multi_host) dist_executor = DetectionDistributedExecutor( strategy=strategy, params=params, model_fn=_model_fn, loss_fn=model_builder.build_loss_fn, is_multi_host=is_multi_host, predict_post_process_fn=model_builder.post_processing, trainable_variables_filter=model_builder .make_filter_trainable_variables_fn()) if is_multi_host: train_input_fn = functools.partial( train_input_fn, batch_size=params.train.batch_size // strategy.num_replicas_in_sync) return dist_executor.train( train_input_fn=train_input_fn, model_dir=params.model_dir, iterations_per_loop=params.train.iterations_per_loop, total_steps=params.train.total_steps, init_checkpoint=model_builder.make_restore_checkpoint_fn(), custom_callbacks=callbacks, save_config=True) elif mode == 'eval' or mode == 'eval_once': def _model_fn(params): return model_builder.build_model(params, mode=ModeKeys.PREDICT_WITH_GT) logging.info('Eval num_replicas_in_sync %d num_workers %d is_multi_host %s', strategy.num_replicas_in_sync, num_workers, is_multi_host) if is_multi_host: eval_input_fn = functools.partial( eval_input_fn, batch_size=params.eval.batch_size // strategy.num_replicas_in_sync) dist_executor = DetectionDistributedExecutor( strategy=strategy, params=params, model_fn=_model_fn, loss_fn=model_builder.build_loss_fn, is_multi_host=is_multi_host, predict_post_process_fn=model_builder.post_processing, trainable_variables_filter=model_builder .make_filter_trainable_variables_fn()) if mode == 'eval': results = dist_executor.evaluate_from_model_dir( model_dir=params.model_dir, eval_input_fn=eval_input_fn, eval_metric_fn=model_builder.eval_metrics, eval_timeout=params.eval.eval_timeout, min_eval_interval=params.eval.min_eval_interval, total_steps=params.train.total_steps) else: # Run evaluation once for a single checkpoint. if not checkpoint_path: raise ValueError('checkpoint_path cannot be empty.') if tf.io.gfile.isdir(checkpoint_path): checkpoint_path = tf.train.latest_checkpoint(checkpoint_path) summary_writer = executor.SummaryWriter(params.model_dir, 'eval') results, _ = dist_executor.evaluate_checkpoint( checkpoint_path=checkpoint_path, eval_input_fn=eval_input_fn, eval_metric_fn=model_builder.eval_metrics, summary_writer=summary_writer) for k, v in results.items(): logging.info('Final eval metric %s: %f', k, v) return results else: raise ValueError('Mode not found: %s.' % mode) def run(callbacks=None): """Runs the experiment.""" keras_utils.set_session_config(enable_xla=FLAGS.enable_xla) params = config_factory.config_generator(FLAGS.model) params = params_dict.override_params_dict( params, FLAGS.config_file, is_strict=True) params = params_dict.override_params_dict( params, FLAGS.params_override, is_strict=True) params.override( { 'strategy_type': FLAGS.strategy_type, 'model_dir': FLAGS.model_dir, 'strategy_config': executor.strategy_flags_dict(), }, is_strict=False) # Make sure use_tpu and strategy_type are in sync. params.use_tpu = (params.strategy_type == 'tpu') if not params.use_tpu: params.override({ 'architecture': { 'use_bfloat16': False, }, 'norm_activation': { 'use_sync_bn': False, }, }, is_strict=True) params.validate() params.lock() pp = pprint.PrettyPrinter() params_str = pp.pformat(params.as_dict()) logging.info('Model Parameters: %s', params_str) train_input_fn = None eval_input_fn = None training_file_pattern = FLAGS.training_file_pattern or params.train.train_file_pattern eval_file_pattern = FLAGS.eval_file_pattern or params.eval.eval_file_pattern if not training_file_pattern and not eval_file_pattern: raise ValueError('Must provide at least one of training_file_pattern and ' 'eval_file_pattern.') if training_file_pattern: # Use global batch size for single host. train_input_fn = input_reader.InputFn( file_pattern=training_file_pattern, params=params, mode=input_reader.ModeKeys.TRAIN, batch_size=params.train.batch_size) if eval_file_pattern: eval_input_fn = input_reader.InputFn( file_pattern=eval_file_pattern, params=params, mode=input_reader.ModeKeys.PREDICT_WITH_GT, batch_size=params.eval.batch_size, num_examples=params.eval.eval_samples) if callbacks is None: callbacks = [] if FLAGS.log_steps: callbacks.append( keras_utils.TimeHistory( batch_size=params.train.batch_size, log_steps=FLAGS.log_steps, )) return run_executor( params, FLAGS.mode, checkpoint_path=FLAGS.checkpoint_path, train_input_fn=train_input_fn, eval_input_fn=eval_input_fn, callbacks=callbacks) def main(argv): del argv # Unused. run() if __name__ == '__main__': tf.config.set_soft_device_placement(True) app.run(main)
9,015
33.022642
94
py
models
models-master/official/legacy/detection/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/detection/evaluation/factory.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Evaluator factory.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from official.legacy.detection.evaluation import coco_evaluator def evaluator_generator(params): """Generator function for various evaluators.""" if params.type == 'box': evaluator = coco_evaluator.COCOEvaluator( annotation_file=params.val_json_file, include_mask=False) elif params.type == 'box_and_mask': evaluator = coco_evaluator.COCOEvaluator( annotation_file=params.val_json_file, include_mask=True) elif params.type == 'oln_xclass_box': evaluator = coco_evaluator.OlnXclassEvaluator( annotation_file=params.val_json_file, include_mask=False, use_category=False, seen_class=params.seen_class,) elif params.type == 'oln_xclass_box_and_mask': evaluator = coco_evaluator.OlnXclassEvaluator( annotation_file=params.val_json_file, include_mask=True, use_category=False, seen_class=params.seen_class,) elif params.type == 'oln_xdata_box': evaluator = coco_evaluator.OlnXdataEvaluator( annotation_file=params.val_json_file, include_mask=False, use_category=False, seen_class='all',) elif params.type == 'shapemask_box_and_mask': evaluator = coco_evaluator.ShapeMaskCOCOEvaluator( mask_eval_class=params.mask_eval_class, annotation_file=params.val_json_file, include_mask=True) else: raise ValueError('Evaluator %s is not supported.' % params.type) return coco_evaluator.MetricWrapper(evaluator)
2,169
39.943396
74
py
models
models-master/official/legacy/detection/evaluation/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/detection/evaluation/coco_evaluator.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """The COCO-style evaluator. The following snippet demonstrates the use of interfaces: evaluator = COCOEvaluator(...) for _ in range(num_evals): for _ in range(num_batches_per_eval): predictions, groundtruth = predictor.predict(...) # pop a batch. evaluator.update(predictions, groundtruths) # aggregate internal stats. evaluator.evaluate() # finish one full eval. See also: https://github.com/cocodataset/cocoapi/ """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import atexit import copy import tempfile from absl import logging import numpy as np from pycocotools import cocoeval import six import tensorflow as tf from official.legacy.detection.evaluation import coco_utils from official.legacy.detection.utils import class_utils class OlnCOCOevalWrapper(cocoeval.COCOeval): """COCOeval wrapper class. Rewritten based on cocoapi: (pycocotools/cocoeval.py) This class wraps COCOEVAL API object, which provides the following additional functionalities: 1. summarze 'all', 'seen', and 'novel' split output print-out, e.g., AR at different K proposals, AR and AP resutls for 'seen' and 'novel' class splits. """ def __init__(self, coco_gt, coco_dt, iou_type='box'): super(OlnCOCOevalWrapper, self).__init__( cocoGt=coco_gt, cocoDt=coco_dt, iouType=iou_type) def summarize(self): """Compute and display summary metrics for evaluation results. Delta to the standard cocoapi function: More Averate Recall metrics are produced with different top-K proposals. Note this functin can *only* be applied on the default parameter setting. Raises: Exception: Please run accumulate() first. """ def _summarize(ap=1, iou_thr=None, area_rng='all', max_dets=100): p = self.params i_str = (' {:<18} {} @[ IoU={:<9} | area={:>6s} | maxDets={:>3d} ] = ' '{:0.3f}') title_str = 'Average Precision' if ap == 1 else 'Average Recall' type_str = '(AP)' if ap == 1 else '(AR)' iou_str = '{:0.2f}:{:0.2f}'.format( p.iouThrs[0], p.iouThrs[-1]) if iou_thr is None else '{:0.2f}'.format(iou_thr) aind = [i for i, a_rng in enumerate(p.areaRngLbl) if a_rng == area_rng] mind = [i for i, m_det in enumerate(p.maxDets) if m_det == max_dets] if ap == 1: # dimension of precision: [TxRxKxAxM] s = self.eval['precision'] # IoU if iou_thr is not None: t = np.where(iou_thr == p.iouThrs)[0] s = s[t] s = s[:, :, :, aind, mind] else: # dimension of recall: [TxKxAxM] s = self.eval['recall'] if iou_thr is not None: t = np.where(iou_thr == p.iouThrs)[0] s = s[t] s = s[:, :, aind, mind] if not (s[s > -1]).any(): mean_s = -1 else: mean_s = np.mean(s[s > -1]) print( i_str.format(title_str, type_str, iou_str, area_rng, max_dets, mean_s)) return mean_s def _summarize_dets(): stats = np.zeros((14,)) stats[0] = _summarize(1) stats[1] = _summarize( 1, iou_thr=.5, ) stats[2] = _summarize( 1, iou_thr=.75, ) stats[3] = _summarize( 1, area_rng='small', ) stats[4] = _summarize( 1, area_rng='medium', ) stats[5] = _summarize( 1, area_rng='large', ) stats[6] = _summarize(0, max_dets=self.params.maxDets[0]) # 10 stats[7] = _summarize(0, max_dets=self.params.maxDets[1]) # 20 stats[8] = _summarize(0, max_dets=self.params.maxDets[2]) # 50 stats[9] = _summarize(0, max_dets=self.params.maxDets[3]) # 100 stats[10] = _summarize(0, max_dets=self.params.maxDets[4]) # 200 stats[11] = _summarize(0, area_rng='small', max_dets=10) stats[12] = _summarize(0, area_rng='medium', max_dets=10) stats[13] = _summarize(0, area_rng='large', max_dets=10) return stats if not self.eval: raise Exception('Please run accumulate() first') summarize = _summarize_dets self.stats = summarize() class OlnCOCOevalXclassWrapper(OlnCOCOevalWrapper): """COCOeval wrapper class. Rewritten based on cocoapi: (pycocotools/cocoeval.py) Delta to the standard cocoapi: Detections that hit the 'seen' class objects are ignored in top-K proposals. This class wraps COCOEVAL API object, which provides the following additional functionalities: 1. Include ignore-class split (e.g., 'voc' or 'nonvoc'). 2. Do not count (or ignore) box proposals hitting ignore-class when evaluating Average Recall at top-K proposals. """ def __init__(self, coco_gt, coco_dt, iou_type='box'): super(OlnCOCOevalXclassWrapper, self).__init__( coco_gt=coco_gt, coco_dt=coco_dt, iou_type=iou_type) def evaluateImg(self, img_id, cat_id, a_rng, max_det): p = self.params if p.useCats: gt = self._gts[img_id, cat_id] dt = self._dts[img_id, cat_id] else: gt, dt = [], [] for c_id in p.catIds: gt.extend(self._gts[img_id, c_id]) dt.extend(self._dts[img_id, c_id]) if not gt and not dt: return None for g in gt: if g['ignore'] or (g['area'] < a_rng[0] or g['area'] > a_rng[1]): g['_ignore'] = 1 else: g['_ignore'] = 0 # Class manipulation: ignore the 'ignored_split'. if 'ignored_split' in g and g['ignored_split'] == 1: g['_ignore'] = 1 # sort dt highest score first, sort gt ignore last gtind = np.argsort([g['_ignore'] for g in gt], kind='mergesort') gt = [gt[i] for i in gtind] dtind = np.argsort([-d['score'] for d in dt], kind='mergesort') dt = [dt[i] for i in dtind[0:max_det]] iscrowd = [int(o['iscrowd']) for o in gt] # load computed ious # ious = self.ious[img_id, cat_id][:, gtind] if len( # self.ious[img_id, cat_id]) > 0 else self.ious[img_id, cat_id] if self.ious[img_id, cat_id].any(): ious = self.ious[img_id, cat_id][:, gtind] else: ious = self.ious[img_id, cat_id] tt = len(p.iouThrs) gg = len(gt) dd = len(dt) gtm = np.zeros((tt, gg)) dtm = np.zeros((tt, dd)) gt_ig = np.array([g['_ignore'] for g in gt]) dt_ig = np.zeros((tt, dd)) # indicator of whether the gt object class is of ignored_split or not. gt_ig_split = np.array([g['ignored_split'] for g in gt]) dt_ig_split = np.zeros((dd)) if ious.any(): for tind, t in enumerate(p.iouThrs): for dind, d in enumerate(dt): # information about best match so far (m=-1 -> unmatched) iou = min([t, 1 - 1e-10]) m = -1 for gind, g in enumerate(gt): # if this gt already matched, and not a crowd, continue if gtm[tind, gind] > 0 and not iscrowd[gind]: continue # if dt matched to reg gt, and on ignore gt, stop if m > -1 and gt_ig[m] == 0 and gt_ig[gind] == 1: break # continue to next gt unless better match made if ious[dind, gind] < iou: continue # if match successful and best so far, store appropriately iou = ious[dind, gind] m = gind # if match made store id of match for both dt and gt if m == -1: continue dt_ig[tind, dind] = gt_ig[m] dtm[tind, dind] = gt[m]['id'] gtm[tind, m] = d['id'] # Activate to ignore the seen-class detections. if tind == 0: # Register just only once: tind > 0 is also fine. dt_ig_split[dind] = gt_ig_split[m] # set unmatched detections outside of area range to ignore a = np.array([d['area'] < a_rng[0] or d['area'] > a_rng[1] for d in dt ]).reshape((1, len(dt))) dt_ig = np.logical_or(dt_ig, np.logical_and(dtm == 0, np.repeat(a, tt, 0))) # Activate to ignore the seen-class detections. # Take only eval_split (eg, nonvoc) and ignore seen_split (eg, voc). if dt_ig_split.sum() > 0: dtm = dtm[:, dt_ig_split == 0] dt_ig = dt_ig[:, dt_ig_split == 0] len_dt = min(max_det, len(dt)) dt = [dt[i] for i in range(len_dt) if dt_ig_split[i] == 0] # store results for given image and category return { 'image_id': img_id, 'category_id': cat_id, 'aRng': a_rng, 'maxDet': max_det, 'dtIds': [d['id'] for d in dt], 'gtIds': [g['id'] for g in gt], 'dtMatches': dtm, 'gtMatches': gtm, 'dtScores': [d['score'] for d in dt], 'gtIgnore': gt_ig, 'dtIgnore': dt_ig, } class MetricWrapper(object): """Metric Wrapper of the COCO evaluator.""" # This is only a wrapper for COCO metric and works on for numpy array. So it # doesn't inherit from tf.keras.layers.Layer or tf.keras.metrics.Metric. def __init__(self, evaluator): self._evaluator = evaluator def update_state(self, y_true, y_pred): """Update internal states.""" labels = tf.nest.map_structure(lambda x: x.numpy(), y_true) outputs = tf.nest.map_structure(lambda x: x.numpy(), y_pred) groundtruths = {} predictions = {} for key, val in outputs.items(): if isinstance(val, tuple): val = np.concatenate(val) predictions[key] = val for key, val in labels.items(): if isinstance(val, tuple): val = np.concatenate(val) groundtruths[key] = val self._evaluator.update(predictions, groundtruths) def result(self): return self._evaluator.evaluate() def reset_states(self): return self._evaluator.reset() class COCOEvaluator(object): """COCO evaluation metric class.""" def __init__(self, annotation_file, include_mask, need_rescale_bboxes=True): """Constructs COCO evaluation class. The class provides the interface to metrics_fn in TPUEstimator. The _update_op() takes detections from each image and push them to self.detections. The _evaluate() loads a JSON file in COCO annotation format as the groundtruths and runs COCO evaluation. Args: annotation_file: a JSON file that stores annotations of the eval dataset. If `annotation_file` is None, groundtruth annotations will be loaded from the dataloader. include_mask: a boolean to indicate whether or not to include the mask eval. need_rescale_bboxes: If true bboxes in `predictions` will be rescaled back to absolute values (`image_info` is needed in this case). """ if annotation_file: if annotation_file.startswith('gs://'): _, local_val_json = tempfile.mkstemp(suffix='.json') tf.io.gfile.remove(local_val_json) tf.io.gfile.copy(annotation_file, local_val_json) atexit.register(tf.io.gfile.remove, local_val_json) else: local_val_json = annotation_file self._coco_gt = coco_utils.COCOWrapper( eval_type=('mask' if include_mask else 'box'), annotation_file=local_val_json) self._annotation_file = annotation_file self._include_mask = include_mask self._metric_names = [ 'AP', 'AP50', 'AP75', 'APs', 'APm', 'APl', 'ARmax1', 'ARmax10', 'ARmax100', 'ARs', 'ARm', 'ARl' ] self._required_prediction_fields = [ 'source_id', 'num_detections', 'detection_classes', 'detection_scores', 'detection_boxes' ] self._need_rescale_bboxes = need_rescale_bboxes if self._need_rescale_bboxes: self._required_prediction_fields.append('image_info') self._required_groundtruth_fields = [ 'source_id', 'height', 'width', 'classes', 'boxes' ] if self._include_mask: mask_metric_names = ['mask_' + x for x in self._metric_names] self._metric_names.extend(mask_metric_names) self._required_prediction_fields.extend(['detection_masks']) self._required_groundtruth_fields.extend(['masks']) self.reset() def reset(self): """Resets internal states for a fresh run.""" self._predictions = {} if not self._annotation_file: self._groundtruths = {} def evaluate(self): """Evaluates with detections from all images with COCO API. Returns: coco_metric: float numpy array with shape [24] representing the coco-style evaluation metrics (box and mask). """ if not self._annotation_file: logging.info('Thre is no annotation_file in COCOEvaluator.') gt_dataset = coco_utils.convert_groundtruths_to_coco_dataset( self._groundtruths) coco_gt = coco_utils.COCOWrapper( eval_type=('mask' if self._include_mask else 'box'), gt_dataset=gt_dataset) else: logging.info('Using annotation file: %s', self._annotation_file) coco_gt = self._coco_gt coco_predictions = coco_utils.convert_predictions_to_coco_annotations( self._predictions) coco_dt = coco_gt.loadRes(predictions=coco_predictions) image_ids = [ann['image_id'] for ann in coco_predictions] coco_eval = cocoeval.COCOeval(coco_gt, coco_dt, iouType='bbox') coco_eval.params.imgIds = image_ids coco_eval.evaluate() coco_eval.accumulate() coco_eval.summarize() coco_metrics = coco_eval.stats if self._include_mask: mcoco_eval = cocoeval.COCOeval(coco_gt, coco_dt, iouType='segm') mcoco_eval.params.imgIds = image_ids mcoco_eval.evaluate() mcoco_eval.accumulate() mcoco_eval.summarize() mask_coco_metrics = mcoco_eval.stats if self._include_mask: metrics = np.hstack((coco_metrics, mask_coco_metrics)) else: metrics = coco_metrics # Cleans up the internal variables in order for a fresh eval next time. self.reset() metrics_dict = {} for i, name in enumerate(self._metric_names): metrics_dict[name] = metrics[i].astype(np.float32) return metrics_dict def _process_predictions(self, predictions): image_scale = np.tile(predictions['image_info'][:, 2:3, :], (1, 1, 2)) predictions['detection_boxes'] = ( predictions['detection_boxes'].astype(np.float32)) predictions['detection_boxes'] /= image_scale if 'detection_outer_boxes' in predictions: predictions['detection_outer_boxes'] = ( predictions['detection_outer_boxes'].astype(np.float32)) predictions['detection_outer_boxes'] /= image_scale def update(self, predictions, groundtruths=None): """Update and aggregate detection results and groundtruth data. Args: predictions: a dictionary of numpy arrays including the fields below. See different parsers under `../dataloader` for more details. Required fields: - source_id: a numpy array of int or string of shape [batch_size]. - image_info [if `need_rescale_bboxes` is True]: a numpy array of float of shape [batch_size, 4, 2]. - num_detections: a numpy array of int of shape [batch_size]. - detection_boxes: a numpy array of float of shape [batch_size, K, 4]. - detection_classes: a numpy array of int of shape [batch_size, K]. - detection_scores: a numpy array of float of shape [batch_size, K]. Optional fields: - detection_masks: a numpy array of float of shape [batch_size, K, mask_height, mask_width]. groundtruths: a dictionary of numpy arrays including the fields below. See also different parsers under `../dataloader` for more details. Required fields: - source_id: a numpy array of int or string of shape [batch_size]. - height: a numpy array of int of shape [batch_size]. - width: a numpy array of int of shape [batch_size]. - num_detections: a numpy array of int of shape [batch_size]. - boxes: a numpy array of float of shape [batch_size, K, 4]. - classes: a numpy array of int of shape [batch_size, K]. Optional fields: - is_crowds: a numpy array of int of shape [batch_size, K]. If the field is absent, it is assumed that this instance is not crowd. - areas: a numy array of float of shape [batch_size, K]. If the field is absent, the area is calculated using either boxes or masks depending on which one is available. - masks: a numpy array of float of shape [batch_size, K, mask_height, mask_width], Raises: ValueError: if the required prediction or groundtruth fields are not present in the incoming `predictions` or `groundtruths`. """ for k in self._required_prediction_fields: if k not in predictions: raise ValueError( 'Missing the required key `{}` in predictions!'.format(k)) if self._need_rescale_bboxes: self._process_predictions(predictions) for k, v in six.iteritems(predictions): if k not in self._predictions: self._predictions[k] = [v] else: self._predictions[k].append(v) if not self._annotation_file: assert groundtruths for k in self._required_groundtruth_fields: if k not in groundtruths: raise ValueError( 'Missing the required key `{}` in groundtruths!'.format(k)) for k, v in six.iteritems(groundtruths): if k not in self._groundtruths: self._groundtruths[k] = [v] else: self._groundtruths[k].append(v) class OlnXclassEvaluator(COCOEvaluator): """COCO evaluation metric class.""" def __init__(self, annotation_file, include_mask, need_rescale_bboxes=True, use_category=True, seen_class='all'): """Constructs COCO evaluation class. The class provides the interface to metrics_fn in TPUEstimator. The _update_op() takes detections from each image and push them to self.detections. The _evaluate() loads a JSON file in COCO annotation format as the groundtruths and runs COCO evaluation. Args: annotation_file: a JSON file that stores annotations of the eval dataset. If `annotation_file` is None, groundtruth annotations will be loaded from the dataloader. include_mask: a boolean to indicate whether or not to include the mask eval. need_rescale_bboxes: If true bboxes in `predictions` will be rescaled back to absolute values (`image_info` is needed in this case). use_category: if `False`, treat all object in all classes in one foreground category. seen_class: 'all' or 'voc' or 'nonvoc' """ super(OlnXclassEvaluator, self).__init__( annotation_file=annotation_file, include_mask=include_mask, need_rescale_bboxes=need_rescale_bboxes) self._use_category = use_category self._seen_class = seen_class self._seen_class_ids = class_utils.coco_split_class_ids(seen_class) self._metric_names = [ 'AP', 'AP50', 'AP75', 'APs', 'APm', 'APl', 'ARmax10', 'ARmax20', 'ARmax50', 'ARmax100', 'ARmax200', 'ARmax10s', 'ARmax10m', 'ARmax10l' ] if self._seen_class != 'all': self._metric_names.extend([ 'AP_seen', 'AP50_seen', 'AP75_seen', 'APs_seen', 'APm_seen', 'APl_seen', 'ARmax10_seen', 'ARmax20_seen', 'ARmax50_seen', 'ARmax100_seen', 'ARmax200_seen', 'ARmax10s_seen', 'ARmax10m_seen', 'ARmax10l_seen', 'AP_novel', 'AP50_novel', 'AP75_novel', 'APs_novel', 'APm_novel', 'APl_novel', 'ARmax10_novel', 'ARmax20_novel', 'ARmax50_novel', 'ARmax100_novel', 'ARmax200_novel', 'ARmax10s_novel', 'ARmax10m_novel', 'ARmax10l_novel', ]) if self._include_mask: mask_metric_names = ['mask_' + x for x in self._metric_names] self._metric_names.extend(mask_metric_names) self._required_prediction_fields.extend(['detection_masks']) self._required_groundtruth_fields.extend(['masks']) self.reset() def evaluate(self): """Evaluates with detections from all images with COCO API. Returns: coco_metric: float numpy array with shape [24] representing the coco-style evaluation metrics (box and mask). """ if not self._annotation_file: logging.info('Thre is no annotation_file in COCOEvaluator.') gt_dataset = coco_utils.convert_groundtruths_to_coco_dataset( self._groundtruths) coco_gt = coco_utils.COCOWrapper( eval_type=('mask' if self._include_mask else 'box'), gt_dataset=gt_dataset) else: logging.info('Using annotation file: %s', self._annotation_file) coco_gt = self._coco_gt coco_predictions = coco_utils.convert_predictions_to_coco_annotations( self._predictions) coco_dt = coco_gt.loadRes(predictions=coco_predictions) image_ids = [ann['image_id'] for ann in coco_predictions] # Class manipulation: 'all' split samples -> ignored_split = 0. for idx, ann in enumerate(coco_gt.dataset['annotations']): coco_gt.dataset['annotations'][idx]['ignored_split'] = 0 coco_eval = cocoeval.OlnCOCOevalXclassWrapper( coco_gt, coco_dt, iou_type='bbox') coco_eval.params.maxDets = [10, 20, 50, 100, 200] coco_eval.params.imgIds = image_ids coco_eval.params.useCats = 0 if not self._use_category else 1 coco_eval.evaluate() coco_eval.accumulate() coco_eval.summarize() coco_metrics = coco_eval.stats if self._include_mask: mcoco_eval = cocoeval.OlnCOCOevalXclassWrapper( coco_gt, coco_dt, iou_type='segm') mcoco_eval.params.maxDets = [10, 20, 50, 100, 200] mcoco_eval.params.imgIds = image_ids mcoco_eval.params.useCats = 0 if not self._use_category else 1 mcoco_eval.evaluate() mcoco_eval.accumulate() mcoco_eval.summarize() mask_coco_metrics = mcoco_eval.stats if self._include_mask: metrics = np.hstack((coco_metrics, mask_coco_metrics)) else: metrics = coco_metrics if self._seen_class != 'all': # for seen class eval, samples of novel_class are ignored. coco_gt_seen = copy.deepcopy(coco_gt) for idx, ann in enumerate(coco_gt.dataset['annotations']): if ann['category_id'] in self._seen_class_ids: coco_gt_seen.dataset['annotations'][idx]['ignored_split'] = 0 else: coco_gt_seen.dataset['annotations'][idx]['ignored_split'] = 1 coco_eval_seen = cocoeval.OlnCOCOevalXclassWrapper( coco_gt_seen, coco_dt, iou_type='bbox') coco_eval_seen.params.maxDets = [10, 20, 50, 100, 200] coco_eval_seen.params.imgIds = image_ids coco_eval_seen.params.useCats = 0 if not self._use_category else 1 coco_eval_seen.evaluate() coco_eval_seen.accumulate() coco_eval_seen.summarize() coco_metrics_seen = coco_eval_seen.stats if self._include_mask: mcoco_eval_seen = cocoeval.OlnCOCOevalXclassWrapper( coco_gt_seen, coco_dt, iou_type='segm') mcoco_eval_seen.params.maxDets = [10, 20, 50, 100, 200] mcoco_eval_seen.params.imgIds = image_ids mcoco_eval_seen.params.useCats = 0 if not self._use_category else 1 mcoco_eval_seen.evaluate() mcoco_eval_seen.accumulate() mcoco_eval_seen.summarize() mask_coco_metrics_seen = mcoco_eval_seen.stats # for novel class eval, samples of seen_class are ignored. coco_gt_novel = copy.deepcopy(coco_gt) for idx, ann in enumerate(coco_gt.dataset['annotations']): if ann['category_id'] in self._seen_class_ids: coco_gt_novel.dataset['annotations'][idx]['ignored_split'] = 1 else: coco_gt_novel.dataset['annotations'][idx]['ignored_split'] = 0 coco_eval_novel = cocoeval.OlnCOCOevalXclassWrapper( coco_gt_novel, coco_dt, iou_type='bbox') coco_eval_novel.params.maxDets = [10, 20, 50, 100, 200] coco_eval_novel.params.imgIds = image_ids coco_eval_novel.params.useCats = 0 if not self._use_category else 1 coco_eval_novel.evaluate() coco_eval_novel.accumulate() coco_eval_novel.summarize() coco_metrics_novel = coco_eval_novel.stats if self._include_mask: mcoco_eval_novel = cocoeval.OlnCOCOevalXclassWrapper( coco_gt_novel, coco_dt, iou_type='segm') mcoco_eval_novel.params.maxDets = [10, 20, 50, 100, 200] mcoco_eval_novel.params.imgIds = image_ids mcoco_eval_novel.params.useCats = 0 if not self._use_category else 1 mcoco_eval_novel.evaluate() mcoco_eval_novel.accumulate() mcoco_eval_novel.summarize() mask_coco_metrics_novel = mcoco_eval_novel.stats # Combine all splits. if self._include_mask: metrics = np.hstack(( coco_metrics, coco_metrics_seen, coco_metrics_novel, mask_coco_metrics, mask_coco_metrics_seen, mask_coco_metrics_novel)) else: metrics = np.hstack(( coco_metrics, coco_metrics_seen, coco_metrics_novel)) # Cleans up the internal variables in order for a fresh eval next time. self.reset() metrics_dict = {} for i, name in enumerate(self._metric_names): metrics_dict[name] = metrics[i].astype(np.float32) return metrics_dict class OlnXdataEvaluator(OlnXclassEvaluator): """COCO evaluation metric class.""" def __init__(self, annotation_file, include_mask, need_rescale_bboxes=True, use_category=True, seen_class='all'): """Constructs COCO evaluation class. The class provides the interface to metrics_fn in TPUEstimator. The _update_op() takes detections from each image and push them to self.detections. The _evaluate() loads a JSON file in COCO annotation format as the groundtruths and runs COCO evaluation. Args: annotation_file: a JSON file that stores annotations of the eval dataset. If `annotation_file` is None, groundtruth annotations will be loaded from the dataloader. include_mask: a boolean to indicate whether or not to include the mask eval. need_rescale_bboxes: If true bboxes in `predictions` will be rescaled back to absolute values (`image_info` is needed in this case). use_category: if `False`, treat all object in all classes in one foreground category. seen_class: 'all' or 'voc' or 'nonvoc' """ super(OlnXdataEvaluator, self).__init__( annotation_file=annotation_file, include_mask=include_mask, need_rescale_bboxes=need_rescale_bboxes, use_category=False, seen_class='all') def evaluate(self): """Evaluates with detections from all images with COCO API. Returns: coco_metric: float numpy array with shape [24] representing the coco-style evaluation metrics (box and mask). """ if not self._annotation_file: logging.info('Thre is no annotation_file in COCOEvaluator.') gt_dataset = coco_utils.convert_groundtruths_to_coco_dataset( self._groundtruths) coco_gt = coco_utils.COCOWrapper( eval_type=('mask' if self._include_mask else 'box'), gt_dataset=gt_dataset) else: logging.info('Using annotation file: %s', self._annotation_file) coco_gt = self._coco_gt coco_predictions = coco_utils.convert_predictions_to_coco_annotations( self._predictions) coco_dt = coco_gt.loadRes(predictions=coco_predictions) image_ids = [ann['image_id'] for ann in coco_predictions] # Class manipulation: 'all' split samples -> ignored_split = 0. for idx, _ in enumerate(coco_gt.dataset['annotations']): coco_gt.dataset['annotations'][idx]['ignored_split'] = 0 coco_eval = cocoeval.OlnCOCOevalWrapper(coco_gt, coco_dt, iou_type='bbox') coco_eval.params.maxDets = [10, 20, 50, 100, 200] coco_eval.params.imgIds = image_ids coco_eval.params.useCats = 0 if not self._use_category else 1 coco_eval.evaluate() coco_eval.accumulate() coco_eval.summarize() coco_metrics = coco_eval.stats if self._include_mask: mcoco_eval = cocoeval.OlnCOCOevalWrapper(coco_gt, coco_dt, iou_type='segm') mcoco_eval.params.maxDets = [10, 20, 50, 100, 200] mcoco_eval.params.imgIds = image_ids mcoco_eval.params.useCats = 0 if not self._use_category else 1 mcoco_eval.evaluate() mcoco_eval.accumulate() mcoco_eval.summarize() mask_coco_metrics = mcoco_eval.stats if self._include_mask: metrics = np.hstack((coco_metrics, mask_coco_metrics)) else: metrics = coco_metrics # Cleans up the internal variables in order for a fresh eval next time. self.reset() metrics_dict = {} for i, name in enumerate(self._metric_names): metrics_dict[name] = metrics[i].astype(np.float32) return metrics_dict class ShapeMaskCOCOEvaluator(COCOEvaluator): """COCO evaluation metric class for ShapeMask.""" def __init__(self, mask_eval_class, **kwargs): """Constructs COCO evaluation class. The class provides the interface to metrics_fn in TPUEstimator. The _update_op() takes detections from each image and push them to self.detections. The _evaluate() loads a JSON file in COCO annotation format as the groundtruths and runs COCO evaluation. Args: mask_eval_class: the set of classes for mask evaluation. **kwargs: other keyword arguments passed to the parent class initializer. """ super(ShapeMaskCOCOEvaluator, self).__init__(**kwargs) self._mask_eval_class = mask_eval_class self._eval_categories = class_utils.coco_split_class_ids(mask_eval_class) if mask_eval_class != 'all': self._metric_names = [ x.replace('mask', 'novel_mask') for x in self._metric_names ] def evaluate(self): """Evaluates with detections from all images with COCO API. Returns: coco_metric: float numpy array with shape [24] representing the coco-style evaluation metrics (box and mask). """ if not self._annotation_file: gt_dataset = coco_utils.convert_groundtruths_to_coco_dataset( self._groundtruths) coco_gt = coco_utils.COCOWrapper( eval_type=('mask' if self._include_mask else 'box'), gt_dataset=gt_dataset) else: coco_gt = self._coco_gt coco_predictions = coco_utils.convert_predictions_to_coco_annotations( self._predictions) coco_dt = coco_gt.loadRes(predictions=coco_predictions) image_ids = [ann['image_id'] for ann in coco_predictions] coco_eval = cocoeval.COCOeval(coco_gt, coco_dt, iouType='bbox') coco_eval.params.imgIds = image_ids coco_eval.evaluate() coco_eval.accumulate() coco_eval.summarize() coco_metrics = coco_eval.stats if self._include_mask: mcoco_eval = cocoeval.COCOeval(coco_gt, coco_dt, iouType='segm') mcoco_eval.params.imgIds = image_ids mcoco_eval.evaluate() mcoco_eval.accumulate() mcoco_eval.summarize() if self._mask_eval_class == 'all': metrics = np.hstack((coco_metrics, mcoco_eval.stats)) else: mask_coco_metrics = mcoco_eval.category_stats val_catg_idx = np.isin(mcoco_eval.params.catIds, self._eval_categories) # Gather the valid evaluation of the eval categories. if np.any(val_catg_idx): mean_val_metrics = [] for mid in range(len(self._metric_names) // 2): mean_val_metrics.append( np.nanmean(mask_coco_metrics[mid][val_catg_idx])) mean_val_metrics = np.array(mean_val_metrics) else: mean_val_metrics = np.zeros(len(self._metric_names) // 2) metrics = np.hstack((coco_metrics, mean_val_metrics)) else: metrics = coco_metrics # Cleans up the internal variables in order for a fresh eval next time. self.reset() metrics_dict = {} for i, name in enumerate(self._metric_names): metrics_dict[name] = metrics[i].astype(np.float32) return metrics_dict
32,813
37.695755
80
py
models
models-master/official/legacy/detection/evaluation/coco_utils.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Util functions related to pycocotools and COCO eval.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import copy import json from absl import logging import numpy as np from PIL import Image from pycocotools import coco from pycocotools import mask as mask_api import six import tensorflow as tf from official.legacy.detection.dataloader import tf_example_decoder from official.legacy.detection.utils import box_utils from official.legacy.detection.utils import mask_utils class COCOWrapper(coco.COCO): """COCO wrapper class. This class wraps COCO API object, which provides the following additional functionalities: 1. Support string type image id. 2. Support loading the groundtruth dataset using the external annotation dictionary. 3. Support loading the prediction results using the external annotation dictionary. """ def __init__(self, eval_type='box', annotation_file=None, gt_dataset=None): """Instantiates a COCO-style API object. Args: eval_type: either 'box' or 'mask'. annotation_file: a JSON file that stores annotations of the eval dataset. This is required if `gt_dataset` is not provided. gt_dataset: the groundtruth eval datatset in COCO API format. """ if ((annotation_file and gt_dataset) or ((not annotation_file) and (not gt_dataset))): raise ValueError('One and only one of `annotation_file` and `gt_dataset` ' 'needs to be specified.') if eval_type not in ['box', 'mask']: raise ValueError('The `eval_type` can only be either `box` or `mask`.') coco.COCO.__init__(self, annotation_file=annotation_file) self._eval_type = eval_type if gt_dataset: self.dataset = gt_dataset self.createIndex() def loadRes(self, predictions): """Loads result file and return a result api object. Args: predictions: a list of dictionary each representing an annotation in COCO format. The required fields are `image_id`, `category_id`, `score`, `bbox`, `segmentation`. Returns: res: result COCO api object. Raises: ValueError: if the set of image id from predctions is not the subset of the set of image id of the groundtruth dataset. """ res = coco.COCO() res.dataset['images'] = copy.deepcopy(self.dataset['images']) res.dataset['categories'] = copy.deepcopy(self.dataset['categories']) image_ids = [ann['image_id'] for ann in predictions] if set(image_ids) != (set(image_ids) & set(self.getImgIds())): raise ValueError('Results do not correspond to the current dataset!') for ann in predictions: x1, x2, y1, y2 = [ann['bbox'][0], ann['bbox'][0] + ann['bbox'][2], ann['bbox'][1], ann['bbox'][1] + ann['bbox'][3]] if self._eval_type == 'box': ann['area'] = ann['bbox'][2] * ann['bbox'][3] ann['segmentation'] = [ [x1, y1, x1, y2, x2, y2, x2, y1]] elif self._eval_type == 'mask': ann['area'] = mask_api.area(ann['segmentation']) res.dataset['annotations'] = copy.deepcopy(predictions) res.createIndex() return res def convert_predictions_to_coco_annotations(predictions): """Converts a batch of predictions to annotations in COCO format. Args: predictions: a dictionary of lists of numpy arrays including the following fields. K below denotes the maximum number of instances per image. Required fields: - source_id: a list of numpy arrays of int or string of shape [batch_size]. - num_detections: a list of numpy arrays of int of shape [batch_size]. - detection_boxes: a list of numpy arrays of float of shape [batch_size, K, 4], where coordinates are in the original image space (not the scaled image space). - detection_classes: a list of numpy arrays of int of shape [batch_size, K]. - detection_scores: a list of numpy arrays of float of shape [batch_size, K]. Optional fields: - detection_masks: a list of numpy arrays of float of shape [batch_size, K, mask_height, mask_width]. Returns: coco_predictions: prediction in COCO annotation format. """ coco_predictions = [] num_batches = len(predictions['source_id']) batch_size = predictions['source_id'][0].shape[0] max_num_detections = predictions['detection_classes'][0].shape[1] use_outer_box = 'detection_outer_boxes' in predictions for i in range(num_batches): predictions['detection_boxes'][i] = box_utils.yxyx_to_xywh( predictions['detection_boxes'][i]) if use_outer_box: predictions['detection_outer_boxes'][i] = box_utils.yxyx_to_xywh( predictions['detection_outer_boxes'][i]) mask_boxes = predictions['detection_outer_boxes'] else: mask_boxes = predictions['detection_boxes'] for j in range(batch_size): if 'detection_masks' in predictions: image_masks = mask_utils.paste_instance_masks( predictions['detection_masks'][i][j], mask_boxes[i][j], int(predictions['image_info'][i][j, 0, 0]), int(predictions['image_info'][i][j, 0, 1])) binary_masks = (image_masks > 0.0).astype(np.uint8) encoded_masks = [ mask_api.encode(np.asfortranarray(binary_mask)) for binary_mask in list(binary_masks)] for k in range(max_num_detections): ann = {} ann['image_id'] = predictions['source_id'][i][j] ann['category_id'] = predictions['detection_classes'][i][j, k] ann['bbox'] = predictions['detection_boxes'][i][j, k] ann['score'] = predictions['detection_scores'][i][j, k] if 'detection_masks' in predictions: ann['segmentation'] = encoded_masks[k] coco_predictions.append(ann) for i, ann in enumerate(coco_predictions): ann['id'] = i + 1 return coco_predictions def convert_groundtruths_to_coco_dataset(groundtruths, label_map=None): """Converts groundtruths to the dataset in COCO format. Args: groundtruths: a dictionary of numpy arrays including the fields below. Note that each element in the list represent the number for a single example without batch dimension. K below denotes the actual number of instances for each image. Required fields: - source_id: a list of numpy arrays of int or string of shape [batch_size]. - height: a list of numpy arrays of int of shape [batch_size]. - width: a list of numpy arrays of int of shape [batch_size]. - num_detections: a list of numpy arrays of int of shape [batch_size]. - boxes: a list of numpy arrays of float of shape [batch_size, K, 4], where coordinates are in the original image space (not the normalized coordinates). - classes: a list of numpy arrays of int of shape [batch_size, K]. Optional fields: - is_crowds: a list of numpy arrays of int of shape [batch_size, K]. If th field is absent, it is assumed that this instance is not crowd. - areas: a list of numy arrays of float of shape [batch_size, K]. If the field is absent, the area is calculated using either boxes or masks depending on which one is available. - masks: a list of numpy arrays of string of shape [batch_size, K], label_map: (optional) a dictionary that defines items from the category id to the category name. If `None`, collect the category mappping from the `groundtruths`. Returns: coco_groundtruths: the groundtruth dataset in COCO format. """ source_ids = np.concatenate(groundtruths['source_id'], axis=0) heights = np.concatenate(groundtruths['height'], axis=0) widths = np.concatenate(groundtruths['width'], axis=0) gt_images = [{'id': int(i), 'height': int(h), 'width': int(w)} for i, h, w in zip(source_ids, heights, widths)] gt_annotations = [] num_batches = len(groundtruths['source_id']) batch_size = groundtruths['source_id'][0].shape[0] for i in range(num_batches): for j in range(batch_size): num_instances = groundtruths['num_detections'][i][j] for k in range(num_instances): ann = {} ann['image_id'] = int(groundtruths['source_id'][i][j]) if 'is_crowds' in groundtruths: ann['iscrowd'] = int(groundtruths['is_crowds'][i][j, k]) else: ann['iscrowd'] = 0 ann['category_id'] = int(groundtruths['classes'][i][j, k]) boxes = groundtruths['boxes'][i] ann['bbox'] = [ float(boxes[j, k, 1]), float(boxes[j, k, 0]), float(boxes[j, k, 3] - boxes[j, k, 1]), float(boxes[j, k, 2] - boxes[j, k, 0])] if 'areas' in groundtruths: ann['area'] = float(groundtruths['areas'][i][j, k]) else: ann['area'] = float( (boxes[j, k, 3] - boxes[j, k, 1]) * (boxes[j, k, 2] - boxes[j, k, 0])) if 'masks' in groundtruths: mask = Image.open(six.BytesIO(groundtruths['masks'][i][j, k])) np_mask = np.array(mask, dtype=np.uint8) np_mask[np_mask > 0] = 255 encoded_mask = mask_api.encode(np.asfortranarray(np_mask)) ann['segmentation'] = encoded_mask if 'areas' not in groundtruths: ann['area'] = mask_api.area(encoded_mask) gt_annotations.append(ann) for i, ann in enumerate(gt_annotations): ann['id'] = i + 1 if label_map: gt_categories = [{'id': i, 'name': label_map[i]} for i in label_map] else: category_ids = [gt['category_id'] for gt in gt_annotations] gt_categories = [{'id': i} for i in set(category_ids)] gt_dataset = { 'images': gt_images, 'categories': gt_categories, 'annotations': copy.deepcopy(gt_annotations), } return gt_dataset class COCOGroundtruthGenerator(object): """Generates the groundtruth annotations from a single example.""" def __init__(self, file_pattern, num_examples, include_mask): self._file_pattern = file_pattern self._num_examples = num_examples self._include_mask = include_mask self._dataset_fn = tf.data.TFRecordDataset def _parse_single_example(self, example): """Parses a single serialized tf.Example proto. Args: example: a serialized tf.Example proto string. Returns: A dictionary of groundtruth with the following fields: source_id: a scalar tensor of int64 representing the image source_id. height: a scalar tensor of int64 representing the image height. width: a scalar tensor of int64 representing the image width. boxes: a float tensor of shape [K, 4], representing the groundtruth boxes in absolute coordinates with respect to the original image size. classes: a int64 tensor of shape [K], representing the class labels of each instances. is_crowds: a bool tensor of shape [K], indicating whether the instance is crowd. areas: a float tensor of shape [K], indicating the area of each instance. masks: a string tensor of shape [K], containing the bytes of the png mask of each instance. """ decoder = tf_example_decoder.TfExampleDecoder( include_mask=self._include_mask) decoded_tensors = decoder.decode(example) image = decoded_tensors['image'] image_size = tf.shape(image)[0:2] boxes = box_utils.denormalize_boxes( decoded_tensors['groundtruth_boxes'], image_size) groundtruths = { 'source_id': tf.string_to_number( decoded_tensors['source_id'], out_type=tf.int64), 'height': decoded_tensors['height'], 'width': decoded_tensors['width'], 'num_detections': tf.shape(decoded_tensors['groundtruth_classes'])[0], 'boxes': boxes, 'classes': decoded_tensors['groundtruth_classes'], 'is_crowds': decoded_tensors['groundtruth_is_crowd'], 'areas': decoded_tensors['groundtruth_area'], } if self._include_mask: groundtruths.update({ 'masks': decoded_tensors['groundtruth_instance_masks_png'], }) return groundtruths def _build_pipeline(self): """Builds data pipeline to generate groundtruth annotations.""" dataset = tf.data.Dataset.list_files(self._file_pattern, shuffle=False) dataset = dataset.apply( tf.data.experimental.parallel_interleave( lambda filename: self._dataset_fn(filename).prefetch(1), cycle_length=32, sloppy=False)) dataset = dataset.map(self._parse_single_example, num_parallel_calls=64) dataset = dataset.prefetch(tf.data.experimental.AUTOTUNE) dataset = dataset.batch(1, drop_remainder=False) return dataset def __call__(self): with tf.Graph().as_default(): dataset = self._build_pipeline() groundtruth = dataset.make_one_shot_iterator().get_next() with tf.Session() as sess: for _ in range(self._num_examples): groundtruth_result = sess.run(groundtruth) yield groundtruth_result def scan_and_generator_annotation_file(file_pattern, num_samples, include_mask, annotation_file): """Scans and generate the COCO-style annotation JSON file given a dataset.""" groundtruth_generator = COCOGroundtruthGenerator( file_pattern, num_samples, include_mask) generate_annotation_file(groundtruth_generator, annotation_file) def generate_annotation_file(groundtruth_generator, annotation_file): """Generates COCO-style annotation JSON file given a groundtruth generator.""" groundtruths = {} logging.info('Loading groundtruth annotations from dataset to memory...') for groundtruth in groundtruth_generator(): for k, v in six.iteritems(groundtruth): if k not in groundtruths: groundtruths[k] = [v] else: groundtruths[k].append(v) gt_dataset = convert_groundtruths_to_coco_dataset(groundtruths) logging.info('Saving groundtruth annotations to the JSON file...') with tf.io.gfile.GFile(annotation_file, 'w') as f: f.write(json.dumps(gt_dataset)) logging.info('Done saving the JSON file...')
15,082
39.436997
80
py
models
models-master/official/legacy/detection/configs/olnmask_config.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Config template to train Object Localization Network (OLN).""" from official.legacy.detection.configs import base_config from official.modeling.hyperparams import params_dict # pylint: disable=line-too-long OLNMASK_CFG = params_dict.ParamsDict(base_config.BASE_CFG) OLNMASK_CFG.override({ 'type': 'olnmask', 'eval': { 'type': 'oln_xclass_box', 'use_category': False, 'seen_class': 'voc', 'num_images_to_visualize': 0, }, 'architecture': { 'parser': 'olnmask_parser', 'min_level': 2, 'max_level': 6, 'include_rpn_class': False, 'include_frcnn_class': False, 'include_frcnn_box': True, 'include_mask': False, 'mask_target_size': 28, 'num_classes': 2, }, 'olnmask_parser': { 'output_size': [640, 640], 'num_channels': 3, 'rpn_match_threshold': 0.7, 'rpn_unmatched_threshold': 0.3, 'rpn_batch_size_per_im': 256, 'rpn_fg_fraction': 0.5, 'aug_rand_hflip': True, 'aug_scale_min': 0.5, 'aug_scale_max': 2.0, 'skip_crowd_during_training': True, 'max_num_instances': 100, 'mask_crop_size': 112, # centerness targets. 'has_centerness': True, 'rpn_center_match_iou_threshold': 0.3, 'rpn_center_unmatched_iou_threshold': 0.1, 'rpn_num_center_samples_per_im': 256, # class manipulation. 'class_agnostic': True, 'train_class': 'voc', }, 'anchor': { 'num_scales': 1, 'aspect_ratios': [1.0], 'anchor_size': 8, }, 'rpn_head': { 'num_convs': 2, 'num_filters': 256, 'use_separable_conv': False, 'use_batch_norm': False, # RPN-Centerness learning { 'has_centerness': True, # } }, 'frcnn_head': { 'num_convs': 0, 'num_filters': 256, 'use_separable_conv': False, 'num_fcs': 2, 'fc_dims': 1024, 'use_batch_norm': False, 'has_scoring': True, }, 'mrcnn_head': { 'num_convs': 4, 'num_filters': 256, 'use_separable_conv': False, 'use_batch_norm': False, 'has_scoring': False, }, 'rpn_score_loss': { 'rpn_batch_size_per_im': 256, }, 'rpn_box_loss': { 'huber_loss_delta': 1.0 / 9.0, }, 'frcnn_box_loss': { 'huber_loss_delta': 1.0, }, 'frcnn_box_score_loss': { 'ignore_threshold': 0.3, }, 'roi_proposal': { 'rpn_pre_nms_top_k': 2000, 'rpn_post_nms_top_k': 2000, 'rpn_nms_threshold': 0.7, 'rpn_score_threshold': 0.0, 'rpn_min_size_threshold': 0.0, 'test_rpn_pre_nms_top_k': 2000, 'test_rpn_post_nms_top_k': 2000, 'test_rpn_nms_threshold': 0.7, 'test_rpn_score_threshold': 0.0, 'test_rpn_min_size_threshold': 0.0, 'use_batched_nms': False, }, 'roi_sampling': { 'num_samples_per_image': 512, 'fg_fraction': 0.25, 'fg_iou_thresh': 0.5, 'bg_iou_thresh_hi': 0.5, 'bg_iou_thresh_lo': 0.0, 'mix_gt_boxes': True, }, 'mask_sampling': { 'num_mask_samples_per_image': 128, # Typically = `num_samples_per_image` * `fg_fraction`. }, 'postprocess': { 'use_batched_nms': False, 'max_total_size': 100, 'nms_iou_threshold': 0.5, 'score_threshold': 0.00, 'pre_nms_num_boxes': 2000, }, }, is_strict=False) OLNMASK_RESTRICTIONS = [ # 'anchor.aspect_ratios == [1.0]', # 'anchor.scales == 1', ] # pylint: enable=line-too-long
4,271
28.666667
98
py
models
models-master/official/legacy/detection/configs/base_config.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Base config template.""" BACKBONES = [ 'resnet', 'spinenet', ] MULTILEVEL_FEATURES = [ 'fpn', 'identity', ] # pylint: disable=line-too-long # For ResNet, this freezes the variables of the first conv1 and conv2_x # layers [1], which leads to higher training speed and slightly better testing # accuracy. The intuition is that the low-level architecture (e.g., ResNet-50) # is able to capture low-level features such as edges; therefore, it does not # need to be fine-tuned for the detection task. # Note that we need to trailing `/` to avoid the incorrect match. # [1]: https://github.com/facebookresearch/Detectron/blob/master/detectron/core/config.py#L198 RESNET_FROZEN_VAR_PREFIX = r'(resnet\d+)\/(conv2d(|_([1-9]|10))|batch_normalization(|_([1-9]|10)))\/' REGULARIZATION_VAR_REGEX = r'.*(kernel|weight):0$' BASE_CFG = { 'model_dir': '', 'use_tpu': True, 'strategy_type': 'tpu', 'isolate_session_state': False, 'train': { 'iterations_per_loop': 100, 'batch_size': 64, 'total_steps': 22500, 'num_cores_per_replica': None, 'input_partition_dims': None, 'optimizer': { 'type': 'momentum', 'momentum': 0.9, 'nesterov': True, # `False` is better for TPU v3-128. }, 'learning_rate': { 'type': 'step', 'warmup_learning_rate': 0.0067, 'warmup_steps': 500, 'init_learning_rate': 0.08, 'learning_rate_levels': [0.008, 0.0008], 'learning_rate_steps': [15000, 20000], }, 'checkpoint': { 'path': '', 'prefix': '', }, # One can use 'RESNET_FROZEN_VAR_PREFIX' to speed up ResNet training # when loading from the checkpoint. 'frozen_variable_prefix': '', 'train_file_pattern': '', 'train_dataset_type': 'tfrecord', # TODO(b/142174042): Support transpose_input option. 'transpose_input': False, 'regularization_variable_regex': REGULARIZATION_VAR_REGEX, 'l2_weight_decay': 0.0001, 'gradient_clip_norm': 0.0, 'input_sharding': False, }, 'eval': { 'input_sharding': True, 'batch_size': 8, 'eval_samples': 5000, 'min_eval_interval': 180, 'eval_timeout': None, 'num_steps_per_eval': 1000, 'type': 'box', 'use_json_file': True, 'val_json_file': '', 'eval_file_pattern': '', 'eval_dataset_type': 'tfrecord', # When visualizing images, set evaluation batch size to 40 to avoid # potential OOM. 'num_images_to_visualize': 0, }, 'predict': { 'batch_size': 8, }, 'architecture': { 'backbone': 'resnet', 'min_level': 3, 'max_level': 7, 'multilevel_features': 'fpn', 'use_bfloat16': True, # Note that `num_classes` is the total number of classes including # one background classes whose index is 0. 'num_classes': 91, }, 'anchor': { 'num_scales': 3, 'aspect_ratios': [1.0, 2.0, 0.5], 'anchor_size': 4.0, }, 'norm_activation': { 'activation': 'relu', 'batch_norm_momentum': 0.997, 'batch_norm_epsilon': 1e-4, 'batch_norm_trainable': True, 'use_sync_bn': False, }, 'resnet': { 'resnet_depth': 50, }, 'spinenet': { 'model_id': '49', }, 'fpn': { 'fpn_feat_dims': 256, 'use_separable_conv': False, 'use_batch_norm': True, }, 'postprocess': { 'use_batched_nms': False, 'max_total_size': 100, 'nms_iou_threshold': 0.5, 'score_threshold': 0.05, 'pre_nms_num_boxes': 5000, }, 'enable_summary': False, } # pylint: enable=line-too-long
4,468
30.695035
101
py
models
models-master/official/legacy/detection/configs/factory.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Factory to provide model configs.""" from official.legacy.detection.configs import maskrcnn_config from official.legacy.detection.configs import olnmask_config from official.legacy.detection.configs import retinanet_config from official.legacy.detection.configs import shapemask_config from official.modeling.hyperparams import params_dict def config_generator(model): """Model function generator.""" if model == 'retinanet': default_config = retinanet_config.RETINANET_CFG restrictions = retinanet_config.RETINANET_RESTRICTIONS elif model == 'mask_rcnn': default_config = maskrcnn_config.MASKRCNN_CFG restrictions = maskrcnn_config.MASKRCNN_RESTRICTIONS elif model == 'olnmask': default_config = olnmask_config.OLNMASK_CFG restrictions = olnmask_config.OLNMASK_RESTRICTIONS elif model == 'shapemask': default_config = shapemask_config.SHAPEMASK_CFG restrictions = shapemask_config.SHAPEMASK_RESTRICTIONS else: raise ValueError('Model %s is not supported.' % model) return params_dict.ParamsDict(default_config, restrictions)
1,692
39.309524
74
py
models
models-master/official/legacy/detection/configs/retinanet_config.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Config template to train Retinanet.""" from official.legacy.detection.configs import base_config from official.modeling.hyperparams import params_dict # pylint: disable=line-too-long RETINANET_CFG = params_dict.ParamsDict(base_config.BASE_CFG) RETINANET_CFG.override({ 'type': 'retinanet', 'architecture': { 'parser': 'retinanet_parser', }, 'retinanet_parser': { 'output_size': [640, 640], 'num_channels': 3, 'match_threshold': 0.5, 'unmatched_threshold': 0.5, 'aug_rand_hflip': True, 'aug_scale_min': 1.0, 'aug_scale_max': 1.0, 'use_autoaugment': False, 'autoaugment_policy_name': 'v0', 'skip_crowd_during_training': True, 'max_num_instances': 100, }, 'retinanet_head': { 'num_convs': 4, 'num_filters': 256, 'use_separable_conv': False, }, 'retinanet_loss': { 'focal_loss_alpha': 0.25, 'focal_loss_gamma': 1.5, 'huber_loss_delta': 0.1, 'box_loss_weight': 50, }, 'enable_summary': True, }, is_strict=False) RETINANET_RESTRICTIONS = [ ] # pylint: enable=line-too-long
1,776
29.118644
74
py
models
models-master/official/legacy/detection/configs/maskrcnn_config.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Config template to train Mask R-CNN.""" from official.legacy.detection.configs import base_config from official.modeling.hyperparams import params_dict # pylint: disable=line-too-long MASKRCNN_CFG = params_dict.ParamsDict(base_config.BASE_CFG) MASKRCNN_CFG.override({ 'type': 'mask_rcnn', 'eval': { 'type': 'box_and_mask', 'num_images_to_visualize': 0, }, 'architecture': { 'parser': 'maskrcnn_parser', 'min_level': 2, 'max_level': 6, 'include_mask': True, 'mask_target_size': 28, }, 'maskrcnn_parser': { 'output_size': [1024, 1024], 'num_channels': 3, 'rpn_match_threshold': 0.7, 'rpn_unmatched_threshold': 0.3, 'rpn_batch_size_per_im': 256, 'rpn_fg_fraction': 0.5, 'aug_rand_hflip': True, 'aug_scale_min': 1.0, 'aug_scale_max': 1.0, 'skip_crowd_during_training': True, 'max_num_instances': 100, 'mask_crop_size': 112, }, 'anchor': { 'num_scales': 1, 'anchor_size': 8, }, 'rpn_head': { 'num_convs': 2, 'num_filters': 256, 'use_separable_conv': False, 'use_batch_norm': False, }, 'frcnn_head': { 'num_convs': 0, 'num_filters': 256, 'use_separable_conv': False, 'num_fcs': 2, 'fc_dims': 1024, 'use_batch_norm': False, }, 'mrcnn_head': { 'num_convs': 4, 'num_filters': 256, 'use_separable_conv': False, 'use_batch_norm': False, }, 'rpn_score_loss': { 'rpn_batch_size_per_im': 256, }, 'rpn_box_loss': { 'huber_loss_delta': 1.0 / 9.0, }, 'frcnn_box_loss': { 'huber_loss_delta': 1.0, }, 'roi_proposal': { 'rpn_pre_nms_top_k': 2000, 'rpn_post_nms_top_k': 1000, 'rpn_nms_threshold': 0.7, 'rpn_score_threshold': 0.0, 'rpn_min_size_threshold': 0.0, 'test_rpn_pre_nms_top_k': 1000, 'test_rpn_post_nms_top_k': 1000, 'test_rpn_nms_threshold': 0.7, 'test_rpn_score_threshold': 0.0, 'test_rpn_min_size_threshold': 0.0, 'use_batched_nms': False, }, 'roi_sampling': { 'num_samples_per_image': 512, 'fg_fraction': 0.25, 'fg_iou_thresh': 0.5, 'bg_iou_thresh_hi': 0.5, 'bg_iou_thresh_lo': 0.0, 'mix_gt_boxes': True, }, 'mask_sampling': { 'num_mask_samples_per_image': 128, # Typically = `num_samples_per_image` * `fg_fraction`. }, 'postprocess': { 'pre_nms_num_boxes': 1000, }, }, is_strict=False) MASKRCNN_RESTRICTIONS = [ ] # pylint: enable=line-too-long
3,328
27.698276
98
py
models
models-master/official/legacy/detection/configs/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/detection/configs/shapemask_config.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Config to train shapemask on COCO.""" from official.legacy.detection.configs import base_config from official.modeling.hyperparams import params_dict SHAPEMASK_RESNET_FROZEN_VAR_PREFIX = r'(conv2d(|_([1-9]|10))|batch_normalization(|_([1-9]|10)))\/' SHAPEMASK_CFG = params_dict.ParamsDict(base_config.BASE_CFG) SHAPEMASK_CFG.override({ 'type': 'shapemask', 'architecture': { 'parser': 'shapemask_parser', 'backbone': 'resnet', 'multilevel_features': 'fpn', 'outer_box_scale': 1.25, }, 'train': { 'total_steps': 45000, 'learning_rate': { 'learning_rate_steps': [30000, 40000], }, 'frozen_variable_prefix': SHAPEMASK_RESNET_FROZEN_VAR_PREFIX, 'regularization_variable_regex': None, }, 'eval': { 'type': 'shapemask_box_and_mask', 'mask_eval_class': 'all', # 'all', 'voc', or 'nonvoc'. }, 'shapemask_parser': { 'output_size': [640, 640], 'num_channels': 3, 'match_threshold': 0.5, 'unmatched_threshold': 0.5, 'aug_rand_hflip': True, 'aug_scale_min': 0.8, 'aug_scale_max': 1.2, 'skip_crowd_during_training': True, 'max_num_instances': 100, # Shapemask specific parameters 'mask_train_class': 'all', # 'all', 'voc', or 'nonvoc'. 'use_category': True, 'outer_box_scale': 1.25, 'num_sampled_masks': 8, 'mask_crop_size': 32, 'mask_min_level': 3, 'mask_max_level': 5, 'box_jitter_scale': 0.025, 'upsample_factor': 4, }, 'retinanet_head': { 'num_convs': 4, 'num_filters': 256, 'use_separable_conv': False, 'use_batch_norm': True, }, 'shapemask_head': { 'num_downsample_channels': 128, 'mask_crop_size': 32, 'use_category_for_mask': True, 'num_convs': 4, 'upsample_factor': 4, 'shape_prior_path': '', }, 'retinanet_loss': { 'focal_loss_alpha': 0.4, 'focal_loss_gamma': 1.5, 'huber_loss_delta': 0.15, 'box_loss_weight': 50, }, 'shapemask_loss': { 'shape_prior_loss_weight': 0.1, 'coarse_mask_loss_weight': 1.0, 'fine_mask_loss_weight': 1.0, }, }, is_strict=False) SHAPEMASK_RESTRICTIONS = [ 'shapemask_head.mask_crop_size == shapemask_parser.mask_crop_size', 'shapemask_head.upsample_factor == shapemask_parser.upsample_factor', 'shapemask_parser.outer_box_scale == architecture.outer_box_scale', ] # pylint: enable=line-too-long
3,204
31.704082
98
py
models
models-master/official/legacy/detection/executor/distributed_executor.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Custom training loop for running TensorFlow 2.0 models.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import os from typing import Optional, Dict, List, Text, Callable, Union, Iterator, Any from absl import flags from absl import logging import numpy as np import tensorflow as tf # pylint: disable=unused-import,g-import-not-at-top,redefined-outer-name,reimported from official.common import distribute_utils from official.modeling.hyperparams import params_dict from official.utils import hyperparams_flags from official.utils.misc import keras_utils FLAGS = flags.FLAGS strategy_flags_dict = hyperparams_flags.strategy_flags_dict hparam_flags_dict = hyperparams_flags.hparam_flags_dict def _save_checkpoint(checkpoint, model_dir, checkpoint_prefix): """Saves model to model_dir with provided checkpoint prefix.""" checkpoint_path = os.path.join(model_dir, checkpoint_prefix) saved_path = checkpoint.save(checkpoint_path) logging.info('Saving model as TF checkpoint: %s', saved_path) def _steps_to_run(current_step, total_steps, steps_per_loop): """Calculates steps to run on device.""" if steps_per_loop <= 0: raise ValueError('steps_per_loop should be positive integer.') return min(total_steps - current_step, steps_per_loop) def _no_metric(): return None def metrics_as_dict(metric): """Puts input metric(s) into a list. Args: metric: metric(s) to be put into the list. `metric` could be an object, a list, or a dict of tf.keras.metrics.Metric or has the `required_method`. Returns: A dictionary of valid metrics. """ if isinstance(metric, tf.keras.metrics.Metric): metrics = {metric.name: metric} elif isinstance(metric, list): metrics = {m.name: m for m in metric} elif isinstance(metric, dict): metrics = metric elif not metric: return {} else: metrics = {'metric': metric} return metrics def metric_results(metric): """Collects results from the given metric(s).""" metrics = metrics_as_dict(metric) metric_result = { name: m.result().numpy().astype(float) for name, m in metrics.items() } return metric_result def reset_states(metric): """Resets states of the given metric(s).""" metrics = metrics_as_dict(metric) for m in metrics.values(): m.reset_states() class SummaryWriter(object): """Simple SummaryWriter for writing dictionary of metrics. Attributes: writer: The tf.SummaryWriter. """ def __init__(self, model_dir: Text, name: Text): """Inits SummaryWriter with paths. Args: model_dir: the model folder path. name: the summary subfolder name. """ self.writer = tf.summary.create_file_writer(os.path.join(model_dir, name)) def __call__(self, metrics: Union[Dict[Text, float], float], step: int): """Write metrics to summary with the given writer. Args: metrics: a dictionary of metrics values. Prefer dictionary. step: integer. The training step. """ if not isinstance(metrics, dict): # Support scalar metric without name. logging.warning('Warning: summary writer prefer metrics as dictionary.') metrics = {'metric': metrics} with self.writer.as_default(): for k, v in metrics.items(): tf.summary.scalar(k, v, step=step) self.writer.flush() class DistributedExecutor(object): """Interface to train and eval models with tf.distribute.Strategy.""" def __init__(self, strategy, params, model_fn, loss_fn, is_multi_host=False): """Constructor. Args: strategy: an instance of tf.distribute.Strategy. params: Model configuration needed to run distribution strategy. model_fn: Keras model function. Signature: (params: ParamsDict) -> tf.keras.models.Model. loss_fn: loss function. Signature: (y_true: Tensor, y_pred: Tensor) -> Tensor is_multi_host: Set to True when using multi hosts for training, like multi worker GPU or TPU pod (slice). Otherwise, False. """ self._params = params self._model_fn = model_fn self._loss_fn = loss_fn self._strategy = strategy self._checkpoint_name = 'ctl_step_{step}.ckpt' self._is_multi_host = is_multi_host self.train_summary_writer = None self.eval_summary_writer = None self.global_train_step = None @property def checkpoint_name(self): """Returns default checkpoint name.""" return self._checkpoint_name @checkpoint_name.setter def checkpoint_name(self, name): """Sets default summary writer for the current thread.""" self._checkpoint_name = name def loss_fn(self): return self._loss_fn() def model_fn(self, params): return self._model_fn(params) def _save_config(self, model_dir): """Save parameters to config files if model_dir is defined.""" logging.info('Save config to model_dir %s.', model_dir) if model_dir: if not tf.io.gfile.exists(model_dir): tf.io.gfile.makedirs(model_dir) self._params.lock() params_dict.save_params_dict_to_yaml(self._params, model_dir + '/params.yaml') else: logging.warning('model_dir is empty, so skip the save config.') def _get_input_iterator( self, input_fn: Callable[..., tf.data.Dataset], strategy: tf.distribute.Strategy) -> Optional[Iterator[Any]]: """Returns distributed dataset iterator. Args: input_fn: (params: dict) -> tf.data.Dataset. strategy: an instance of tf.distribute.Strategy. Returns: An iterator that yields input tensors. """ if input_fn is None: return None # When training with multiple TPU workers, datasets needs to be cloned # across workers. Since Dataset instance cannot be cloned in eager mode, # we instead pass callable that returns a dataset. if self._is_multi_host: return iter(strategy.distribute_datasets_from_function(input_fn)) else: input_data = input_fn() return iter(strategy.experimental_distribute_dataset(input_data)) def _create_replicated_step(self, strategy, model, loss_fn, optimizer, metric=None): """Creates a single training step. Args: strategy: an instance of tf.distribute.Strategy. model: (Tensor, bool) -> Tensor. model function. loss_fn: (y_true: Tensor, y_pred: Tensor) -> Tensor. optimizer: tf.keras.optimizers.Optimizer. metric: tf.keras.metrics.Metric subclass. Returns: The training step callable. """ metrics = metrics_as_dict(metric) def _replicated_step(inputs): """Replicated training step.""" inputs, labels = inputs with tf.GradientTape() as tape: outputs = model(inputs, training=True) prediction_loss = loss_fn(labels, outputs) loss = tf.reduce_mean(prediction_loss) loss = loss / strategy.num_replicas_in_sync for m in metrics.values(): m.update_state(labels, outputs) grads = tape.gradient(loss, model.trainable_variables) optimizer.apply_gradients(zip(grads, model.trainable_variables)) return loss return _replicated_step def _create_train_step(self, strategy, model, loss_fn, optimizer, metric=None): """Creates a distributed training step. Args: strategy: an instance of tf.distribute.Strategy. model: (Tensor, bool) -> Tensor. model function. loss_fn: (y_true: Tensor, y_pred: Tensor) -> Tensor. optimizer: tf.keras.optimizers.Optimizer. metric: tf.keras.metrics.Metric subclass. Returns: The training step callable. """ replicated_step = self._create_replicated_step(strategy, model, loss_fn, optimizer, metric) @tf.function def train_step(iterator, num_steps): """Performs a distributed training step. Args: iterator: an iterator that yields input tensors. num_steps: the number of steps in the loop. Returns: The loss tensor. """ if not isinstance(num_steps, tf.Tensor): raise ValueError('steps should be an Tensor. Python object may cause ' 'retracing.') per_replica_losses = strategy.run(replicated_step, args=(next(iterator),)) for _ in tf.range(num_steps - 1): per_replica_losses = strategy.run( replicated_step, args=(next(iterator),)) # For reporting, we returns the mean of losses. losses = tf.nest.map_structure( lambda x: strategy.reduce(tf.distribute.ReduceOp.MEAN, x, axis=None), per_replica_losses) return losses return train_step def _create_test_step(self, strategy, model, metric): """Creates a distributed test step.""" metrics = metrics_as_dict(metric) @tf.function def test_step(iterator): """Calculates evaluation metrics on distributed devices.""" if not metric: logging.info('Skip test_step because metric is None (%s)', metric) return None, None def _test_step_fn(inputs): """Replicated accuracy calculation.""" inputs, labels = inputs model_outputs = model(inputs, training=False) for m in metrics.values(): m.update_state(labels, model_outputs) return labels, model_outputs return strategy.run(_test_step_fn, args=(next(iterator),)) return test_step def train( self, train_input_fn: Callable[[params_dict.ParamsDict], tf.data.Dataset], eval_input_fn: Optional[Callable[[params_dict.ParamsDict], tf.data.Dataset]] = None, model_dir: Optional[Text] = None, total_steps: int = 1, iterations_per_loop: int = 1, train_metric_fn: Optional[Callable[[], Any]] = None, eval_metric_fn: Optional[Callable[[], Any]] = None, summary_writer_fn: Callable[[Text, Text], SummaryWriter] = SummaryWriter, init_checkpoint: Optional[Callable[[tf.keras.Model], Any]] = None, custom_callbacks: Optional[List[tf.keras.callbacks.Callback]] = None, continuous_eval: bool = False, save_config: bool = True): """Runs distributed training. Args: train_input_fn: (params: dict) -> tf.data.Dataset training data input function. eval_input_fn: (Optional) same type as train_input_fn. If not None, will trigger evaluating metric on eval data. If None, will not run the eval step. model_dir: the folder path for model checkpoints. total_steps: total training steps. iterations_per_loop: train steps per loop. After each loop, this job will update metrics like loss and save checkpoint. train_metric_fn: metric_fn for evaluation in train_step. eval_metric_fn: metric_fn for evaluation in test_step. summary_writer_fn: function to create summary writer. init_checkpoint: function to load checkpoint. custom_callbacks: A list of Keras Callbacks objects to run during training. More specifically, `on_batch_begin()`, `on_batch_end()`, methods are invoked during training. continuous_eval: If `True`, will continously run evaluation on every available checkpoints. If `False`, will do the evaluation once after the final step. save_config: bool. Whether to save params to model_dir. Returns: The training loss and eval metrics. """ assert train_input_fn is not None if train_metric_fn and not callable(train_metric_fn): raise ValueError('if `train_metric_fn` is specified, ' 'train_metric_fn must be a callable.') if eval_metric_fn and not callable(eval_metric_fn): raise ValueError('if `eval_metric_fn` is specified, ' 'eval_metric_fn must be a callable.') train_metric_fn = train_metric_fn or _no_metric eval_metric_fn = eval_metric_fn or _no_metric if custom_callbacks and iterations_per_loop != 1: logging.warning( 'It is sematically wrong to run callbacks when ' 'iterations_per_loop is not one (%s)', iterations_per_loop) custom_callbacks = custom_callbacks or [] def _run_callbacks_on_batch_begin(batch): """Runs custom callbacks at the start of every step.""" if not custom_callbacks: return for callback in custom_callbacks: if callback: callback.on_batch_begin(batch) def _run_callbacks_on_batch_end(batch): """Runs custom callbacks at the end of every step.""" if not custom_callbacks: return for callback in custom_callbacks: if callback: callback.on_batch_end(batch) if save_config: self._save_config(model_dir) if FLAGS.save_checkpoint_freq: save_freq = FLAGS.save_checkpoint_freq else: save_freq = iterations_per_loop params = self._params strategy = self._strategy # To reduce unnecessary send/receive input pipeline operation, we place # input pipeline ops in worker task. train_iterator = self._get_input_iterator(train_input_fn, strategy) train_loss = None train_metric_result = None eval_metric_result = None tf.keras.backend.set_learning_phase(1) with strategy.scope(): # To correctly place the model weights on accelerators, # model and optimizer should be created in scope. model = self.model_fn(params.as_dict()) if not hasattr(model, 'optimizer'): raise ValueError('User should set optimizer attribute to model ' 'inside `model_fn`.') optimizer = model.optimizer # Training loop starts here. checkpoint = tf.train.Checkpoint(model=model, optimizer=optimizer) latest_checkpoint_file = tf.train.latest_checkpoint(model_dir) initial_step = 0 if latest_checkpoint_file: logging.info( 'Checkpoint file %s found and restoring from ' 'checkpoint', latest_checkpoint_file) checkpoint.restore(latest_checkpoint_file) initial_step = optimizer.iterations.numpy() logging.info('Loading from checkpoint file completed. Init step %d', initial_step) elif init_checkpoint: logging.info('Restoring from init checkpoint function') init_checkpoint(model) logging.info('Loading from init checkpoint file completed') current_step = optimizer.iterations.numpy() checkpoint_name = self.checkpoint_name eval_metric = eval_metric_fn() train_metric = train_metric_fn() train_summary_writer = summary_writer_fn(model_dir, 'eval_train') self.train_summary_writer = train_summary_writer.writer test_summary_writer = summary_writer_fn(model_dir, 'eval_test') self.eval_summary_writer = test_summary_writer.writer # Use training summary writer in TimeHistory if it's in use for cb in custom_callbacks: if isinstance(cb, keras_utils.TimeHistory): cb.summary_writer = self.train_summary_writer # Continue training loop. train_step = self._create_train_step( strategy=strategy, model=model, loss_fn=self.loss_fn(), optimizer=optimizer, metric=train_metric) test_step = None if eval_input_fn and eval_metric: self.global_train_step = model.optimizer.iterations test_step = self._create_test_step(strategy, model, metric=eval_metric) # Step-0 operations if current_step == 0 and not latest_checkpoint_file: _save_checkpoint(checkpoint, model_dir, checkpoint_name.format(step=current_step)) if test_step: eval_iterator = self._get_input_iterator(eval_input_fn, strategy) eval_metric_result = self._run_evaluation(test_step, current_step, eval_metric, eval_iterator) logging.info('Step: %s evalation metric = %s.', current_step, eval_metric_result) test_summary_writer(metrics=eval_metric_result, step=optimizer.iterations) reset_states(eval_metric) logging.info('Training started') last_save_checkpoint_step = current_step while current_step < total_steps: num_steps = _steps_to_run(current_step, total_steps, iterations_per_loop) _run_callbacks_on_batch_begin(current_step) train_loss = train_step(train_iterator, tf.convert_to_tensor(num_steps, dtype=tf.int32)) current_step += num_steps train_loss = tf.nest.map_structure(lambda x: x.numpy().astype(float), train_loss) _run_callbacks_on_batch_end(current_step - 1) if not isinstance(train_loss, dict): train_loss = {'total_loss': train_loss} if np.isnan(train_loss['total_loss']): raise ValueError('total loss is NaN.') if train_metric: train_metric_result = metric_results(train_metric) train_metric_result.update(train_loss) else: train_metric_result = train_loss if callable(optimizer.lr): train_metric_result.update( {'learning_rate': optimizer.lr(current_step).numpy()}) else: train_metric_result.update({'learning_rate': optimizer.lr.numpy()}) logging.info('Train Step: %d/%d / loss = %s / training metric = %s', current_step, total_steps, train_loss, train_metric_result) train_summary_writer( metrics=train_metric_result, step=optimizer.iterations) # Saves model checkpoints and run validation steps at every # iterations_per_loop steps. # To avoid repeated model saving, we do not save after the last # step of training. if save_freq > 0 and current_step < total_steps and ( current_step - last_save_checkpoint_step) >= save_freq: _save_checkpoint(checkpoint, model_dir, checkpoint_name.format(step=current_step)) last_save_checkpoint_step = current_step if continuous_eval and current_step < total_steps and test_step: eval_iterator = self._get_input_iterator(eval_input_fn, strategy) eval_metric_result = self._run_evaluation(test_step, current_step, eval_metric, eval_iterator) logging.info('Step: %s evalation metric = %s.', current_step, eval_metric_result) test_summary_writer( metrics=eval_metric_result, step=optimizer.iterations) # Re-initialize evaluation metric, except the last step. if eval_metric and current_step < total_steps: reset_states(eval_metric) if train_metric and current_step < total_steps: reset_states(train_metric) # Reaches the end of training and saves the last checkpoint. if last_save_checkpoint_step < total_steps: _save_checkpoint(checkpoint, model_dir, checkpoint_name.format(step=current_step)) if test_step: logging.info('Running final evaluation after training is complete.') eval_iterator = self._get_input_iterator(eval_input_fn, strategy) eval_metric_result = self._run_evaluation(test_step, current_step, eval_metric, eval_iterator) logging.info('Final evaluation metric = %s.', eval_metric_result) test_summary_writer(metrics=eval_metric_result, step=optimizer.iterations) self.train_summary_writer.close() self.eval_summary_writer.close() return train_metric_result, eval_metric_result def _run_evaluation(self, test_step, current_training_step, metric, test_iterator): """Runs validation steps and aggregate metrics.""" if not test_iterator or not metric: logging.warning( 'Both test_iterator (%s) and metrics (%s) must not be None.', test_iterator, metric) return None logging.info('Running evaluation after step: %s.', current_training_step) eval_step = 0 while True: try: with tf.experimental.async_scope(): test_step(test_iterator) eval_step += 1 except (StopIteration, tf.errors.OutOfRangeError): tf.experimental.async_clear_error() break metric_result = metric_results(metric) logging.info('Total eval steps: [%d]', eval_step) logging.info('At training step: [%r] Validation metric = %r', current_training_step, metric_result) return metric_result def evaluate_from_model_dir( self, model_dir: Text, eval_input_fn: Callable[[params_dict.ParamsDict], tf.data.Dataset], eval_metric_fn: Callable[[], Any], total_steps: int = -1, eval_timeout: Optional[int] = None, min_eval_interval: int = 180, summary_writer_fn: Callable[[Text, Text], SummaryWriter] = SummaryWriter): """Runs distributed evaluation on model folder. Args: model_dir: the folder for storing model checkpoints. eval_input_fn: (Optional) same type as train_input_fn. If not None, will trigger evaluting metric on eval data. If None, will not run eval step. eval_metric_fn: metric_fn for evaluation in test_step. total_steps: total training steps. If the current step reaches the total_steps, the evaluation loop will stop. eval_timeout: The maximum number of seconds to wait between checkpoints. If left as None, then the process will wait indefinitely. Used by tf.train.checkpoints_iterator. min_eval_interval: The minimum number of seconds between yielding checkpoints. Used by tf.train.checkpoints_iterator. summary_writer_fn: function to create summary writer. Returns: Eval metrics dictionary of the last checkpoint. """ if not model_dir: raise ValueError('model_dir must be set.') def terminate_eval(): tf.logging.info('Terminating eval after %d seconds of no checkpoints' % eval_timeout) return True summary_writer = summary_writer_fn(model_dir, 'eval') self.eval_summary_writer = summary_writer.writer # Read checkpoints from the given model directory # until `eval_timeout` seconds elapses. for checkpoint_path in tf.train.checkpoints_iterator( model_dir, min_interval_secs=min_eval_interval, timeout=eval_timeout, timeout_fn=terminate_eval): eval_metric_result, current_step = self.evaluate_checkpoint( checkpoint_path=checkpoint_path, eval_input_fn=eval_input_fn, eval_metric_fn=eval_metric_fn, summary_writer=summary_writer) if total_steps > 0 and current_step >= total_steps: logging.info('Evaluation finished after training step %d', current_step) break return eval_metric_result def evaluate_checkpoint(self, checkpoint_path: Text, eval_input_fn: Callable[[params_dict.ParamsDict], tf.data.Dataset], eval_metric_fn: Callable[[], Any], summary_writer: Optional[SummaryWriter] = None): """Runs distributed evaluation on the one checkpoint. Args: checkpoint_path: the checkpoint to evaluate. eval_input_fn: (Optional) same type as train_input_fn. If not None, will trigger evaluting metric on eval data. If None, will not run eval step. eval_metric_fn: metric_fn for evaluation in test_step. summary_writer: function to create summary writer. Returns: Eval metrics dictionary of the last checkpoint. """ if not callable(eval_metric_fn): raise ValueError('if `eval_metric_fn` is specified, ' 'eval_metric_fn must be a callable.') old_phase = tf.keras.backend.learning_phase() tf.keras.backend.set_learning_phase(0) params = self._params strategy = self._strategy # To reduce unnecessary send/receive input pipeline operation, we place # input pipeline ops in worker task. with strategy.scope(): # To correctly place the model weights on accelerators, # model and optimizer should be created in scope. model = self.model_fn(params.as_dict()) checkpoint = tf.train.Checkpoint(model=model) eval_metric = eval_metric_fn() assert eval_metric, 'eval_metric does not exist' test_step = self._create_test_step(strategy, model, metric=eval_metric) logging.info('Starting to evaluate.') if not checkpoint_path: raise ValueError('checkpoint path is empty') reader = tf.compat.v1.train.NewCheckpointReader(checkpoint_path) if reader.has_tensor('optimizer/iter/.ATTRIBUTES/VARIABLE_VALUE'): # Legacy keras optimizer iteration. current_step = reader.get_tensor( 'optimizer/iter/.ATTRIBUTES/VARIABLE_VALUE') else: # New keras optimizer iteration. current_step = reader.get_tensor( 'optimizer/_iterations/.ATTRIBUTES/VARIABLE_VALUE') logging.info('Checkpoint file %s found and restoring from ' 'checkpoint', checkpoint_path) status = checkpoint.restore(checkpoint_path) status.expect_partial().assert_existing_objects_matched() self.global_train_step = model.optimizer.iterations eval_iterator = self._get_input_iterator(eval_input_fn, strategy) eval_metric_result = self._run_evaluation(test_step, current_step, eval_metric, eval_iterator) logging.info('Step: %s evalation metric = %s.', current_step, eval_metric_result) summary_writer(metrics=eval_metric_result, step=current_step) reset_states(eval_metric) tf.keras.backend.set_learning_phase(old_phase) return eval_metric_result, current_step def predict(self): return NotImplementedError('Unimplmented function.') class ExecutorBuilder(object): """Builder of DistributedExecutor. Example 1: Builds an executor with supported Strategy. builder = ExecutorBuilder( strategy_type='tpu', strategy_config={'tpu': '/bns/xxx'}) dist_executor = builder.build_executor( params=params, model_fn=my_model_fn, loss_fn=my_loss_fn, metric_fn=my_metric_fn) Example 2: Builds an executor with customized Strategy. builder = ExecutorBuilder() builder.strategy = <some customized Strategy> dist_executor = builder.build_executor( params=params, model_fn=my_model_fn, loss_fn=my_loss_fn, metric_fn=my_metric_fn) Example 3: Builds a customized executor with customized Strategy. class MyDistributedExecutor(DistributedExecutor): # implementation ... builder = ExecutorBuilder() builder.strategy = <some customized Strategy> dist_executor = builder.build_executor( class_ctor=MyDistributedExecutor, params=params, model_fn=my_model_fn, loss_fn=my_loss_fn, metric_fn=my_metric_fn) """ def __init__(self, strategy_type=None, strategy_config=None): _ = distribute_utils.configure_cluster(strategy_config.worker_hosts, strategy_config.task_index) """Constructor. Args: strategy_type: string. One of 'tpu', 'mirrored', 'multi_worker_mirrored'. If None, the user is responsible to set the strategy before calling build_executor(...). strategy_config: necessary config for constructing the proper Strategy. Check strategy_flags_dict() for examples of the structure. """ self._strategy = distribute_utils.get_distribution_strategy( distribution_strategy=strategy_type, num_gpus=strategy_config.num_gpus, all_reduce_alg=strategy_config.all_reduce_alg, num_packs=strategy_config.num_packs, tpu_address=strategy_config.tpu) @property def strategy(self): """Returns default checkpoint name.""" return self._strategy @strategy.setter def strategy(self, new_strategy): """Sets default summary writer for the current thread.""" self._strategy = new_strategy def build_executor(self, class_ctor=DistributedExecutor, params=None, model_fn=None, loss_fn=None, **kwargs): """Creates an executor according to strategy type. See doc string of the DistributedExecutor.__init__ for more information of the input arguments. Args: class_ctor: A constructor of executor (default: DistributedExecutor). params: ParamsDict, all the model parameters and runtime parameters. model_fn: Keras model function. loss_fn: loss function. **kwargs: other arguments to the executor constructor. Returns: An instance of DistributedExecutor or its subclass. """ if self._strategy is None: raise ValueError('`strategy` should not be None. You need to specify ' '`strategy_type` in the builder contructor or directly ' 'set the `strategy` property of the builder.') return class_ctor( strategy=self._strategy, params=params, model_fn=model_fn, loss_fn=loss_fn, **kwargs)
30,360
36.390394
83
py
models
models-master/official/legacy/detection/executor/detection_executor.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """An executor class for running model on TensorFlow 2.0.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from absl import logging import tensorflow as tf from official.legacy.detection.executor import distributed_executor as executor from official.vision.utils.object_detection import visualization_utils class DetectionDistributedExecutor(executor.DistributedExecutor): """Detection specific customer training loop executor. Subclasses the DistributedExecutor and adds support for numpy based metrics. """ def __init__(self, predict_post_process_fn=None, trainable_variables_filter=None, **kwargs): super(DetectionDistributedExecutor, self).__init__(**kwargs) if predict_post_process_fn: assert callable(predict_post_process_fn) if trainable_variables_filter: assert callable(trainable_variables_filter) self._predict_post_process_fn = predict_post_process_fn self._trainable_variables_filter = trainable_variables_filter self.eval_steps = tf.Variable( 0, trainable=False, dtype=tf.int32, synchronization=tf.VariableSynchronization.ON_READ, aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA, shape=[]) def _create_replicated_step(self, strategy, model, loss_fn, optimizer, metric=None): trainable_variables = model.trainable_variables if self._trainable_variables_filter: trainable_variables = self._trainable_variables_filter( trainable_variables) logging.info('Filter trainable variables from %d to %d', len(model.trainable_variables), len(trainable_variables)) update_state_fn = lambda labels, outputs: None if isinstance(metric, tf.keras.metrics.Metric): update_state_fn = metric.update_state else: logging.error('Detection: train metric is not an instance of ' 'tf.keras.metrics.Metric.') def _replicated_step(inputs): """Replicated training step.""" inputs, labels = inputs with tf.GradientTape() as tape: outputs = model(inputs, training=True) all_losses = loss_fn(labels, outputs) losses = {} for k, v in all_losses.items(): losses[k] = tf.reduce_mean(v) per_replica_loss = losses['total_loss'] / strategy.num_replicas_in_sync update_state_fn(labels, outputs) grads = tape.gradient(per_replica_loss, trainable_variables) clipped_grads, _ = tf.clip_by_global_norm(grads, clip_norm=1.0) optimizer.apply_gradients(zip(clipped_grads, trainable_variables)) return losses return _replicated_step def _create_test_step(self, strategy, model, metric): """Creates a distributed test step.""" @tf.function def test_step(iterator, eval_steps): """Calculates evaluation metrics on distributed devices.""" def _test_step_fn(inputs, eval_steps): """Replicated accuracy calculation.""" inputs, labels = inputs model_outputs = model(inputs, training=False) if self._predict_post_process_fn: labels, prediction_outputs = self._predict_post_process_fn( labels, model_outputs) num_remaining_visualizations = ( self._params.eval.num_images_to_visualize - eval_steps) # If there are remaining number of visualizations that needs to be # done, add next batch outputs for visualization. # # TODO(hongjunchoi): Once dynamic slicing is supported on TPU, only # write correct slice of outputs to summary file. if num_remaining_visualizations > 0: visualization_utils.visualize_images_with_bounding_boxes( inputs, prediction_outputs['detection_boxes'], self.global_train_step, self.eval_summary_writer) return labels, prediction_outputs labels, outputs = strategy.run( _test_step_fn, args=( next(iterator), eval_steps, )) outputs = tf.nest.map_structure(strategy.experimental_local_results, outputs) labels = tf.nest.map_structure(strategy.experimental_local_results, labels) eval_steps.assign_add(self._params.eval.batch_size) return labels, outputs return test_step def _run_evaluation(self, test_step, current_training_step, metric, test_iterator): """Runs validation steps and aggregate metrics.""" self.eval_steps.assign(0) if not test_iterator or not metric: logging.warning( 'Both test_iterator (%s) and metrics (%s) must not be None.', test_iterator, metric) return None logging.info('Running evaluation after step: %s.', current_training_step) while True: try: labels, outputs = test_step(test_iterator, self.eval_steps) if metric: metric.update_state(labels, outputs) except (StopIteration, tf.errors.OutOfRangeError): break metric_result = metric.result() if isinstance(metric, tf.keras.metrics.Metric): metric_result = tf.nest.map_structure(lambda x: x.numpy().astype(float), metric_result) logging.info('Step: [%d] Validation metric = %s', current_training_step, metric_result) return metric_result
6,259
38.125
79
py
models
models-master/official/legacy/detection/executor/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/detection/utils/box_utils.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Utility functions for bounding box processing.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import numpy as np import tensorflow as tf EPSILON = 1e-8 BBOX_XFORM_CLIP = np.log(1000. / 16.) def visualize_images_with_bounding_boxes(images, box_outputs, step, summary_writer): """Records subset of evaluation images with bounding boxes.""" image_shape = tf.shape(images[0]) image_height = tf.cast(image_shape[0], tf.float32) image_width = tf.cast(image_shape[1], tf.float32) normalized_boxes = normalize_boxes(box_outputs, [image_height, image_width]) bounding_box_color = tf.constant([[1.0, 1.0, 0.0, 1.0]]) image_summary = tf.image.draw_bounding_boxes(images, normalized_boxes, bounding_box_color) with summary_writer.as_default(): tf.summary.image('bounding_box_summary', image_summary, step=step) summary_writer.flush() def yxyx_to_xywh(boxes): """Converts boxes from ymin, xmin, ymax, xmax to xmin, ymin, width, height. Args: boxes: a numpy array whose last dimension is 4 representing the coordinates of boxes in ymin, xmin, ymax, xmax order. Returns: boxes: a numpy array whose shape is the same as `boxes` in new format. Raises: ValueError: If the last dimension of boxes is not 4. """ if boxes.shape[-1] != 4: raise ValueError('boxes.shape[-1] is {:d}, but must be 4.'.format( boxes.shape[-1])) boxes_ymin = boxes[..., 0] boxes_xmin = boxes[..., 1] boxes_width = boxes[..., 3] - boxes[..., 1] boxes_height = boxes[..., 2] - boxes[..., 0] new_boxes = np.stack([boxes_xmin, boxes_ymin, boxes_width, boxes_height], axis=-1) return new_boxes def jitter_boxes(boxes, noise_scale=0.025): """Jitter the box coordinates by some noise distribution. Args: boxes: a tensor whose last dimension is 4 representing the coordinates of boxes in ymin, xmin, ymax, xmax order. noise_scale: a python float which specifies the magnitude of noise. The rule of thumb is to set this between (0, 0.1]. The default value is found to mimic the noisy detections best empirically. Returns: jittered_boxes: a tensor whose shape is the same as `boxes` representing the jittered boxes. Raises: ValueError: If the last dimension of boxes is not 4. """ if boxes.shape[-1] != 4: raise ValueError('boxes.shape[-1] is {:d}, but must be 4.'.format( boxes.shape[-1])) with tf.name_scope('jitter_boxes'): bbox_jitters = tf.random.normal(boxes.get_shape(), stddev=noise_scale) ymin = boxes[..., 0:1] xmin = boxes[..., 1:2] ymax = boxes[..., 2:3] xmax = boxes[..., 3:4] width = xmax - xmin height = ymax - ymin new_center_x = (xmin + xmax) / 2.0 + bbox_jitters[..., 0:1] * width new_center_y = (ymin + ymax) / 2.0 + bbox_jitters[..., 1:2] * height new_width = width * tf.math.exp(bbox_jitters[..., 2:3]) new_height = height * tf.math.exp(bbox_jitters[..., 3:4]) jittered_boxes = tf.concat([ new_center_y - new_height * 0.5, new_center_x - new_width * 0.5, new_center_y + new_height * 0.5, new_center_x + new_width * 0.5 ], axis=-1) return jittered_boxes def normalize_boxes(boxes, image_shape): """Converts boxes to the normalized coordinates. Args: boxes: a tensor whose last dimension is 4 representing the coordinates of boxes in ymin, xmin, ymax, xmax order. image_shape: a list of two integers, a two-element vector or a tensor such that all but the last dimensions are `broadcastable` to `boxes`. The last dimension is 2, which represents [height, width]. Returns: normalized_boxes: a tensor whose shape is the same as `boxes` representing the normalized boxes. Raises: ValueError: If the last dimension of boxes is not 4. """ if boxes.shape[-1] != 4: raise ValueError('boxes.shape[-1] is {:d}, but must be 4.'.format( boxes.shape[-1])) with tf.name_scope('normalize_boxes'): if isinstance(image_shape, list) or isinstance(image_shape, tuple): height, width = image_shape else: image_shape = tf.cast(image_shape, dtype=boxes.dtype) height = image_shape[..., 0:1] width = image_shape[..., 1:2] ymin = boxes[..., 0:1] / height xmin = boxes[..., 1:2] / width ymax = boxes[..., 2:3] / height xmax = boxes[..., 3:4] / width normalized_boxes = tf.concat([ymin, xmin, ymax, xmax], axis=-1) return normalized_boxes def denormalize_boxes(boxes, image_shape): """Converts boxes normalized by [height, width] to pixel coordinates. Args: boxes: a tensor whose last dimension is 4 representing the coordinates of boxes in ymin, xmin, ymax, xmax order. image_shape: a list of two integers, a two-element vector or a tensor such that all but the last dimensions are `broadcastable` to `boxes`. The last dimension is 2, which represents [height, width]. Returns: denormalized_boxes: a tensor whose shape is the same as `boxes` representing the denormalized boxes. Raises: ValueError: If the last dimension of boxes is not 4. """ with tf.name_scope('denormalize_boxes'): if isinstance(image_shape, list) or isinstance(image_shape, tuple): height, width = image_shape else: image_shape = tf.cast(image_shape, dtype=boxes.dtype) height, width = tf.split(image_shape, 2, axis=-1) ymin, xmin, ymax, xmax = tf.split(boxes, 4, axis=-1) ymin = ymin * height xmin = xmin * width ymax = ymax * height xmax = xmax * width denormalized_boxes = tf.concat([ymin, xmin, ymax, xmax], axis=-1) return denormalized_boxes def clip_boxes(boxes, image_shape): """Clips boxes to image boundaries. Args: boxes: a tensor whose last dimension is 4 representing the coordinates of boxes in ymin, xmin, ymax, xmax order. image_shape: a list of two integers, a two-element vector or a tensor such that all but the last dimensions are `broadcastable` to `boxes`. The last dimension is 2, which represents [height, width]. Returns: clipped_boxes: a tensor whose shape is the same as `boxes` representing the clipped boxes. Raises: ValueError: If the last dimension of boxes is not 4. """ if boxes.shape[-1] != 4: raise ValueError('boxes.shape[-1] is {:d}, but must be 4.'.format( boxes.shape[-1])) with tf.name_scope('clip_boxes'): if isinstance(image_shape, list) or isinstance(image_shape, tuple): height, width = image_shape max_length = [height - 1.0, width - 1.0, height - 1.0, width - 1.0] else: image_shape = tf.cast(image_shape, dtype=boxes.dtype) height, width = tf.unstack(image_shape, axis=-1) max_length = tf.stack( [height - 1.0, width - 1.0, height - 1.0, width - 1.0], axis=-1) clipped_boxes = tf.math.maximum(tf.math.minimum(boxes, max_length), 0.0) return clipped_boxes def compute_outer_boxes(boxes, image_shape, scale=1.0): """Compute outer box encloses an object with a margin. Args: boxes: a tensor whose last dimension is 4 representing the coordinates of boxes in ymin, xmin, ymax, xmax order. image_shape: a list of two integers, a two-element vector or a tensor such that all but the last dimensions are `broadcastable` to `boxes`. The last dimension is 2, which represents [height, width]. scale: a float number specifying the scale of output outer boxes to input `boxes`. Returns: outer_boxes: a tensor whose shape is the same as `boxes` representing the outer boxes. """ if scale < 1.0: raise ValueError( 'scale is {}, but outer box scale must be greater than 1.0.'.format( scale)) centers_y = (boxes[..., 0] + boxes[..., 2]) / 2.0 centers_x = (boxes[..., 1] + boxes[..., 3]) / 2.0 box_height = (boxes[..., 2] - boxes[..., 0]) * scale box_width = (boxes[..., 3] - boxes[..., 1]) * scale outer_boxes = tf.stack([ centers_y - box_height / 2.0, centers_x - box_width / 2.0, centers_y + box_height / 2.0, centers_x + box_width / 2.0 ], axis=1) outer_boxes = clip_boxes(outer_boxes, image_shape) return outer_boxes def encode_boxes(boxes, anchors, weights=None): """Encode boxes to targets. Args: boxes: a tensor whose last dimension is 4 representing the coordinates of boxes in ymin, xmin, ymax, xmax order. anchors: a tensor whose shape is the same as, or `broadcastable` to `boxes`, representing the coordinates of anchors in ymin, xmin, ymax, xmax order. weights: None or a list of four float numbers used to scale coordinates. Returns: encoded_boxes: a tensor whose shape is the same as `boxes` representing the encoded box targets. Raises: ValueError: If the last dimension of boxes is not 4. """ if boxes.shape[-1] != 4: raise ValueError('boxes.shape[-1] is {:d}, but must be 4.'.format( boxes.shape[-1])) with tf.name_scope('encode_boxes'): boxes = tf.cast(boxes, dtype=anchors.dtype) ymin = boxes[..., 0:1] xmin = boxes[..., 1:2] ymax = boxes[..., 2:3] xmax = boxes[..., 3:4] box_h = ymax - ymin + 1.0 box_w = xmax - xmin + 1.0 box_yc = ymin + 0.5 * box_h box_xc = xmin + 0.5 * box_w anchor_ymin = anchors[..., 0:1] anchor_xmin = anchors[..., 1:2] anchor_ymax = anchors[..., 2:3] anchor_xmax = anchors[..., 3:4] anchor_h = anchor_ymax - anchor_ymin + 1.0 anchor_w = anchor_xmax - anchor_xmin + 1.0 anchor_yc = anchor_ymin + 0.5 * anchor_h anchor_xc = anchor_xmin + 0.5 * anchor_w encoded_dy = (box_yc - anchor_yc) / anchor_h encoded_dx = (box_xc - anchor_xc) / anchor_w encoded_dh = tf.math.log(box_h / anchor_h) encoded_dw = tf.math.log(box_w / anchor_w) if weights: encoded_dy *= weights[0] encoded_dx *= weights[1] encoded_dh *= weights[2] encoded_dw *= weights[3] encoded_boxes = tf.concat([encoded_dy, encoded_dx, encoded_dh, encoded_dw], axis=-1) return encoded_boxes def decode_boxes(encoded_boxes, anchors, weights=None): """Decode boxes. Args: encoded_boxes: a tensor whose last dimension is 4 representing the coordinates of encoded boxes in ymin, xmin, ymax, xmax order. anchors: a tensor whose shape is the same as, or `broadcastable` to `boxes`, representing the coordinates of anchors in ymin, xmin, ymax, xmax order. weights: None or a list of four float numbers used to scale coordinates. Returns: encoded_boxes: a tensor whose shape is the same as `boxes` representing the decoded box targets. """ if encoded_boxes.shape[-1] != 4: raise ValueError('encoded_boxes.shape[-1] is {:d}, but must be 4.'.format( encoded_boxes.shape[-1])) with tf.name_scope('decode_boxes'): encoded_boxes = tf.cast(encoded_boxes, dtype=anchors.dtype) dy = encoded_boxes[..., 0:1] dx = encoded_boxes[..., 1:2] dh = encoded_boxes[..., 2:3] dw = encoded_boxes[..., 3:4] if weights: dy /= weights[0] dx /= weights[1] dh /= weights[2] dw /= weights[3] dh = tf.math.minimum(dh, BBOX_XFORM_CLIP) dw = tf.math.minimum(dw, BBOX_XFORM_CLIP) anchor_ymin = anchors[..., 0:1] anchor_xmin = anchors[..., 1:2] anchor_ymax = anchors[..., 2:3] anchor_xmax = anchors[..., 3:4] anchor_h = anchor_ymax - anchor_ymin + 1.0 anchor_w = anchor_xmax - anchor_xmin + 1.0 anchor_yc = anchor_ymin + 0.5 * anchor_h anchor_xc = anchor_xmin + 0.5 * anchor_w decoded_boxes_yc = dy * anchor_h + anchor_yc decoded_boxes_xc = dx * anchor_w + anchor_xc decoded_boxes_h = tf.math.exp(dh) * anchor_h decoded_boxes_w = tf.math.exp(dw) * anchor_w decoded_boxes_ymin = decoded_boxes_yc - 0.5 * decoded_boxes_h decoded_boxes_xmin = decoded_boxes_xc - 0.5 * decoded_boxes_w decoded_boxes_ymax = decoded_boxes_ymin + decoded_boxes_h - 1.0 decoded_boxes_xmax = decoded_boxes_xmin + decoded_boxes_w - 1.0 decoded_boxes = tf.concat([ decoded_boxes_ymin, decoded_boxes_xmin, decoded_boxes_ymax, decoded_boxes_xmax ], axis=-1) return decoded_boxes def encode_boxes_lrtb(boxes, anchors, weights=None): """Encode boxes to targets on lrtb (=left,right,top,bottom) format. Args: boxes: a tensor whose last dimension is 4 representing the coordinates of boxes in ymin, xmin, ymax, xmax order. anchors: a tensor whose shape is the same as, or `broadcastable` to `boxes`, representing the coordinates of anchors in ymin, xmin, ymax, xmax order. weights: None or a list of four float numbers used to scale coordinates. Returns: encoded_boxes_lrtb: a tensor whose shape is the same as `boxes` representing the encoded box targets. The box targets encode the left, right, top, bottom distances from an anchor location to the four borders of the matched groundtruth bounding box. center_targets: centerness targets defined by the left, right, top, and bottom distance targets. The centerness is defined as the deviation of the anchor location from the groundtruth object center. Formally, centerness = sqrt(min(left, right)/max(left, right)*min(top, bottom)/max(top, bottom)). Raises: ValueError: If the last dimension of boxes is not 4. """ if boxes.shape[-1] != 4: raise ValueError( 'boxes.shape[-1] is {:d}, but must be 4.'.format(boxes.shape[-1])) with tf.name_scope('encode_boxes_lrtb'): boxes = tf.cast(boxes, dtype=anchors.dtype) ymin = boxes[..., 0:1] xmin = boxes[..., 1:2] ymax = boxes[..., 2:3] xmax = boxes[..., 3:4] # box_h = ymax - ymin + 1.0 # box_w = xmax - xmin + 1.0 box_h = ymax - ymin box_w = xmax - xmin anchor_ymin = anchors[..., 0:1] anchor_xmin = anchors[..., 1:2] anchor_ymax = anchors[..., 2:3] anchor_xmax = anchors[..., 3:4] # anchor_h = anchor_ymax - anchor_ymin + 1.0 # anchor_w = anchor_xmax - anchor_xmin + 1.0 anchor_h = anchor_ymax - anchor_ymin anchor_w = anchor_xmax - anchor_xmin anchor_yc = anchor_ymin + 0.5 * anchor_h anchor_xc = anchor_xmin + 0.5 * anchor_w box_h += EPSILON box_w += EPSILON anchor_h += EPSILON anchor_w += EPSILON left = (anchor_xc - xmin) / anchor_w right = (xmax - anchor_xc) / anchor_w top = (anchor_yc - ymin) / anchor_h bottom = (ymax - anchor_yc) / anchor_h # Create centerness target. { lrtb_targets = tf.concat([left, right, top, bottom], axis=-1) valid_match = tf.greater(tf.reduce_min(lrtb_targets, -1), 0.0) # Centerness score. left_right = tf.concat([left, right], axis=-1) left_right = tf.where(tf.stack([valid_match, valid_match], -1), left_right, tf.zeros_like(left_right)) top_bottom = tf.concat([top, bottom], axis=-1) top_bottom = tf.where(tf.stack([valid_match, valid_match], -1), top_bottom, tf.zeros_like(top_bottom)) center_targets = tf.sqrt( (tf.reduce_min(left_right, -1) / (tf.reduce_max(left_right, -1) + EPSILON)) * (tf.reduce_min(top_bottom, -1) / (tf.reduce_max(top_bottom, -1) + EPSILON))) center_targets = tf.where(valid_match, center_targets, tf.zeros_like(center_targets)) if weights: left *= weights[0] right *= weights[1] top *= weights[2] bottom *= weights[3] encoded_boxes_lrtb = tf.concat( [left, right, top, bottom], axis=-1) return encoded_boxes_lrtb, center_targets def decode_boxes_lrtb(encoded_boxes_lrtb, anchors, weights=None): """Decode boxes. Args: encoded_boxes_lrtb: a tensor whose last dimension is 4 representing the coordinates of encoded boxes in left, right, top, bottom order. anchors: a tensor whose shape is the same as, or `broadcastable` to `boxes`, representing the coordinates of anchors in ymin, xmin, ymax, xmax order. weights: None or a list of four float numbers used to scale coordinates. Returns: decoded_boxes_lrtb: a tensor whose shape is the same as `boxes` representing the decoded box targets in lrtb (=left,right,top,bottom) format. The box decoded box coordinates represent the left, right, top, and bottom distances from an anchor location to the four borders of the matched groundtruth bounding box. """ if encoded_boxes_lrtb.shape[-1] != 4: raise ValueError( 'encoded_boxes_lrtb.shape[-1] is {:d}, but must be 4.' .format(encoded_boxes_lrtb.shape[-1])) with tf.name_scope('decode_boxes_lrtb'): encoded_boxes_lrtb = tf.cast(encoded_boxes_lrtb, dtype=anchors.dtype) left = encoded_boxes_lrtb[..., 0:1] right = encoded_boxes_lrtb[..., 1:2] top = encoded_boxes_lrtb[..., 2:3] bottom = encoded_boxes_lrtb[..., 3:4] if weights: left /= weights[0] right /= weights[1] top /= weights[2] bottom /= weights[3] anchor_ymin = anchors[..., 0:1] anchor_xmin = anchors[..., 1:2] anchor_ymax = anchors[..., 2:3] anchor_xmax = anchors[..., 3:4] anchor_h = anchor_ymax - anchor_ymin anchor_w = anchor_xmax - anchor_xmin anchor_yc = anchor_ymin + 0.5 * anchor_h anchor_xc = anchor_xmin + 0.5 * anchor_w anchor_h += EPSILON anchor_w += EPSILON decoded_boxes_ymin = anchor_yc - top * anchor_h decoded_boxes_xmin = anchor_xc - left * anchor_w decoded_boxes_ymax = anchor_yc + bottom * anchor_h decoded_boxes_xmax = anchor_xc + right * anchor_w decoded_boxes_lrtb = tf.concat( [decoded_boxes_ymin, decoded_boxes_xmin, decoded_boxes_ymax, decoded_boxes_xmax], axis=-1) return decoded_boxes_lrtb def filter_boxes(boxes, scores, image_shape, min_size_threshold): """Filter and remove boxes that are too small or fall outside the image. Args: boxes: a tensor whose last dimension is 4 representing the coordinates of boxes in ymin, xmin, ymax, xmax order. scores: a tensor whose shape is the same as tf.shape(boxes)[:-1] representing the original scores of the boxes. image_shape: a tensor whose shape is the same as, or `broadcastable` to `boxes` except the last dimension, which is 2, representing [height, width] of the scaled image. min_size_threshold: a float representing the minimal box size in each side (w.r.t. the scaled image). Boxes whose sides are smaller than it will be filtered out. Returns: filtered_boxes: a tensor whose shape is the same as `boxes` but with the position of the filtered boxes are filled with 0. filtered_scores: a tensor whose shape is the same as 'scores' but with the positinon of the filtered boxes filled with 0. """ if boxes.shape[-1] != 4: raise ValueError('boxes.shape[1] is {:d}, but must be 4.'.format( boxes.shape[-1])) with tf.name_scope('filter_boxes'): if isinstance(image_shape, list) or isinstance(image_shape, tuple): height, width = image_shape else: image_shape = tf.cast(image_shape, dtype=boxes.dtype) height = image_shape[..., 0] width = image_shape[..., 1] ymin = boxes[..., 0] xmin = boxes[..., 1] ymax = boxes[..., 2] xmax = boxes[..., 3] h = ymax - ymin + 1.0 w = xmax - xmin + 1.0 yc = ymin + 0.5 * h xc = xmin + 0.5 * w min_size = tf.cast( tf.math.maximum(min_size_threshold, 1.0), dtype=boxes.dtype) filtered_size_mask = tf.math.logical_and( tf.math.greater(h, min_size), tf.math.greater(w, min_size)) filtered_center_mask = tf.logical_and( tf.math.logical_and(tf.math.greater(yc, 0.0), tf.math.less(yc, height)), tf.math.logical_and(tf.math.greater(xc, 0.0), tf.math.less(xc, width))) filtered_mask = tf.math.logical_and(filtered_size_mask, filtered_center_mask) filtered_scores = tf.where(filtered_mask, scores, tf.zeros_like(scores)) filtered_boxes = tf.cast( tf.expand_dims(filtered_mask, axis=-1), dtype=boxes.dtype) * boxes return filtered_boxes, filtered_scores def filter_boxes_by_scores(boxes, scores, min_score_threshold): """Filter and remove boxes whose scores are smaller than the threshold. Args: boxes: a tensor whose last dimension is 4 representing the coordinates of boxes in ymin, xmin, ymax, xmax order. scores: a tensor whose shape is the same as tf.shape(boxes)[:-1] representing the original scores of the boxes. min_score_threshold: a float representing the minimal box score threshold. Boxes whose score are smaller than it will be filtered out. Returns: filtered_boxes: a tensor whose shape is the same as `boxes` but with the position of the filtered boxes are filled with -1. filtered_scores: a tensor whose shape is the same as 'scores' but with the """ if boxes.shape[-1] != 4: raise ValueError('boxes.shape[1] is {:d}, but must be 4.'.format( boxes.shape[-1])) with tf.name_scope('filter_boxes_by_scores'): filtered_mask = tf.math.greater(scores, min_score_threshold) filtered_scores = tf.where(filtered_mask, scores, -tf.ones_like(scores)) filtered_boxes = tf.cast( tf.expand_dims(filtered_mask, axis=-1), dtype=boxes.dtype) * boxes return filtered_boxes, filtered_scores def top_k_boxes(boxes, scores, k): """Sort and select top k boxes according to the scores. Args: boxes: a tensor of shape [batch_size, N, 4] representing the coordiante of the boxes. N is the number of boxes per image. scores: a tensor of shsape [batch_size, N] representing the socre of the boxes. k: an integer or a tensor indicating the top k number. Returns: selected_boxes: a tensor of shape [batch_size, k, 4] representing the selected top k box coordinates. selected_scores: a tensor of shape [batch_size, k] representing the selected top k box scores. """ with tf.name_scope('top_k_boxes'): selected_scores, top_k_indices = tf.nn.top_k(scores, k=k, sorted=True) batch_size, _ = scores.get_shape().as_list() if batch_size == 1: selected_boxes = tf.squeeze( tf.gather(boxes, top_k_indices, axis=1), axis=1) else: top_k_indices_shape = tf.shape(top_k_indices) batch_indices = ( tf.expand_dims(tf.range(top_k_indices_shape[0]), axis=-1) * tf.ones([1, top_k_indices_shape[-1]], dtype=tf.int32)) gather_nd_indices = tf.stack([batch_indices, top_k_indices], axis=-1) selected_boxes = tf.gather_nd(boxes, gather_nd_indices) return selected_boxes, selected_scores def bbox_overlap(boxes, gt_boxes): """Calculates the overlap between proposal and ground truth boxes. Some `gt_boxes` may have been padded. The returned `iou` tensor for these boxes will be -1. Args: boxes: a tensor with a shape of [batch_size, N, 4]. N is the number of proposals before groundtruth assignment (e.g., rpn_post_nms_topn). The last dimension is the pixel coordinates in [ymin, xmin, ymax, xmax] form. gt_boxes: a tensor with a shape of [batch_size, MAX_NUM_INSTANCES, 4]. This tensor might have paddings with a negative value. Returns: iou: a tensor with as a shape of [batch_size, N, MAX_NUM_INSTANCES]. """ with tf.name_scope('bbox_overlap'): bb_y_min, bb_x_min, bb_y_max, bb_x_max = tf.split( value=boxes, num_or_size_splits=4, axis=2) gt_y_min, gt_x_min, gt_y_max, gt_x_max = tf.split( value=gt_boxes, num_or_size_splits=4, axis=2) # Calculates the intersection area. i_xmin = tf.math.maximum(bb_x_min, tf.transpose(gt_x_min, [0, 2, 1])) i_xmax = tf.math.minimum(bb_x_max, tf.transpose(gt_x_max, [0, 2, 1])) i_ymin = tf.math.maximum(bb_y_min, tf.transpose(gt_y_min, [0, 2, 1])) i_ymax = tf.math.minimum(bb_y_max, tf.transpose(gt_y_max, [0, 2, 1])) i_area = tf.math.maximum((i_xmax - i_xmin), 0) * tf.math.maximum( (i_ymax - i_ymin), 0) # Calculates the union area. bb_area = (bb_y_max - bb_y_min) * (bb_x_max - bb_x_min) gt_area = (gt_y_max - gt_y_min) * (gt_x_max - gt_x_min) # Adds a small epsilon to avoid divide-by-zero. u_area = bb_area + tf.transpose(gt_area, [0, 2, 1]) - i_area + 1e-8 # Calculates IoU. iou = i_area / u_area # Fills -1 for IoU entries between the padded ground truth boxes. gt_invalid_mask = tf.less( tf.reduce_max(gt_boxes, axis=-1, keepdims=True), 0.0) padding_mask = tf.logical_or( tf.zeros_like(bb_x_min, dtype=tf.bool), tf.transpose(gt_invalid_mask, [0, 2, 1])) iou = tf.where(padding_mask, -tf.ones_like(iou), iou) return iou def get_non_empty_box_indices(boxes): """Get indices for non-empty boxes.""" # Selects indices if box height or width is 0. height = boxes[:, 2] - boxes[:, 0] width = boxes[:, 3] - boxes[:, 1] indices = tf.where( tf.logical_and(tf.greater(height, 0), tf.greater(width, 0))) return indices[:, 0]
26,050
36.162625
80
py
models
models-master/official/legacy/detection/utils/dataloader_utils.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Utility functions for dataloader.""" import tensorflow as tf from official.legacy.detection.utils import input_utils def process_source_id(source_id): """Processes source_id to the right format.""" if source_id.dtype == tf.string: source_id = tf.cast(tf.strings.to_number(source_id), tf.int64) with tf.control_dependencies([source_id]): source_id = tf.cond( pred=tf.equal(tf.size(input=source_id), 0), true_fn=lambda: tf.cast(tf.constant(-1), tf.int64), false_fn=lambda: tf.identity(source_id)) return source_id def pad_groundtruths_to_fixed_size(gt, n): """Pads the first dimension of groundtruths labels to the fixed size.""" gt['boxes'] = input_utils.pad_to_fixed_size(gt['boxes'], n, -1) gt['is_crowds'] = input_utils.pad_to_fixed_size(gt['is_crowds'], n, 0) gt['areas'] = input_utils.pad_to_fixed_size(gt['areas'], n, -1) gt['classes'] = input_utils.pad_to_fixed_size(gt['classes'], n, -1) return gt
1,575
37.439024
74
py
models
models-master/official/legacy/detection/utils/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/detection/utils/class_utils.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Utility functions for handling dataset object categories.""" def coco_split_class_ids(split_name): """Return the COCO class split ids based on split name and training mode. Args: split_name: The name of dataset split. Returns: class_ids: a python list of integer. """ if split_name == 'all': return [] elif split_name == 'voc': return [ 1, 2, 3, 4, 5, 6, 7, 9, 16, 17, 18, 19, 20, 21, 44, 62, 63, 64, 67, 72 ] elif split_name == 'nonvoc': return [ 8, 10, 11, 13, 14, 15, 22, 23, 24, 25, 27, 28, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 65, 70, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 84, 85, 86, 87, 88, 89, 90 ] else: raise ValueError('Invalid split name {}!!!'.format(split_name))
1,465
31.577778
79
py
models
models-master/official/legacy/detection/utils/mask_utils.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Utility functions for segmentations.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import math import cv2 import numpy as np def paste_instance_masks(masks, detected_boxes, image_height, image_width): """Paste instance masks to generate the image segmentation results. Args: masks: a numpy array of shape [N, mask_height, mask_width] representing the instance masks w.r.t. the `detected_boxes`. detected_boxes: a numpy array of shape [N, 4] representing the reference bounding boxes. image_height: an integer representing the height of the image. image_width: an integer representing the width of the image. Returns: segms: a numpy array of shape [N, image_height, image_width] representing the instance masks *pasted* on the image canvas. """ def expand_boxes(boxes, scale): """Expands an array of boxes by a given scale.""" # Reference: https://github.com/facebookresearch/Detectron/blob/master/detectron/utils/boxes.py#L227 # pylint: disable=line-too-long # The `boxes` in the reference implementation is in [x1, y1, x2, y2] form, # whereas `boxes` here is in [x1, y1, w, h] form w_half = boxes[:, 2] * .5 h_half = boxes[:, 3] * .5 x_c = boxes[:, 0] + w_half y_c = boxes[:, 1] + h_half w_half *= scale h_half *= scale boxes_exp = np.zeros(boxes.shape) boxes_exp[:, 0] = x_c - w_half boxes_exp[:, 2] = x_c + w_half boxes_exp[:, 1] = y_c - h_half boxes_exp[:, 3] = y_c + h_half return boxes_exp # Reference: https://github.com/facebookresearch/Detectron/blob/master/detectron/core/test.py#L812 # pylint: disable=line-too-long # To work around an issue with cv2.resize (it seems to automatically pad # with repeated border values), we manually zero-pad the masks by 1 pixel # prior to resizing back to the original image resolution. This prevents # "top hat" artifacts. We therefore need to expand the reference boxes by an # appropriate factor. _, mask_height, mask_width = masks.shape scale = max((mask_width + 2.0) / mask_width, (mask_height + 2.0) / mask_height) ref_boxes = expand_boxes(detected_boxes, scale) ref_boxes = ref_boxes.astype(np.int32) padded_mask = np.zeros((mask_height + 2, mask_width + 2), dtype=np.float32) segms = [] for mask_ind, mask in enumerate(masks): im_mask = np.zeros((image_height, image_width), dtype=np.uint8) # Process mask inside bounding boxes. padded_mask[1:-1, 1:-1] = mask[:, :] ref_box = ref_boxes[mask_ind, :] w = ref_box[2] - ref_box[0] + 1 h = ref_box[3] - ref_box[1] + 1 w = np.maximum(w, 1) h = np.maximum(h, 1) mask = cv2.resize(padded_mask, (w, h)) mask = np.array(mask > 0.5, dtype=np.uint8) x_0 = min(max(ref_box[0], 0), image_width) x_1 = min(max(ref_box[2] + 1, 0), image_width) y_0 = min(max(ref_box[1], 0), image_height) y_1 = min(max(ref_box[3] + 1, 0), image_height) im_mask[y_0:y_1, x_0:x_1] = mask[(y_0 - ref_box[1]):(y_1 - ref_box[1]), (x_0 - ref_box[0]):(x_1 - ref_box[0])] segms.append(im_mask) segms = np.array(segms) assert masks.shape[0] == segms.shape[0] return segms def paste_instance_masks_v2(masks, detected_boxes, image_height, image_width): """Paste instance masks to generate the image segmentation (v2). Args: masks: a numpy array of shape [N, mask_height, mask_width] representing the instance masks w.r.t. the `detected_boxes`. detected_boxes: a numpy array of shape [N, 4] representing the reference bounding boxes. image_height: an integer representing the height of the image. image_width: an integer representing the width of the image. Returns: segms: a numpy array of shape [N, image_height, image_width] representing the instance masks *pasted* on the image canvas. """ _, mask_height, mask_width = masks.shape segms = [] for i, mask in enumerate(masks): box = detected_boxes[i, :] xmin = box[0] ymin = box[1] xmax = xmin + box[2] ymax = ymin + box[3] # Sample points of the cropped mask w.r.t. the image grid. # Note that these coordinates may fall beyond the image. # Pixel clipping will happen after warping. xmin_int = int(math.floor(xmin)) xmax_int = int(math.ceil(xmax)) ymin_int = int(math.floor(ymin)) ymax_int = int(math.ceil(ymax)) alpha = box[2] / (1.0 * mask_width) beta = box[3] / (1.0 * mask_height) # pylint: disable=invalid-name # Transformation from mask pixel indices to image coordinate. M_mask_to_image = np.array([[alpha, 0, xmin], [0, beta, ymin], [0, 0, 1]], dtype=np.float32) # Transformation from image to cropped mask coordinate. M_image_to_crop = np.array( [[1, 0, -xmin_int], [0, 1, -ymin_int], [0, 0, 1]], dtype=np.float32) M = np.dot(M_image_to_crop, M_mask_to_image) # Compensate the half pixel offset that OpenCV has in the # warpPerspective implementation: the top-left pixel is sampled # at (0,0), but we want it to be at (0.5, 0.5). M = np.dot( np.dot( np.array([[1, 0, -0.5], [0, 1, -0.5], [0, 0, 1]], np.float32), M), np.array([[1, 0, 0.5], [0, 1, 0.5], [0, 0, 1]], np.float32)) # pylint: enable=invalid-name cropped_mask = cv2.warpPerspective( mask.astype(np.float32), M, (xmax_int - xmin_int, ymax_int - ymin_int)) cropped_mask = np.array(cropped_mask > 0.5, dtype=np.uint8) img_mask = np.zeros((image_height, image_width)) x0 = max(min(xmin_int, image_width), 0) x1 = max(min(xmax_int, image_width), 0) y0 = max(min(ymin_int, image_height), 0) y1 = max(min(ymax_int, image_height), 0) img_mask[y0:y1, x0:x1] = cropped_mask[(y0 - ymin_int):(y1 - ymin_int), (x0 - xmin_int):(x1 - xmin_int)] segms.append(img_mask) segms = np.array(segms) return segms
6,647
37.651163
137
py
models
models-master/official/legacy/detection/utils/input_utils.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Utility functions for input processing.""" import math import tensorflow as tf from official.legacy.detection.utils import box_utils from official.vision.utils.object_detection import preprocessor def pad_to_fixed_size(input_tensor, size, constant_values=0): """Pads data to a fixed length at the first dimension. Args: input_tensor: `Tensor` with any dimension. size: `int` number for the first dimension of output Tensor. constant_values: `int` value assigned to the paddings. Returns: `Tensor` with the first dimension padded to `size`. """ input_shape = input_tensor.get_shape().as_list() padding_shape = [] # Computes the padding length on the first dimension. padding_length = tf.maximum(0, size - tf.shape(input_tensor)[0]) assert_length = tf.Assert( tf.greater_equal(padding_length, 0), [padding_length]) with tf.control_dependencies([assert_length]): padding_shape.append(padding_length) # Copies shapes of the rest of input shape dimensions. for i in range(1, len(input_shape)): padding_shape.append(tf.shape(input=input_tensor)[i]) # Pads input tensor to the fixed first dimension. paddings = tf.cast(constant_values * tf.ones(padding_shape), input_tensor.dtype) padded_tensor = tf.concat([input_tensor, paddings], axis=0) output_shape = input_shape output_shape[0] = size padded_tensor.set_shape(output_shape) return padded_tensor def normalize_image(image, offset=(0.485, 0.456, 0.406), scale=(0.229, 0.224, 0.225)): """Normalizes the image to zero mean and unit variance.""" image = tf.image.convert_image_dtype(image, dtype=tf.float32) offset = tf.constant(offset) offset = tf.expand_dims(offset, axis=0) offset = tf.expand_dims(offset, axis=0) image -= offset scale = tf.constant(scale) scale = tf.expand_dims(scale, axis=0) scale = tf.expand_dims(scale, axis=0) image /= scale return image def compute_padded_size(desired_size, stride): """Compute the padded size given the desired size and the stride. The padded size will be the smallest rectangle, such that each dimension is the smallest multiple of the stride which is larger than the desired dimension. For example, if desired_size = (100, 200) and stride = 32, the output padded_size = (128, 224). Args: desired_size: a `Tensor` or `int` list/tuple of two elements representing [height, width] of the target output image size. stride: an integer, the stride of the backbone network. Returns: padded_size: a `Tensor` or `int` list/tuple of two elements representing [height, width] of the padded output image size. """ if isinstance(desired_size, list) or isinstance(desired_size, tuple): padded_size = [ int(math.ceil(d * 1.0 / stride) * stride) for d in desired_size ] else: padded_size = tf.cast( tf.math.ceil(tf.cast(desired_size, dtype=tf.float32) / stride) * stride, tf.int32) return padded_size def resize_and_crop_image(image, desired_size, padded_size, aug_scale_min=1.0, aug_scale_max=1.0, seed=1, method=tf.image.ResizeMethod.BILINEAR): """Resizes the input image to output size. Resize and pad images given the desired output size of the image and stride size. Here are the preprocessing steps. 1. For a given image, keep its aspect ratio and rescale the image to make it the largest rectangle to be bounded by the rectangle specified by the `desired_size`. 2. Pad the rescaled image to the padded_size. Args: image: a `Tensor` of shape [height, width, 3] representing an image. desired_size: a `Tensor` or `int` list/tuple of two elements representing [height, width] of the desired actual output image size. padded_size: a `Tensor` or `int` list/tuple of two elements representing [height, width] of the padded output image size. Padding will be applied after scaling the image to the desired_size. aug_scale_min: a `float` with range between [0, 1.0] representing minimum random scale applied to desired_size for training scale jittering. aug_scale_max: a `float` with range between [1.0, inf] representing maximum random scale applied to desired_size for training scale jittering. seed: seed for random scale jittering. method: function to resize input image to scaled image. Returns: output_image: `Tensor` of shape [height, width, 3] where [height, width] equals to `output_size`. image_info: a 2D `Tensor` that encodes the information of the image and the applied preprocessing. It is in the format of [[original_height, original_width], [desired_height, desired_width], [y_scale, x_scale], [y_offset, x_offset]], where [desired_height, desireed_width] is the actual scaled image size, and [y_scale, x_scale] is the scaling factory, which is the ratio of scaled dimension / original dimension. """ with tf.name_scope('resize_and_crop_image'): image_size = tf.cast(tf.shape(input=image)[0:2], tf.float32) random_jittering = (aug_scale_min != 1.0 or aug_scale_max != 1.0) if random_jittering: random_scale = tf.random.uniform([], aug_scale_min, aug_scale_max, seed=seed) scaled_size = tf.round(random_scale * desired_size) else: scaled_size = desired_size scale = tf.minimum(scaled_size[0] / image_size[0], scaled_size[1] / image_size[1]) scaled_size = tf.round(image_size * scale) # Computes 2D image_scale. image_scale = scaled_size / image_size # Selects non-zero random offset (x, y) if scaled image is larger than # desired_size. if random_jittering: max_offset = scaled_size - desired_size max_offset = tf.where( tf.less(max_offset, 0), tf.zeros_like(max_offset), max_offset) offset = max_offset * tf.random.uniform([ 2, ], 0, 1, seed=seed) offset = tf.cast(offset, tf.int32) else: offset = tf.zeros((2,), tf.int32) scaled_image = tf.image.resize( image, tf.cast(scaled_size, tf.int32), method=method) if random_jittering: scaled_image = scaled_image[offset[0]:offset[0] + desired_size[0], offset[1]:offset[1] + desired_size[1], :] output_image = tf.image.pad_to_bounding_box(scaled_image, 0, 0, padded_size[0], padded_size[1]) image_info = tf.stack([ image_size, tf.cast(desired_size, dtype=tf.float32), image_scale, tf.cast(offset, tf.float32) ]) return output_image, image_info def resize_and_crop_image_v2(image, short_side, long_side, padded_size, aug_scale_min=1.0, aug_scale_max=1.0, seed=1, method=tf.image.ResizeMethod.BILINEAR): """Resizes the input image to output size (Faster R-CNN style). Resize and pad images given the specified short / long side length and the stride size. Here are the preprocessing steps. 1. For a given image, keep its aspect ratio and first try to rescale the short side of the original image to `short_side`. 2. If the scaled image after 1 has a long side that exceeds `long_side`, keep the aspect ratio and rescal the long side of the image to `long_side`. 2. Pad the rescaled image to the padded_size. Args: image: a `Tensor` of shape [height, width, 3] representing an image. short_side: a scalar `Tensor` or `int` representing the desired short side to be rescaled to. long_side: a scalar `Tensor` or `int` representing the desired long side to be rescaled to. padded_size: a `Tensor` or `int` list/tuple of two elements representing [height, width] of the padded output image size. Padding will be applied after scaling the image to the desired_size. aug_scale_min: a `float` with range between [0, 1.0] representing minimum random scale applied to desired_size for training scale jittering. aug_scale_max: a `float` with range between [1.0, inf] representing maximum random scale applied to desired_size for training scale jittering. seed: seed for random scale jittering. method: function to resize input image to scaled image. Returns: output_image: `Tensor` of shape [height, width, 3] where [height, width] equals to `output_size`. image_info: a 2D `Tensor` that encodes the information of the image and the applied preprocessing. It is in the format of [[original_height, original_width], [desired_height, desired_width], [y_scale, x_scale], [y_offset, x_offset]], where [desired_height, desired_width] is the actual scaled image size, and [y_scale, x_scale] is the scaling factor, which is the ratio of scaled dimension / original dimension. """ with tf.name_scope('resize_and_crop_image_v2'): image_size = tf.cast(tf.shape(image)[0:2], tf.float32) scale_using_short_side = ( short_side / tf.math.minimum(image_size[0], image_size[1])) scale_using_long_side = ( long_side / tf.math.maximum(image_size[0], image_size[1])) scaled_size = tf.math.round(image_size * scale_using_short_side) scaled_size = tf.where( tf.math.greater( tf.math.maximum(scaled_size[0], scaled_size[1]), long_side), tf.math.round(image_size * scale_using_long_side), scaled_size) desired_size = scaled_size random_jittering = (aug_scale_min != 1.0 or aug_scale_max != 1.0) if random_jittering: random_scale = tf.random.uniform([], aug_scale_min, aug_scale_max, seed=seed) scaled_size = tf.math.round(random_scale * scaled_size) # Computes 2D image_scale. image_scale = scaled_size / image_size # Selects non-zero random offset (x, y) if scaled image is larger than # desired_size. if random_jittering: max_offset = scaled_size - desired_size max_offset = tf.where( tf.math.less(max_offset, 0), tf.zeros_like(max_offset), max_offset) offset = max_offset * tf.random.uniform([ 2, ], 0, 1, seed=seed) offset = tf.cast(offset, tf.int32) else: offset = tf.zeros((2,), tf.int32) scaled_image = tf.image.resize( image, tf.cast(scaled_size, tf.int32), method=method) if random_jittering: scaled_image = scaled_image[offset[0]:offset[0] + desired_size[0], offset[1]:offset[1] + desired_size[1], :] output_image = tf.image.pad_to_bounding_box(scaled_image, 0, 0, padded_size[0], padded_size[1]) image_info = tf.stack([ image_size, tf.cast(desired_size, dtype=tf.float32), image_scale, tf.cast(offset, tf.float32) ]) return output_image, image_info def resize_and_crop_boxes(boxes, image_scale, output_size, offset): """Resizes boxes to output size with scale and offset. Args: boxes: `Tensor` of shape [N, 4] representing ground truth boxes. image_scale: 2D float `Tensor` representing scale factors that apply to [height, width] of input image. output_size: 2D `Tensor` or `int` representing [height, width] of target output image size. offset: 2D `Tensor` representing top-left corner [y0, x0] to crop scaled boxes. Returns: boxes: `Tensor` of shape [N, 4] representing the scaled boxes. """ # Adjusts box coordinates based on image_scale and offset. boxes *= tf.tile(tf.expand_dims(image_scale, axis=0), [1, 2]) boxes -= tf.tile(tf.expand_dims(offset, axis=0), [1, 2]) # Clips the boxes. boxes = box_utils.clip_boxes(boxes, output_size) return boxes def resize_and_crop_masks(masks, image_scale, output_size, offset): """Resizes boxes to output size with scale and offset. Args: masks: `Tensor` of shape [N, H, W, 1] representing ground truth masks. image_scale: 2D float `Tensor` representing scale factors that apply to [height, width] of input image. output_size: 2D `Tensor` or `int` representing [height, width] of target output image size. offset: 2D `Tensor` representing top-left corner [y0, x0] to crop scaled boxes. Returns: masks: `Tensor` of shape [N, H, W, 1] representing the scaled masks. """ mask_size = tf.shape(input=masks)[1:3] scaled_size = tf.cast(image_scale * tf.cast(mask_size, image_scale.dtype), tf.int32) scaled_masks = tf.image.resize( masks, scaled_size, method=tf.image.ResizeMethod.NEAREST_NEIGHBOR) offset = tf.cast(offset, tf.int32) scaled_masks = scaled_masks[:, offset[0]:offset[0] + output_size[0], offset[1]:offset[1] + output_size[1], :] output_masks = tf.image.pad_to_bounding_box(scaled_masks, 0, 0, output_size[0], output_size[1]) return output_masks def random_horizontal_flip(image, boxes=None, masks=None): """Randomly flips input image and bounding boxes.""" return preprocessor.random_horizontal_flip(image, boxes, masks)
14,297
38.716667
80
py
models
models-master/official/legacy/detection/modeling/base_model.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Base Model definition.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import abc import re import tensorflow as tf from official.legacy.detection.modeling import checkpoint_utils from official.legacy.detection.modeling import learning_rates from official.legacy.detection.modeling import optimizers def _make_filter_trainable_variables_fn(frozen_variable_prefix): """Creates a function for filtering trainable varialbes.""" def _filter_trainable_variables(variables): """Filters trainable varialbes. Args: variables: a list of tf.Variable to be filtered. Returns: filtered_variables: a list of tf.Variable filtered out the frozen ones. """ # frozen_variable_prefix: a regex string specifing the prefix pattern of # the frozen variables' names. filtered_variables = [ v for v in variables if not frozen_variable_prefix or not re.match(frozen_variable_prefix, v.name) ] return filtered_variables return _filter_trainable_variables class Model(object): """Base class for model function.""" __metaclass__ = abc.ABCMeta def __init__(self, params): self._use_bfloat16 = params.architecture.use_bfloat16 if params.architecture.use_bfloat16: tf.compat.v2.keras.mixed_precision.set_global_policy('mixed_bfloat16') # Optimization. self._optimizer_fn = optimizers.OptimizerFactory(params.train.optimizer) self._learning_rate = learning_rates.learning_rate_generator( params.train.total_steps, params.train.learning_rate) self._frozen_variable_prefix = params.train.frozen_variable_prefix self._regularization_var_regex = params.train.regularization_variable_regex self._l2_weight_decay = params.train.l2_weight_decay # Checkpoint restoration. self._checkpoint = params.train.checkpoint.as_dict() # Summary. self._enable_summary = params.enable_summary self._model_dir = params.model_dir @abc.abstractmethod def build_outputs(self, inputs, mode): """Build the graph of the forward path.""" pass @abc.abstractmethod def build_model(self, params, mode): """Build the model object.""" pass @abc.abstractmethod def build_loss_fn(self): """Build the model object.""" pass def post_processing(self, labels, outputs): """Post-processing function.""" return labels, outputs def model_outputs(self, inputs, mode): """Build the model outputs.""" return self.build_outputs(inputs, mode) def build_optimizer(self): """Returns train_op to optimize total loss.""" # Sets up the optimizer. return self._optimizer_fn(self._learning_rate) def make_filter_trainable_variables_fn(self): """Creates a function for filtering trainable varialbes.""" return _make_filter_trainable_variables_fn(self._frozen_variable_prefix) def weight_decay_loss(self, trainable_variables): reg_variables = [ v for v in trainable_variables if self._regularization_var_regex is None or re.match(self._regularization_var_regex, v.name) ] return self._l2_weight_decay * tf.add_n( [tf.nn.l2_loss(v) for v in reg_variables]) def make_restore_checkpoint_fn(self): """Returns scaffold function to restore parameters from v1 checkpoint.""" if 'skip_checkpoint_variables' in self._checkpoint: skip_regex = self._checkpoint['skip_checkpoint_variables'] else: skip_regex = None return checkpoint_utils.make_restore_checkpoint_fn( self._checkpoint['path'], prefix=self._checkpoint['prefix'], skip_regex=skip_regex) def eval_metrics(self): """Returns tuple of metric function and its inputs for evaluation.""" raise NotImplementedError('Unimplemented eval_metrics')
4,439
31.647059
79
py
models
models-master/official/legacy/detection/modeling/losses.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Losses used for detection models.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from absl import logging import tensorflow as tf def focal_loss(logits, targets, alpha, gamma, normalizer): """Compute the focal loss between `logits` and the golden `target` values. Focal loss = -(1-pt)^gamma * log(pt) where pt is the probability of being classified to the true class. Args: logits: A float32 tensor of size [batch, height_in, width_in, num_predictions]. targets: A float32 tensor of size [batch, height_in, width_in, num_predictions]. alpha: A float32 scalar multiplying alpha to the loss from positive examples and (1-alpha) to the loss from negative examples. gamma: A float32 scalar modulating loss from hard and easy examples. normalizer: A float32 scalar normalizes the total loss from all examples. Returns: loss: A float32 Tensor of size [batch, height_in, width_in, num_predictions] representing normalized loss on the prediction map. """ with tf.name_scope('focal_loss'): positive_label_mask = tf.math.equal(targets, 1.0) cross_entropy = ( tf.nn.sigmoid_cross_entropy_with_logits(labels=targets, logits=logits)) # Below are comments/derivations for computing modulator. # For brevity, let x = logits, z = targets, r = gamma, and p_t = sigmod(x) # for positive samples and 1 - sigmoid(x) for negative examples. # # The modulator, defined as (1 - P_t)^r, is a critical part in focal loss # computation. For r > 0, it puts more weights on hard examples, and less # weights on easier ones. However if it is directly computed as (1 - P_t)^r, # its back-propagation is not stable when r < 1. The implementation here # resolves the issue. # # For positive samples (labels being 1), # (1 - p_t)^r # = (1 - sigmoid(x))^r # = (1 - (1 / (1 + exp(-x))))^r # = (exp(-x) / (1 + exp(-x)))^r # = exp(log((exp(-x) / (1 + exp(-x)))^r)) # = exp(r * log(exp(-x)) - r * log(1 + exp(-x))) # = exp(- r * x - r * log(1 + exp(-x))) # # For negative samples (labels being 0), # (1 - p_t)^r # = (sigmoid(x))^r # = (1 / (1 + exp(-x)))^r # = exp(log((1 / (1 + exp(-x)))^r)) # = exp(-r * log(1 + exp(-x))) # # Therefore one unified form for positive (z = 1) and negative (z = 0) # samples is: # (1 - p_t)^r = exp(-r * z * x - r * log(1 + exp(-x))). neg_logits = -1.0 * logits modulator = tf.math.exp(gamma * targets * neg_logits - gamma * tf.math.log1p(tf.math.exp(neg_logits))) loss = modulator * cross_entropy weighted_loss = tf.where(positive_label_mask, alpha * loss, (1.0 - alpha) * loss) weighted_loss /= normalizer return weighted_loss class RpnScoreLoss(object): """Region Proposal Network score loss function.""" def __init__(self, params): self._rpn_batch_size_per_im = params.rpn_batch_size_per_im self._binary_crossentropy = tf.keras.losses.BinaryCrossentropy( reduction=tf.keras.losses.Reduction.SUM, from_logits=True) def __call__(self, score_outputs, labels): """Computes total RPN detection loss. Computes total RPN detection loss including box and score from all levels. Args: score_outputs: an OrderDict with keys representing levels and values representing scores in [batch_size, height, width, num_anchors]. labels: the dictionary that returned from dataloader that includes groundturth targets. Returns: rpn_score_loss: a scalar tensor representing total score loss. """ with tf.name_scope('rpn_loss'): levels = sorted(score_outputs.keys()) score_losses = [] for level in levels: score_losses.append( self._rpn_score_loss( score_outputs[level], labels[level], normalizer=tf.cast( tf.shape(score_outputs[level])[0] * self._rpn_batch_size_per_im, dtype=tf.float32))) # Sums per level losses to total loss. return tf.math.add_n(score_losses) def _rpn_score_loss(self, score_outputs, score_targets, normalizer=1.0): """Computes score loss.""" # score_targets has three values: # (1) score_targets[i]=1, the anchor is a positive sample. # (2) score_targets[i]=0, negative. # (3) score_targets[i]=-1, the anchor is don't care (ignore). with tf.name_scope('rpn_score_loss'): mask = tf.math.logical_or(tf.math.equal(score_targets, 1), tf.math.equal(score_targets, 0)) score_targets = tf.math.maximum(score_targets, tf.zeros_like(score_targets)) score_targets = tf.expand_dims(score_targets, axis=-1) score_outputs = tf.expand_dims(score_outputs, axis=-1) score_loss = self._binary_crossentropy( score_targets, score_outputs, sample_weight=mask) score_loss /= normalizer return score_loss class RpnBoxLoss(object): """Region Proposal Network box regression loss function.""" def __init__(self, params): logging.info('RpnBoxLoss huber_loss_delta %s', params.huber_loss_delta) # The delta is typically around the mean value of regression target. # for instances, the regression targets of 512x512 input with 6 anchors on # P2-P6 pyramid is about [0.1, 0.1, 0.2, 0.2]. self._huber_loss = tf.keras.losses.Huber( delta=params.huber_loss_delta, reduction=tf.keras.losses.Reduction.SUM) def __call__(self, box_outputs, labels): """Computes total RPN detection loss. Computes total RPN detection loss including box and score from all levels. Args: box_outputs: an OrderDict with keys representing levels and values representing box regression targets in [batch_size, height, width, num_anchors * 4]. labels: the dictionary that returned from dataloader that includes groundturth targets. Returns: rpn_box_loss: a scalar tensor representing total box regression loss. """ with tf.name_scope('rpn_loss'): levels = sorted(box_outputs.keys()) box_losses = [] for level in levels: box_losses.append(self._rpn_box_loss(box_outputs[level], labels[level])) # Sum per level losses to total loss. return tf.add_n(box_losses) def _rpn_box_loss(self, box_outputs, box_targets, normalizer=1.0): """Computes box regression loss.""" with tf.name_scope('rpn_box_loss'): mask = tf.cast(tf.not_equal(box_targets, 0.0), dtype=tf.float32) box_targets = tf.expand_dims(box_targets, axis=-1) box_outputs = tf.expand_dims(box_outputs, axis=-1) box_loss = self._huber_loss(box_targets, box_outputs, sample_weight=mask) # The loss is normalized by the sum of non-zero weights and additional # normalizer provided by the function caller. Using + 0.01 here to avoid # division by zero. box_loss /= normalizer * (tf.reduce_sum(mask) + 0.01) return box_loss class OlnRpnCenterLoss(object): """Object Localization Network RPN centerness regression loss function.""" def __init__(self): self._l1_loss = tf.keras.losses.MeanAbsoluteError( reduction=tf.keras.losses.Reduction.SUM) def __call__(self, center_outputs, labels): """Computes total RPN centerness regression loss. Computes total RPN centerness score regression loss from all levels. Args: center_outputs: an OrderDict with keys representing levels and values representing anchor centerness regression targets in [batch_size, height, width, num_anchors * 4]. labels: the dictionary that returned from dataloader that includes groundturth targets. Returns: rpn_center_loss: a scalar tensor representing total centerness regression loss. """ with tf.name_scope('rpn_loss'): # Normalizer. levels = sorted(center_outputs.keys()) num_valid = 0 # 0<pos<1, neg=0, ign=-1 for level in levels: num_valid += tf.reduce_sum(tf.cast( tf.greater(labels[level], -1.0), tf.float32)) # in and out of box num_valid += 1e-12 # Centerness loss over multi levels. center_losses = [] for level in levels: center_losses.append( self._rpn_center_l1_loss( center_outputs[level], labels[level], normalizer=num_valid)) # Sum per level losses to total loss. return tf.add_n(center_losses) def _rpn_center_l1_loss(self, center_outputs, center_targets, normalizer=1.0): """Computes centerness regression loss.""" # for instances, the regression targets of 512x512 input with 6 anchors on # P2-P6 pyramid is about [0.1, 0.1, 0.2, 0.2]. with tf.name_scope('rpn_center_loss'): # mask = tf.greater(center_targets, 0.0) # inside box only. mask = tf.greater(center_targets, -1.0) # in and out of box. center_targets = tf.maximum(center_targets, tf.zeros_like(center_targets)) center_outputs = tf.sigmoid(center_outputs) center_targets = tf.expand_dims(center_targets, -1) center_outputs = tf.expand_dims(center_outputs, -1) mask = tf.cast(mask, dtype=tf.float32) center_loss = self._l1_loss(center_targets, center_outputs, sample_weight=mask) center_loss /= normalizer return center_loss class OlnRpnIoULoss(object): """Object Localization Network RPN box-lrtb regression iou loss function.""" def __call__(self, box_outputs, labels, center_targets): """Computes total RPN detection loss. Computes total RPN box regression loss from all levels. Args: box_outputs: an OrderDict with keys representing levels and values representing box regression targets in [batch_size, height, width, num_anchors * 4]. last channel: (left, right, top, bottom). labels: the dictionary that returned from dataloader that includes groundturth targets (left, right, top, bottom). center_targets: valid_target mask. Returns: rpn_iou_loss: a scalar tensor representing total box regression loss. """ with tf.name_scope('rpn_loss'): # Normalizer. levels = sorted(box_outputs.keys()) normalizer = 0. for level in levels: # center_targets pos>0, neg=0, ign=-1. mask_ = tf.cast(tf.logical_and( tf.greater(center_targets[level][..., 0], 0.0), tf.greater(tf.reduce_min(labels[level], -1), 0.0)), tf.float32) normalizer += tf.reduce_sum(mask_) normalizer += 1e-8 # iou_loss over multi levels. iou_losses = [] for level in levels: iou_losses.append( self._rpn_iou_loss( box_outputs[level], labels[level], center_weight=center_targets[level][..., 0], normalizer=normalizer)) # Sum per level losses to total loss. return tf.add_n(iou_losses) def _rpn_iou_loss(self, box_outputs, box_targets, center_weight=None, normalizer=1.0): """Computes box regression loss.""" # for instances, the regression targets of 512x512 input with 6 anchors on # P2-P6 pyramid is about [0.1, 0.1, 0.2, 0.2]. with tf.name_scope('rpn_iou_loss'): mask = tf.logical_and( tf.greater(center_weight, 0.0), tf.greater(tf.reduce_min(box_targets, -1), 0.0)) pred_left = box_outputs[..., 0] pred_right = box_outputs[..., 1] pred_top = box_outputs[..., 2] pred_bottom = box_outputs[..., 3] gt_left = box_targets[..., 0] gt_right = box_targets[..., 1] gt_top = box_targets[..., 2] gt_bottom = box_targets[..., 3] inter_width = (tf.minimum(pred_left, gt_left) + tf.minimum(pred_right, gt_right)) inter_height = (tf.minimum(pred_top, gt_top) + tf.minimum(pred_bottom, gt_bottom)) inter_area = inter_width * inter_height union_area = ((pred_left + pred_right) * (pred_top + pred_bottom) + (gt_left + gt_right) * (gt_top + gt_bottom) - inter_area) iou = inter_area / (union_area + 1e-8) mask_ = tf.cast(mask, tf.float32) iou = tf.clip_by_value(iou, clip_value_min=1e-8, clip_value_max=1.0) neg_log_iou = -tf.math.log(iou) iou_loss = tf.reduce_sum(neg_log_iou * mask_) iou_loss /= normalizer return iou_loss class FastrcnnClassLoss(object): """Fast R-CNN classification loss function.""" def __init__(self): self._categorical_crossentropy = tf.keras.losses.CategoricalCrossentropy( reduction=tf.keras.losses.Reduction.SUM, from_logits=True) def __call__(self, class_outputs, class_targets): """Computes the class loss (Fast-RCNN branch) of Mask-RCNN. This function implements the classification loss of the Fast-RCNN. The classification loss is softmax on all RoIs. Reference: https://github.com/facebookresearch/Detectron/blob/master/detectron/modeling/fast_rcnn_heads.py # pylint: disable=line-too-long Args: class_outputs: a float tensor representing the class prediction for each box with a shape of [batch_size, num_boxes, num_classes]. class_targets: a float tensor representing the class label for each box with a shape of [batch_size, num_boxes]. Returns: a scalar tensor representing total class loss. """ with tf.name_scope('fast_rcnn_loss'): batch_size, num_boxes, num_classes = class_outputs.get_shape().as_list() class_targets = tf.cast(class_targets, dtype=tf.int32) class_targets_one_hot = tf.one_hot(class_targets, num_classes) return self._fast_rcnn_class_loss(class_outputs, class_targets_one_hot, normalizer=batch_size * num_boxes / 2.0) def _fast_rcnn_class_loss(self, class_outputs, class_targets_one_hot, normalizer): """Computes classification loss.""" with tf.name_scope('fast_rcnn_class_loss'): class_loss = self._categorical_crossentropy(class_targets_one_hot, class_outputs) class_loss /= normalizer return class_loss class FastrcnnBoxLoss(object): """Fast R-CNN box regression loss function.""" def __init__(self, params): logging.info('FastrcnnBoxLoss huber_loss_delta %s', params.huber_loss_delta) # The delta is typically around the mean value of regression target. # for instances, the regression targets of 512x512 input with 6 anchors on # P2-P6 pyramid is about [0.1, 0.1, 0.2, 0.2]. self._huber_loss = tf.keras.losses.Huber( delta=params.huber_loss_delta, reduction=tf.keras.losses.Reduction.SUM) def __call__(self, box_outputs, class_targets, box_targets): """Computes the box loss (Fast-RCNN branch) of Mask-RCNN. This function implements the box regression loss of the Fast-RCNN. As the `box_outputs` produces `num_classes` boxes for each RoI, the reference model expands `box_targets` to match the shape of `box_outputs` and selects only the target that the RoI has a maximum overlap. (Reference: https://github.com/facebookresearch/Detectron/blob/master/detectron/roi_data/fast_rcnn.py) # pylint: disable=line-too-long Instead, this function selects the `box_outputs` by the `class_targets` so that it doesn't expand `box_targets`. The box loss is smooth L1-loss on only positive samples of RoIs. Reference: https://github.com/facebookresearch/Detectron/blob/master/detectron/modeling/fast_rcnn_heads.py # pylint: disable=line-too-long Args: box_outputs: a float tensor representing the box prediction for each box with a shape of [batch_size, num_boxes, num_classes * 4]. class_targets: a float tensor representing the class label for each box with a shape of [batch_size, num_boxes]. box_targets: a float tensor representing the box label for each box with a shape of [batch_size, num_boxes, 4]. Returns: box_loss: a scalar tensor representing total box regression loss. """ with tf.name_scope('fast_rcnn_loss'): class_targets = tf.cast(class_targets, dtype=tf.int32) # Selects the box from `box_outputs` based on `class_targets`, with which # the box has the maximum overlap. (batch_size, num_rois, num_class_specific_boxes) = box_outputs.get_shape().as_list() num_classes = num_class_specific_boxes // 4 box_outputs = tf.reshape(box_outputs, [batch_size, num_rois, num_classes, 4]) box_indices = tf.reshape( class_targets + tf.tile( tf.expand_dims( tf.range(batch_size) * num_rois * num_classes, 1), [1, num_rois]) + tf.tile( tf.expand_dims(tf.range(num_rois) * num_classes, 0), [batch_size, 1]), [-1]) box_outputs = tf.matmul( tf.one_hot( box_indices, batch_size * num_rois * num_classes, dtype=box_outputs.dtype), tf.reshape(box_outputs, [-1, 4])) box_outputs = tf.reshape(box_outputs, [batch_size, -1, 4]) return self._fast_rcnn_box_loss(box_outputs, box_targets, class_targets) def _fast_rcnn_box_loss(self, box_outputs, box_targets, class_targets, normalizer=1.0): """Computes box regression loss.""" with tf.name_scope('fast_rcnn_box_loss'): mask = tf.tile(tf.expand_dims(tf.greater(class_targets, 0), axis=2), [1, 1, 4]) mask = tf.cast(mask, dtype=tf.float32) box_targets = tf.expand_dims(box_targets, axis=-1) box_outputs = tf.expand_dims(box_outputs, axis=-1) box_loss = self._huber_loss(box_targets, box_outputs, sample_weight=mask) # The loss is normalized by the number of ones in mask, # additianal normalizer provided by the user and using 0.01 here to avoid # division by 0. box_loss /= normalizer * (tf.reduce_sum(mask) + 0.01) return box_loss class OlnBoxScoreLoss(object): """Object Localization Network Box-Iou scoring function.""" def __init__(self, params): self._ignore_threshold = params.ignore_threshold self._l1_loss = tf.keras.losses.MeanAbsoluteError( reduction=tf.keras.losses.Reduction.SUM) def __call__(self, score_outputs, score_targets): """Computes the class loss (Fast-RCNN branch) of Mask-RCNN. This function implements the classification loss of the Fast-RCNN. The classification loss is softmax on all RoIs. Reference: https://github.com/facebookresearch/Detectron/blob/master/detectron/modeling/fast_rcnn_heads.py # pylint: disable=line-too-long Args: score_outputs: a float tensor representing the class prediction for each box with a shape of [batch_size, num_boxes, num_classes]. score_targets: a float tensor representing the class label for each box with a shape of [batch_size, num_boxes]. Returns: a scalar tensor representing total score loss. """ with tf.name_scope('fast_rcnn_loss'): score_outputs = tf.squeeze(score_outputs, -1) mask = tf.greater(score_targets, self._ignore_threshold) num_valid = tf.reduce_sum(tf.cast(mask, tf.float32)) score_targets = tf.maximum(score_targets, tf.zeros_like(score_targets)) score_outputs = tf.sigmoid(score_outputs) score_targets = tf.expand_dims(score_targets, -1) score_outputs = tf.expand_dims(score_outputs, -1) mask = tf.cast(mask, dtype=tf.float32) score_loss = self._l1_loss(score_targets, score_outputs, sample_weight=mask) score_loss /= (num_valid + 1e-10) return score_loss class MaskrcnnLoss(object): """Mask R-CNN instance segmentation mask loss function.""" def __init__(self): self._binary_crossentropy = tf.keras.losses.BinaryCrossentropy( reduction=tf.keras.losses.Reduction.SUM, from_logits=True) def __call__(self, mask_outputs, mask_targets, select_class_targets): """Computes the mask loss of Mask-RCNN. This function implements the mask loss of Mask-RCNN. As the `mask_outputs` produces `num_classes` masks for each RoI, the reference model expands `mask_targets` to match the shape of `mask_outputs` and selects only the target that the RoI has a maximum overlap. (Reference: https://github.com/facebookresearch/Detectron/blob/master/detectron/roi_data/mask_rcnn.py) # pylint: disable=line-too-long Instead, this implementation selects the `mask_outputs` by the `class_targets` so that it doesn't expand `mask_targets`. Note that the selection logic is done in the post-processing of mask_rcnn_fn in mask_rcnn_architecture.py. Args: mask_outputs: a float tensor representing the prediction for each mask, with a shape of [batch_size, num_masks, mask_height, mask_width]. mask_targets: a float tensor representing the binary mask of ground truth labels for each mask with a shape of [batch_size, num_masks, mask_height, mask_width]. select_class_targets: a tensor with a shape of [batch_size, num_masks], representing the foreground mask targets. Returns: mask_loss: a float tensor representing total mask loss. """ with tf.name_scope('mask_rcnn_loss'): (batch_size, num_masks, mask_height, mask_width) = mask_outputs.get_shape().as_list() weights = tf.tile( tf.reshape(tf.greater(select_class_targets, 0), [batch_size, num_masks, 1, 1]), [1, 1, mask_height, mask_width]) weights = tf.cast(weights, dtype=tf.float32) mask_targets = tf.expand_dims(mask_targets, axis=-1) mask_outputs = tf.expand_dims(mask_outputs, axis=-1) mask_loss = self._binary_crossentropy(mask_targets, mask_outputs, sample_weight=weights) # The loss is normalized by the number of 1's in weights and # + 0.01 is used to avoid division by zero. return mask_loss / (tf.reduce_sum(weights) + 0.01) class RetinanetClassLoss(object): """RetinaNet class loss.""" def __init__(self, params, num_classes): self._num_classes = num_classes self._focal_loss_alpha = params.focal_loss_alpha self._focal_loss_gamma = params.focal_loss_gamma def __call__(self, cls_outputs, labels, num_positives): """Computes total detection loss. Computes total detection loss including box and class loss from all levels. Args: cls_outputs: an OrderDict with keys representing levels and values representing logits in [batch_size, height, width, num_anchors * num_classes]. labels: the dictionary that returned from dataloader that includes class groundturth targets. num_positives: number of positive examples in the minibatch. Returns: an integar tensor representing total class loss. """ # Sums all positives in a batch for normalization and avoids zero # num_positives_sum, which would lead to inf loss during training num_positives_sum = tf.reduce_sum(input_tensor=num_positives) + 1.0 cls_losses = [] for level in cls_outputs.keys(): cls_losses.append(self.class_loss( cls_outputs[level], labels[level], num_positives_sum)) # Sums per level losses to total loss. return tf.add_n(cls_losses) def class_loss(self, cls_outputs, cls_targets, num_positives, ignore_label=-2): """Computes RetinaNet classification loss.""" # Onehot encoding for classification labels. cls_targets_one_hot = tf.one_hot(cls_targets, self._num_classes) bs, height, width, _, _ = cls_targets_one_hot.get_shape().as_list() cls_targets_one_hot = tf.reshape(cls_targets_one_hot, [bs, height, width, -1]) loss = focal_loss(tf.cast(cls_outputs, dtype=tf.float32), tf.cast(cls_targets_one_hot, dtype=tf.float32), self._focal_loss_alpha, self._focal_loss_gamma, num_positives) ignore_loss = tf.where( tf.equal(cls_targets, ignore_label), tf.zeros_like(cls_targets, dtype=tf.float32), tf.ones_like(cls_targets, dtype=tf.float32), ) ignore_loss = tf.expand_dims(ignore_loss, -1) ignore_loss = tf.tile(ignore_loss, [1, 1, 1, 1, self._num_classes]) ignore_loss = tf.reshape(ignore_loss, tf.shape(input=loss)) return tf.reduce_sum(input_tensor=ignore_loss * loss) class RetinanetBoxLoss(object): """RetinaNet box loss.""" def __init__(self, params): self._huber_loss = tf.keras.losses.Huber( delta=params.huber_loss_delta, reduction=tf.keras.losses.Reduction.SUM) def __call__(self, box_outputs, labels, num_positives): """Computes box detection loss. Computes total detection loss including box and class loss from all levels. Args: box_outputs: an OrderDict with keys representing levels and values representing box regression targets in [batch_size, height, width, num_anchors * 4]. labels: the dictionary that returned from dataloader that includes box groundturth targets. num_positives: number of positive examples in the minibatch. Returns: an integer tensor representing total box regression loss. """ # Sums all positives in a batch for normalization and avoids zero # num_positives_sum, which would lead to inf loss during training num_positives_sum = tf.reduce_sum(input_tensor=num_positives) + 1.0 box_losses = [] for level in box_outputs.keys(): box_targets_l = labels[level] box_losses.append( self.box_loss(box_outputs[level], box_targets_l, num_positives_sum)) # Sums per level losses to total loss. return tf.add_n(box_losses) def box_loss(self, box_outputs, box_targets, num_positives): """Computes RetinaNet box regression loss.""" # The delta is typically around the mean value of regression target. # for instances, the regression targets of 512x512 input with 6 anchors on # P3-P7 pyramid is about [0.1, 0.1, 0.2, 0.2]. normalizer = num_positives * 4.0 mask = tf.cast(tf.not_equal(box_targets, 0.0), dtype=tf.float32) box_targets = tf.expand_dims(box_targets, axis=-1) box_outputs = tf.expand_dims(box_outputs, axis=-1) box_loss = self._huber_loss(box_targets, box_outputs, sample_weight=mask) box_loss /= normalizer return box_loss class ShapemaskMseLoss(object): """ShapeMask mask Mean Squared Error loss function wrapper.""" def __call__(self, probs, labels, valid_mask): """Compute instance segmentation loss. Args: probs: A Tensor of shape [batch_size * num_points, height, width, num_classes]. The logits are not necessarily between 0 and 1. labels: A float32/float16 Tensor of shape [batch_size, num_instances, mask_size, mask_size], where mask_size = mask_crop_size * gt_upsample_scale for fine mask, or mask_crop_size for coarse masks and shape priors. valid_mask: a binary mask indicating valid training masks. Returns: loss: an float tensor representing total mask classification loss. """ with tf.name_scope('shapemask_prior_loss'): batch_size, num_instances = valid_mask.get_shape().as_list()[:2] diff = (tf.cast(labels, dtype=tf.float32) - tf.cast(probs, dtype=tf.float32)) diff *= tf.cast( tf.reshape(valid_mask, [batch_size, num_instances, 1, 1]), tf.float32) # Adding 0.001 in the denominator to avoid division by zero. loss = tf.nn.l2_loss(diff) / (tf.reduce_sum(labels) + 0.001) return loss class ShapemaskLoss(object): """ShapeMask mask loss function wrapper.""" def __init__(self): self._binary_crossentropy = tf.keras.losses.BinaryCrossentropy( reduction=tf.keras.losses.Reduction.SUM, from_logits=True) def __call__(self, logits, labels, valid_mask): """ShapeMask mask cross entropy loss function wrapper. Args: logits: A Tensor of shape [batch_size * num_instances, height, width, num_classes]. The logits are not necessarily between 0 and 1. labels: A float16/float32 Tensor of shape [batch_size, num_instances, mask_size, mask_size], where mask_size = mask_crop_size * gt_upsample_scale for fine mask, or mask_crop_size for coarse masks and shape priors. valid_mask: a binary mask of shape [batch_size, num_instances] indicating valid training masks. Returns: loss: an float tensor representing total mask classification loss. """ with tf.name_scope('shapemask_loss'): batch_size, num_instances = valid_mask.get_shape().as_list()[:2] labels = tf.cast(labels, tf.float32) logits = tf.cast(logits, tf.float32) loss = self._binary_crossentropy(labels, logits) loss *= tf.cast(tf.reshape( valid_mask, [batch_size, num_instances, 1, 1]), loss.dtype) # Adding 0.001 in the denominator to avoid division by zero. loss = tf.reduce_sum(loss) / (tf.reduce_sum(labels) + 0.001) return loss
30,203
40.603306
186
py
models
models-master/official/legacy/detection/modeling/optimizers.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Optimizers.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import functools import tensorflow as tf class OptimizerFactory(object): """Class to generate optimizer function.""" def __init__(self, params): """Creates optimized based on the specified flags.""" if params.type == 'momentum': self._optimizer = functools.partial( tf.keras.optimizers.SGD, momentum=params.momentum, nesterov=params.nesterov) elif params.type == 'adam': self._optimizer = tf.keras.optimizers.Adam elif params.type == 'adadelta': self._optimizer = tf.keras.optimizers.Adadelta elif params.type == 'adagrad': self._optimizer = tf.keras.optimizers.Adagrad elif params.type == 'rmsprop': self._optimizer = functools.partial( tf.keras.optimizers.RMSprop, momentum=params.momentum) else: raise ValueError('Unsupported optimizer type `{}`.'.format(params.type)) def __call__(self, learning_rate): return self._optimizer(learning_rate=learning_rate)
1,716
33.34
78
py
models
models-master/official/legacy/detection/modeling/learning_rates.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Learning rate schedule.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import numpy as np import tensorflow as tf from official.modeling.hyperparams import params_dict class StepLearningRateWithLinearWarmup( tf.keras.optimizers.schedules.LearningRateSchedule): """Class to generate learning rate tensor.""" def __init__(self, total_steps, params): """Creates the step learning rate tensor with linear warmup.""" super(StepLearningRateWithLinearWarmup, self).__init__() self._total_steps = total_steps assert isinstance(params, (dict, params_dict.ParamsDict)) if isinstance(params, dict): params = params_dict.ParamsDict(params) self._params = params def __call__(self, global_step): warmup_lr = self._params.warmup_learning_rate warmup_steps = self._params.warmup_steps init_lr = self._params.init_learning_rate lr_levels = self._params.learning_rate_levels lr_steps = self._params.learning_rate_steps linear_warmup = ( warmup_lr + tf.cast(global_step, dtype=tf.float32) / warmup_steps * (init_lr - warmup_lr)) learning_rate = tf.where(global_step < warmup_steps, linear_warmup, init_lr) for next_learning_rate, start_step in zip(lr_levels, lr_steps): learning_rate = tf.where(global_step >= start_step, next_learning_rate, learning_rate) return learning_rate def get_config(self): return {'_params': self._params.as_dict()} class CosineLearningRateWithLinearWarmup( tf.keras.optimizers.schedules.LearningRateSchedule): """Class to generate learning rate tensor.""" def __init__(self, total_steps, params): """Creates the cosine learning rate tensor with linear warmup.""" super(CosineLearningRateWithLinearWarmup, self).__init__() self._total_steps = total_steps assert isinstance(params, (dict, params_dict.ParamsDict)) if isinstance(params, dict): params = params_dict.ParamsDict(params) self._params = params def __call__(self, global_step): global_step = tf.cast(global_step, dtype=tf.float32) warmup_lr = self._params.warmup_learning_rate warmup_steps = self._params.warmup_steps init_lr = self._params.init_learning_rate total_steps = self._total_steps linear_warmup = ( warmup_lr + global_step / warmup_steps * (init_lr - warmup_lr)) cosine_learning_rate = ( init_lr * (tf.cos(np.pi * (global_step - warmup_steps) / (total_steps - warmup_steps)) + 1.0) / 2.0) learning_rate = tf.where(global_step < warmup_steps, linear_warmup, cosine_learning_rate) return learning_rate def get_config(self): return {'_params': self._params.as_dict()} def learning_rate_generator(total_steps, params): """The learning rate function generator.""" if params.type == 'step': return StepLearningRateWithLinearWarmup(total_steps, params) elif params.type == 'cosine': return CosineLearningRateWithLinearWarmup(total_steps, params) else: raise ValueError('Unsupported learning rate type: {}.'.format(params.type))
3,800
37.393939
80
py
models
models-master/official/legacy/detection/modeling/checkpoint_utils.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Util functions for loading checkpoints. Especially for loading Tensorflow 1.x checkpoint to Tensorflow 2.x (keras) model. """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import re from absl import logging import tensorflow as tf def _build_assignment_map(keras_model, prefix='', skip_variables_regex=None, var_to_shape_map=None): """Builds the variable assignment map. Compute an assignment mapping for loading older checkpoints into a Keras model. Variable names are remapped from the original TPUEstimator model to the new Keras name. Args: keras_model: tf.keras.Model object to provide variables to assign. prefix: prefix in the variable name to be remove for alignment with names in the checkpoint. skip_variables_regex: regular expression to math the names of variables that do not need to be assign. var_to_shape_map: variable name to shape mapping from the checkpoint. Returns: The variable assignment map. """ assignment_map = {} checkpoint_names = [] if var_to_shape_map: # pylint: disable=g-long-lambda checkpoint_names = list( filter( lambda x: not x.endswith('Momentum') and not x.endswith( 'global_step'), var_to_shape_map.keys())) # pylint: enable=g-long-lambda logging.info('Number of variables in the checkpoint %d', len(checkpoint_names)) for var in keras_model.variables: var_name = var.name if skip_variables_regex and re.match(skip_variables_regex, var_name): continue # Trim the index of the variable. if ':' in var_name: var_name = var_name[:var_name.rindex(':')] if var_name.startswith(prefix): var_name = var_name[len(prefix):] if not var_to_shape_map: assignment_map[var_name] = var continue # Match name with variables in the checkpoint. # pylint: disable=cell-var-from-loop match_names = list(filter(lambda x: x.endswith(var_name), checkpoint_names)) # pylint: enable=cell-var-from-loop try: if match_names: assert len(match_names) == 1, 'more then on matches for {}: {}'.format( var_name, match_names) checkpoint_names.remove(match_names[0]) assignment_map[match_names[0]] = var else: logging.info('Error not found var name: %s', var_name) except Exception as e: logging.info('Error removing the match_name: %s', match_names) logging.info('Exception: %s', e) raise logging.info('Found matching variable in checkpoint: %d', len(assignment_map)) return assignment_map def _get_checkpoint_map(checkpoint_path): reader = tf.train.load_checkpoint(checkpoint_path) return reader.get_variable_to_shape_map() def make_restore_checkpoint_fn(checkpoint_path, prefix='', skip_regex=None): """Returns scaffold function to restore parameters from v1 checkpoint. Args: checkpoint_path: path of the checkpoint folder or file. Example 1: '/path/to/model_dir/' Example 2: '/path/to/model.ckpt-22500' prefix: prefix in the variable name to be remove for alignment with names in the checkpoint. skip_regex: regular expression to math the names of variables that do not need to be assign. Returns: Callable[tf.kears.Model] -> void. Fn to load v1 checkpoint to keras model. """ def _restore_checkpoint_fn(keras_model): """Loads pretrained model through scaffold function.""" if not checkpoint_path: logging.info('checkpoint_path is empty') return var_prefix = prefix if prefix and not prefix.endswith('/'): var_prefix += '/' var_to_shape_map = _get_checkpoint_map(checkpoint_path) assert var_to_shape_map, 'var_to_shape_map should not be empty' vars_to_load = _build_assignment_map( keras_model, prefix=var_prefix, skip_variables_regex=skip_regex, var_to_shape_map=var_to_shape_map) if not vars_to_load: raise ValueError('Variables to load is empty.') tf.compat.v1.train.init_from_checkpoint(checkpoint_path, vars_to_load) return _restore_checkpoint_fn
4,862
33.006993
80
py
models
models-master/official/legacy/detection/modeling/maskrcnn_model.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Model defination for the Mask R-CNN Model.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from official.legacy.detection.dataloader import anchor from official.legacy.detection.dataloader import mode_keys from official.legacy.detection.evaluation import factory as eval_factory from official.legacy.detection.modeling import base_model from official.legacy.detection.modeling import losses from official.legacy.detection.modeling.architecture import factory from official.legacy.detection.ops import postprocess_ops from official.legacy.detection.ops import roi_ops from official.legacy.detection.ops import spatial_transform_ops from official.legacy.detection.ops import target_ops from official.legacy.detection.utils import box_utils class MaskrcnnModel(base_model.Model): """Mask R-CNN model function.""" def __init__(self, params): super(MaskrcnnModel, self).__init__(params) # For eval metrics. self._params = params self._keras_model = None self._include_mask = params.architecture.include_mask # Architecture generators. self._backbone_fn = factory.backbone_generator(params) self._fpn_fn = factory.multilevel_features_generator(params) self._rpn_head_fn = factory.rpn_head_generator(params) self._generate_rois_fn = roi_ops.ROIGenerator(params.roi_proposal) self._sample_rois_fn = target_ops.ROISampler(params.roi_sampling) self._sample_masks_fn = target_ops.MaskSampler( params.architecture.mask_target_size, params.mask_sampling.num_mask_samples_per_image) self._frcnn_head_fn = factory.fast_rcnn_head_generator(params) if self._include_mask: self._mrcnn_head_fn = factory.mask_rcnn_head_generator(params) # Loss function. self._rpn_score_loss_fn = losses.RpnScoreLoss(params.rpn_score_loss) self._rpn_box_loss_fn = losses.RpnBoxLoss(params.rpn_box_loss) self._frcnn_class_loss_fn = losses.FastrcnnClassLoss() self._frcnn_box_loss_fn = losses.FastrcnnBoxLoss(params.frcnn_box_loss) if self._include_mask: self._mask_loss_fn = losses.MaskrcnnLoss() self._generate_detections_fn = postprocess_ops.GenericDetectionGenerator( params.postprocess) self._transpose_input = params.train.transpose_input assert not self._transpose_input, 'Transpose input is not supportted.' def build_outputs(self, inputs, mode): is_training = mode == mode_keys.TRAIN model_outputs = {} image = inputs['image'] _, image_height, image_width, _ = image.get_shape().as_list() backbone_features = self._backbone_fn(image, is_training) fpn_features = self._fpn_fn(backbone_features, is_training) rpn_score_outputs, rpn_box_outputs = self._rpn_head_fn( fpn_features, is_training) model_outputs.update({ 'rpn_score_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), rpn_score_outputs), 'rpn_box_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), rpn_box_outputs), }) input_anchor = anchor.Anchor(self._params.architecture.min_level, self._params.architecture.max_level, self._params.anchor.num_scales, self._params.anchor.aspect_ratios, self._params.anchor.anchor_size, (image_height, image_width)) rpn_rois, _ = self._generate_rois_fn(rpn_box_outputs, rpn_score_outputs, input_anchor.multilevel_boxes, inputs['image_info'][:, 1, :], is_training) if is_training: rpn_rois = tf.stop_gradient(rpn_rois) # Sample proposals. rpn_rois, matched_gt_boxes, matched_gt_classes, matched_gt_indices = ( self._sample_rois_fn(rpn_rois, inputs['gt_boxes'], inputs['gt_classes'])) # Create bounding box training targets. box_targets = box_utils.encode_boxes( matched_gt_boxes, rpn_rois, weights=[10.0, 10.0, 5.0, 5.0]) # If the target is background, the box target is set to all 0s. box_targets = tf.where( tf.tile( tf.expand_dims(tf.equal(matched_gt_classes, 0), axis=-1), [1, 1, 4]), tf.zeros_like(box_targets), box_targets) model_outputs.update({ 'class_targets': matched_gt_classes, 'box_targets': box_targets, }) roi_features = spatial_transform_ops.multilevel_crop_and_resize( fpn_features, rpn_rois, output_size=7) class_outputs, box_outputs = self._frcnn_head_fn(roi_features, is_training) model_outputs.update({ 'class_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), class_outputs), 'box_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), box_outputs), }) # Add this output to train to make the checkpoint loadable in predict mode. # If we skip it in train mode, the heads will be out-of-order and checkpoint # loading will fail. boxes, scores, classes, valid_detections = self._generate_detections_fn( box_outputs, class_outputs, rpn_rois, inputs['image_info'][:, 1:2, :]) model_outputs.update({ 'num_detections': valid_detections, 'detection_boxes': boxes, 'detection_classes': classes, 'detection_scores': scores, }) if not self._include_mask: return model_outputs if is_training: rpn_rois, classes, mask_targets = self._sample_masks_fn( rpn_rois, matched_gt_boxes, matched_gt_classes, matched_gt_indices, inputs['gt_masks']) mask_targets = tf.stop_gradient(mask_targets) classes = tf.cast(classes, dtype=tf.int32) model_outputs.update({ 'mask_targets': mask_targets, 'sampled_class_targets': classes, }) else: rpn_rois = boxes classes = tf.cast(classes, dtype=tf.int32) mask_roi_features = spatial_transform_ops.multilevel_crop_and_resize( fpn_features, rpn_rois, output_size=14) mask_outputs = self._mrcnn_head_fn(mask_roi_features, classes, is_training) if is_training: model_outputs.update({ 'mask_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), mask_outputs), }) else: model_outputs.update({'detection_masks': tf.nn.sigmoid(mask_outputs)}) return model_outputs def build_loss_fn(self): if self._keras_model is None: raise ValueError('build_loss_fn() must be called after build_model().') filter_fn = self.make_filter_trainable_variables_fn() trainable_variables = filter_fn(self._keras_model.trainable_variables) def _total_loss_fn(labels, outputs): rpn_score_loss = self._rpn_score_loss_fn(outputs['rpn_score_outputs'], labels['rpn_score_targets']) rpn_box_loss = self._rpn_box_loss_fn(outputs['rpn_box_outputs'], labels['rpn_box_targets']) frcnn_class_loss = self._frcnn_class_loss_fn(outputs['class_outputs'], outputs['class_targets']) frcnn_box_loss = self._frcnn_box_loss_fn(outputs['box_outputs'], outputs['class_targets'], outputs['box_targets']) if self._include_mask: mask_loss = self._mask_loss_fn(outputs['mask_outputs'], outputs['mask_targets'], outputs['sampled_class_targets']) else: mask_loss = 0.0 model_loss = ( rpn_score_loss + rpn_box_loss + frcnn_class_loss + frcnn_box_loss + mask_loss) l2_regularization_loss = self.weight_decay_loss(trainable_variables) total_loss = model_loss + l2_regularization_loss return { 'total_loss': total_loss, 'loss': total_loss, 'fast_rcnn_class_loss': frcnn_class_loss, 'fast_rcnn_box_loss': frcnn_box_loss, 'mask_loss': mask_loss, 'model_loss': model_loss, 'l2_regularization_loss': l2_regularization_loss, 'rpn_score_loss': rpn_score_loss, 'rpn_box_loss': rpn_box_loss, } return _total_loss_fn def build_input_layers(self, params, mode): is_training = mode == mode_keys.TRAIN input_shape = ( params.maskrcnn_parser.output_size + [params.maskrcnn_parser.num_channels]) if is_training: batch_size = params.train.batch_size input_layer = { 'image': tf.keras.layers.Input( shape=input_shape, batch_size=batch_size, name='image', dtype=tf.bfloat16 if self._use_bfloat16 else tf.float32), 'image_info': tf.keras.layers.Input( shape=[4, 2], batch_size=batch_size, name='image_info', ), 'gt_boxes': tf.keras.layers.Input( shape=[params.maskrcnn_parser.max_num_instances, 4], batch_size=batch_size, name='gt_boxes'), 'gt_classes': tf.keras.layers.Input( shape=[params.maskrcnn_parser.max_num_instances], batch_size=batch_size, name='gt_classes', dtype=tf.int64), } if self._include_mask: input_layer['gt_masks'] = tf.keras.layers.Input( shape=[ params.maskrcnn_parser.max_num_instances, params.maskrcnn_parser.mask_crop_size, params.maskrcnn_parser.mask_crop_size ], batch_size=batch_size, name='gt_masks') else: batch_size = params.eval.batch_size input_layer = { 'image': tf.keras.layers.Input( shape=input_shape, batch_size=batch_size, name='image', dtype=tf.bfloat16 if self._use_bfloat16 else tf.float32), 'image_info': tf.keras.layers.Input( shape=[4, 2], batch_size=batch_size, name='image_info', ), } return input_layer def build_model(self, params, mode): if self._keras_model is None: input_layers = self.build_input_layers(self._params, mode) outputs = self.model_outputs(input_layers, mode) model = tf.keras.models.Model( inputs=input_layers, outputs=outputs, name='maskrcnn') assert model is not None, 'Fail to build tf.keras.Model.' model.optimizer = self.build_optimizer() self._keras_model = model return self._keras_model def post_processing(self, labels, outputs): required_output_fields = ['class_outputs', 'box_outputs'] for field in required_output_fields: if field not in outputs: raise ValueError('"%s" is missing in outputs, requried %s found %s' % (field, required_output_fields, outputs.keys())) predictions = { 'image_info': labels['image_info'], 'num_detections': outputs['num_detections'], 'detection_boxes': outputs['detection_boxes'], 'detection_classes': outputs['detection_classes'], 'detection_scores': outputs['detection_scores'], } if self._include_mask: predictions.update({ 'detection_masks': outputs['detection_masks'], }) if 'groundtruths' in labels: predictions['source_id'] = labels['groundtruths']['source_id'] predictions['gt_source_id'] = labels['groundtruths']['source_id'] predictions['gt_height'] = labels['groundtruths']['height'] predictions['gt_width'] = labels['groundtruths']['width'] predictions['gt_image_info'] = labels['image_info'] predictions['gt_num_detections'] = ( labels['groundtruths']['num_detections']) predictions['gt_boxes'] = labels['groundtruths']['boxes'] predictions['gt_classes'] = labels['groundtruths']['classes'] predictions['gt_areas'] = labels['groundtruths']['areas'] predictions['gt_is_crowds'] = labels['groundtruths']['is_crowds'] return labels, predictions def eval_metrics(self): return eval_factory.evaluator_generator(self._params.eval)
13,458
38.702065
80
py
models
models-master/official/legacy/detection/modeling/olnmask_model.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Model defination for the Object Localization Network (OLN) Model.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from official.legacy.detection.dataloader import anchor from official.legacy.detection.dataloader import mode_keys from official.legacy.detection.modeling import losses from official.legacy.detection.modeling.architecture import factory from official.legacy.detection.modeling.maskrcnn_model import MaskrcnnModel from official.legacy.detection.ops import postprocess_ops from official.legacy.detection.ops import roi_ops from official.legacy.detection.ops import spatial_transform_ops from official.legacy.detection.ops import target_ops from official.legacy.detection.utils import box_utils class OlnMaskModel(MaskrcnnModel): """OLN-Mask model function.""" def __init__(self, params): super(OlnMaskModel, self).__init__(params) self._params = params # Different heads and layers. self._include_rpn_class = params.architecture.include_rpn_class self._include_mask = params.architecture.include_mask self._include_frcnn_class = params.architecture.include_frcnn_class self._include_frcnn_box = params.architecture.include_frcnn_box self._include_centerness = params.rpn_head.has_centerness self._include_box_score = (params.frcnn_head.has_scoring and params.architecture.include_frcnn_box) self._include_mask_score = (params.mrcnn_head.has_scoring and params.architecture.include_mask) # Architecture generators. self._backbone_fn = factory.backbone_generator(params) self._fpn_fn = factory.multilevel_features_generator(params) self._rpn_head_fn = factory.rpn_head_generator(params) if self._include_centerness: self._rpn_head_fn = factory.oln_rpn_head_generator(params) else: self._rpn_head_fn = factory.rpn_head_generator(params) self._generate_rois_fn = roi_ops.OlnROIGenerator(params.roi_proposal) self._sample_rois_fn = target_ops.ROIScoreSampler(params.roi_sampling) self._sample_masks_fn = target_ops.MaskSampler( params.architecture.mask_target_size, params.mask_sampling.num_mask_samples_per_image) if self._include_box_score: self._frcnn_head_fn = factory.oln_box_score_head_generator(params) else: self._frcnn_head_fn = factory.fast_rcnn_head_generator(params) if self._include_mask: if self._include_mask_score: self._mrcnn_head_fn = factory.oln_mask_score_head_generator(params) else: self._mrcnn_head_fn = factory.mask_rcnn_head_generator(params) # Loss function. self._rpn_score_loss_fn = losses.RpnScoreLoss(params.rpn_score_loss) self._rpn_box_loss_fn = losses.RpnBoxLoss(params.rpn_box_loss) if self._include_centerness: self._rpn_iou_loss_fn = losses.OlnRpnIoULoss() self._rpn_center_loss_fn = losses.OlnRpnCenterLoss() self._frcnn_class_loss_fn = losses.FastrcnnClassLoss() self._frcnn_box_loss_fn = losses.FastrcnnBoxLoss(params.frcnn_box_loss) if self._include_box_score: self._frcnn_box_score_loss_fn = losses.OlnBoxScoreLoss( params.frcnn_box_score_loss) if self._include_mask: self._mask_loss_fn = losses.MaskrcnnLoss() self._generate_detections_fn = postprocess_ops.OlnDetectionGenerator( params.postprocess) self._transpose_input = params.train.transpose_input assert not self._transpose_input, 'Transpose input is not supportted.' def build_outputs(self, inputs, mode): is_training = mode == mode_keys.TRAIN model_outputs = {} image = inputs['image'] _, image_height, image_width, _ = image.get_shape().as_list() backbone_features = self._backbone_fn(image, is_training) fpn_features = self._fpn_fn(backbone_features, is_training) # rpn_centerness. if self._include_centerness: rpn_score_outputs, rpn_box_outputs, rpn_center_outputs = ( self._rpn_head_fn(fpn_features, is_training)) model_outputs.update({ 'rpn_center_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), rpn_center_outputs), }) object_scores = rpn_center_outputs else: rpn_score_outputs, rpn_box_outputs = self._rpn_head_fn( fpn_features, is_training) object_scores = None model_outputs.update({ 'rpn_score_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), rpn_score_outputs), 'rpn_box_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), rpn_box_outputs), }) input_anchor = anchor.Anchor(self._params.architecture.min_level, self._params.architecture.max_level, self._params.anchor.num_scales, self._params.anchor.aspect_ratios, self._params.anchor.anchor_size, (image_height, image_width)) rpn_rois, rpn_roi_scores = self._generate_rois_fn( rpn_box_outputs, rpn_score_outputs, input_anchor.multilevel_boxes, inputs['image_info'][:, 1, :], is_training, is_box_lrtb=self._include_centerness, object_scores=object_scores, ) if (not self._include_frcnn_class and not self._include_frcnn_box and not self._include_mask): # if not is_training: # For direct RPN detection, # use dummy box_outputs = (dy,dx,dh,dw = 0,0,0,0) box_outputs = tf.zeros_like(rpn_rois) box_outputs = tf.concat([box_outputs, box_outputs], -1) boxes, scores, classes, valid_detections = self._generate_detections_fn( box_outputs, rpn_roi_scores, rpn_rois, inputs['image_info'][:, 1:2, :], is_single_fg_score=True, # if no_background, no softmax is applied. keep_nms=True) model_outputs.update({ 'num_detections': valid_detections, 'detection_boxes': boxes, 'detection_classes': classes, 'detection_scores': scores, }) return model_outputs # ---- OLN-Proposal finishes here. ---- if is_training: rpn_rois = tf.stop_gradient(rpn_rois) rpn_roi_scores = tf.stop_gradient(rpn_roi_scores) # Sample proposals. (rpn_rois, rpn_roi_scores, matched_gt_boxes, matched_gt_classes, matched_gt_indices) = ( self._sample_rois_fn(rpn_rois, rpn_roi_scores, inputs['gt_boxes'], inputs['gt_classes'])) # Create bounding box training targets. box_targets = box_utils.encode_boxes( matched_gt_boxes, rpn_rois, weights=[10.0, 10.0, 5.0, 5.0]) # If the target is background, the box target is set to all 0s. box_targets = tf.where( tf.tile( tf.expand_dims(tf.equal(matched_gt_classes, 0), axis=-1), [1, 1, 4]), tf.zeros_like(box_targets), box_targets) model_outputs.update({ 'class_targets': matched_gt_classes, 'box_targets': box_targets, }) # Create Box-IoU targets. { box_ious = box_utils.bbox_overlap( rpn_rois, inputs['gt_boxes']) matched_box_ious = tf.reduce_max(box_ious, 2) model_outputs.update({ 'box_iou_targets': matched_box_ious,}) # } roi_features = spatial_transform_ops.multilevel_crop_and_resize( fpn_features, rpn_rois, output_size=7) if not self._include_box_score: class_outputs, box_outputs = self._frcnn_head_fn( roi_features, is_training) else: class_outputs, box_outputs, score_outputs = self._frcnn_head_fn( roi_features, is_training) model_outputs.update({ 'box_score_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), score_outputs),}) model_outputs.update({ 'class_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), class_outputs), 'box_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), box_outputs), }) # Add this output to train to make the checkpoint loadable in predict mode. # If we skip it in train mode, the heads will be out-of-order and checkpoint # loading will fail. if not self._include_frcnn_box: box_outputs = tf.zeros_like(box_outputs) # dummy zeros. if self._include_box_score: score_outputs = tf.cast(tf.squeeze(score_outputs, -1), rpn_roi_scores.dtype) # box-score = (rpn-centerness * box-iou)^(1/2) # TR: rpn_roi_scores: b,1000, score_outputs: b,512 # TS: rpn_roi_scores: b,1000, score_outputs: b,1000 box_scores = tf.pow( rpn_roi_scores * tf.sigmoid(score_outputs), 1/2.) if not self._include_frcnn_class: boxes, scores, classes, valid_detections = self._generate_detections_fn( box_outputs, box_scores, rpn_rois, inputs['image_info'][:, 1:2, :], is_single_fg_score=True, keep_nms=True,) else: boxes, scores, classes, valid_detections = self._generate_detections_fn( box_outputs, class_outputs, rpn_rois, inputs['image_info'][:, 1:2, :], keep_nms=True,) model_outputs.update({ 'num_detections': valid_detections, 'detection_boxes': boxes, 'detection_classes': classes, 'detection_scores': scores, }) # ---- OLN-Box finishes here. ---- if not self._include_mask: return model_outputs if is_training: rpn_rois, classes, mask_targets = self._sample_masks_fn( rpn_rois, matched_gt_boxes, matched_gt_classes, matched_gt_indices, inputs['gt_masks']) mask_targets = tf.stop_gradient(mask_targets) classes = tf.cast(classes, dtype=tf.int32) model_outputs.update({ 'mask_targets': mask_targets, 'sampled_class_targets': classes, }) else: rpn_rois = boxes classes = tf.cast(classes, dtype=tf.int32) mask_roi_features = spatial_transform_ops.multilevel_crop_and_resize( fpn_features, rpn_rois, output_size=14) mask_outputs = self._mrcnn_head_fn(mask_roi_features, classes, is_training) if is_training: model_outputs.update({ 'mask_outputs': tf.nest.map_structure(lambda x: tf.cast(x, tf.float32), mask_outputs), }) else: model_outputs.update({'detection_masks': tf.nn.sigmoid(mask_outputs)}) return model_outputs def build_loss_fn(self): if self._keras_model is None: raise ValueError('build_loss_fn() must be called after build_model().') filter_fn = self.make_filter_trainable_variables_fn() trainable_variables = filter_fn(self._keras_model.trainable_variables) def _total_loss_fn(labels, outputs): if self._include_rpn_class: rpn_score_loss = self._rpn_score_loss_fn(outputs['rpn_score_outputs'], labels['rpn_score_targets']) else: rpn_score_loss = 0.0 if self._include_centerness: rpn_center_loss = self._rpn_center_loss_fn( outputs['rpn_center_outputs'], labels['rpn_center_targets']) rpn_box_loss = self._rpn_iou_loss_fn( outputs['rpn_box_outputs'], labels['rpn_box_targets'], labels['rpn_center_targets']) else: rpn_center_loss = 0.0 rpn_box_loss = self._rpn_box_loss_fn( outputs['rpn_box_outputs'], labels['rpn_box_targets']) if self._include_frcnn_class: frcnn_class_loss = self._frcnn_class_loss_fn( outputs['class_outputs'], outputs['class_targets']) else: frcnn_class_loss = 0.0 if self._include_frcnn_box: frcnn_box_loss = self._frcnn_box_loss_fn( outputs['box_outputs'], outputs['class_targets'], outputs['box_targets']) else: frcnn_box_loss = 0.0 if self._include_box_score: box_score_loss = self._frcnn_box_score_loss_fn( outputs['box_score_outputs'], outputs['box_iou_targets']) else: box_score_loss = 0.0 if self._include_mask: mask_loss = self._mask_loss_fn(outputs['mask_outputs'], outputs['mask_targets'], outputs['sampled_class_targets']) else: mask_loss = 0.0 model_loss = ( rpn_score_loss + rpn_box_loss + rpn_center_loss + frcnn_class_loss + frcnn_box_loss + box_score_loss + mask_loss) l2_regularization_loss = self.weight_decay_loss(trainable_variables) total_loss = model_loss + l2_regularization_loss return { 'total_loss': total_loss, 'loss': total_loss, 'fast_rcnn_class_loss': frcnn_class_loss, 'fast_rcnn_box_loss': frcnn_box_loss, 'fast_rcnn_box_score_loss': box_score_loss, 'mask_loss': mask_loss, 'model_loss': model_loss, 'l2_regularization_loss': l2_regularization_loss, 'rpn_score_loss': rpn_score_loss, 'rpn_box_loss': rpn_box_loss, 'rpn_center_loss': rpn_center_loss, } return _total_loss_fn def build_input_layers(self, params, mode): is_training = mode == mode_keys.TRAIN input_shape = ( params.olnmask_parser.output_size + [params.olnmask_parser.num_channels]) if is_training: batch_size = params.train.batch_size input_layer = { 'image': tf.keras.layers.Input( shape=input_shape, batch_size=batch_size, name='image', dtype=tf.bfloat16 if self._use_bfloat16 else tf.float32), 'image_info': tf.keras.layers.Input( shape=[4, 2], batch_size=batch_size, name='image_info', ), 'gt_boxes': tf.keras.layers.Input( shape=[params.olnmask_parser.max_num_instances, 4], batch_size=batch_size, name='gt_boxes'), 'gt_classes': tf.keras.layers.Input( shape=[params.olnmask_parser.max_num_instances], batch_size=batch_size, name='gt_classes', dtype=tf.int64), } if self._include_mask: input_layer['gt_masks'] = tf.keras.layers.Input( shape=[ params.olnmask_parser.max_num_instances, params.olnmask_parser.mask_crop_size, params.olnmask_parser.mask_crop_size ], batch_size=batch_size, name='gt_masks') else: batch_size = params.eval.batch_size input_layer = { 'image': tf.keras.layers.Input( shape=input_shape, batch_size=batch_size, name='image', dtype=tf.bfloat16 if self._use_bfloat16 else tf.float32), 'image_info': tf.keras.layers.Input( shape=[4, 2], batch_size=batch_size, name='image_info', ), } return input_layer def build_model(self, params, mode): if self._keras_model is None: input_layers = self.build_input_layers(self._params, mode) outputs = self.model_outputs(input_layers, mode) model = tf.keras.models.Model( inputs=input_layers, outputs=outputs, name='olnmask') assert model is not None, 'Fail to build tf.keras.Model.' model.optimizer = self.build_optimizer() self._keras_model = model return self._keras_model
16,774
37.741339
80
py
models
models-master/official/legacy/detection/modeling/retinanet_model.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Model defination for the RetinaNet Model.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from official.legacy.detection.dataloader import mode_keys from official.legacy.detection.evaluation import factory as eval_factory from official.legacy.detection.modeling import base_model from official.legacy.detection.modeling import losses from official.legacy.detection.modeling.architecture import factory from official.legacy.detection.ops import postprocess_ops class RetinanetModel(base_model.Model): """RetinaNet model function.""" def __init__(self, params): super(RetinanetModel, self).__init__(params) # For eval metrics. self._params = params # Architecture generators. self._backbone_fn = factory.backbone_generator(params) self._fpn_fn = factory.multilevel_features_generator(params) self._head_fn = factory.retinanet_head_generator(params) # Loss function. self._cls_loss_fn = losses.RetinanetClassLoss( params.retinanet_loss, params.architecture.num_classes) self._box_loss_fn = losses.RetinanetBoxLoss(params.retinanet_loss) self._box_loss_weight = params.retinanet_loss.box_loss_weight self._keras_model = None # Predict function. self._generate_detections_fn = postprocess_ops.MultilevelDetectionGenerator( params.architecture.min_level, params.architecture.max_level, params.postprocess) self._transpose_input = params.train.transpose_input assert not self._transpose_input, 'Transpose input is not supported.' # Input layer. self._input_layer = tf.keras.layers.Input( shape=(None, None, params.retinanet_parser.num_channels), name='', dtype=tf.bfloat16 if self._use_bfloat16 else tf.float32) def build_outputs(self, inputs, mode): # If the input image is transposed (from NHWC to HWCN), we need to revert it # back to the original shape before it's used in the computation. if self._transpose_input: inputs = tf.transpose(inputs, [3, 0, 1, 2]) backbone_features = self._backbone_fn( inputs, is_training=(mode == mode_keys.TRAIN)) fpn_features = self._fpn_fn( backbone_features, is_training=(mode == mode_keys.TRAIN)) cls_outputs, box_outputs = self._head_fn( fpn_features, is_training=(mode == mode_keys.TRAIN)) if self._use_bfloat16: levels = cls_outputs.keys() for level in levels: cls_outputs[level] = tf.cast(cls_outputs[level], tf.float32) box_outputs[level] = tf.cast(box_outputs[level], tf.float32) model_outputs = { 'cls_outputs': cls_outputs, 'box_outputs': box_outputs, } return model_outputs def build_loss_fn(self): if self._keras_model is None: raise ValueError('build_loss_fn() must be called after build_model().') filter_fn = self.make_filter_trainable_variables_fn() trainable_variables = filter_fn(self._keras_model.trainable_variables) def _total_loss_fn(labels, outputs): cls_loss = self._cls_loss_fn(outputs['cls_outputs'], labels['cls_targets'], labels['num_positives']) box_loss = self._box_loss_fn(outputs['box_outputs'], labels['box_targets'], labels['num_positives']) model_loss = cls_loss + self._box_loss_weight * box_loss l2_regularization_loss = self.weight_decay_loss(trainable_variables) total_loss = model_loss + l2_regularization_loss return { 'total_loss': total_loss, 'cls_loss': cls_loss, 'box_loss': box_loss, 'model_loss': model_loss, 'l2_regularization_loss': l2_regularization_loss, } return _total_loss_fn def build_model(self, params, mode=None): if self._keras_model is None: outputs = self.model_outputs(self._input_layer, mode) model = tf.keras.models.Model( inputs=self._input_layer, outputs=outputs, name='retinanet') assert model is not None, 'Fail to build tf.keras.Model.' model.optimizer = self.build_optimizer() self._keras_model = model return self._keras_model def post_processing(self, labels, outputs): # TODO(yeqing): Moves the output related part into build_outputs. required_output_fields = ['cls_outputs', 'box_outputs'] for field in required_output_fields: if field not in outputs: raise ValueError('"%s" is missing in outputs, requried %s found %s' % (field, required_output_fields, outputs.keys())) required_label_fields = ['image_info', 'groundtruths'] for field in required_label_fields: if field not in labels: raise ValueError('"%s" is missing in outputs, requried %s found %s' % (field, required_label_fields, labels.keys())) boxes, scores, classes, valid_detections = self._generate_detections_fn( outputs['box_outputs'], outputs['cls_outputs'], labels['anchor_boxes'], labels['image_info'][:, 1:2, :]) # Discards the old output tensors to save memory. The `cls_outputs` and # `box_outputs` are pretty big and could potentiall lead to memory issue. outputs = { 'source_id': labels['groundtruths']['source_id'], 'image_info': labels['image_info'], 'num_detections': valid_detections, 'detection_boxes': boxes, 'detection_classes': classes, 'detection_scores': scores, } if 'groundtruths' in labels: labels['source_id'] = labels['groundtruths']['source_id'] labels['boxes'] = labels['groundtruths']['boxes'] labels['classes'] = labels['groundtruths']['classes'] labels['areas'] = labels['groundtruths']['areas'] labels['is_crowds'] = labels['groundtruths']['is_crowds'] return labels, outputs def eval_metrics(self): return eval_factory.evaluator_generator(self._params.eval)
6,661
39.13253
80
py
models
models-master/official/legacy/detection/modeling/factory.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Factory to build detection model.""" from official.legacy.detection.modeling import maskrcnn_model from official.legacy.detection.modeling import olnmask_model from official.legacy.detection.modeling import retinanet_model from official.legacy.detection.modeling import shapemask_model def model_generator(params): """Model function generator.""" if params.type == 'retinanet': model_fn = retinanet_model.RetinanetModel(params) elif params.type == 'mask_rcnn': model_fn = maskrcnn_model.MaskrcnnModel(params) elif params.type == 'olnmask': model_fn = olnmask_model.OlnMaskModel(params) elif params.type == 'shapemask': model_fn = shapemask_model.ShapeMaskModel(params) else: raise ValueError('Model %s is not supported.'% params.type) return model_fn
1,402
35.921053
74
py
models
models-master/official/legacy/detection/modeling/shapemask_model.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Model definition for the ShapeMask Model.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from official.legacy.detection.dataloader import anchor from official.legacy.detection.dataloader import mode_keys from official.legacy.detection.evaluation import factory as eval_factory from official.legacy.detection.modeling import base_model from official.legacy.detection.modeling import losses from official.legacy.detection.modeling.architecture import factory from official.legacy.detection.ops import postprocess_ops from official.legacy.detection.utils import box_utils class ShapeMaskModel(base_model.Model): """ShapeMask model function.""" def __init__(self, params): super(ShapeMaskModel, self).__init__(params) self._params = params self._keras_model = None # Architecture generators. self._backbone_fn = factory.backbone_generator(params) self._fpn_fn = factory.multilevel_features_generator(params) self._retinanet_head_fn = factory.retinanet_head_generator(params) self._shape_prior_head_fn = factory.shapeprior_head_generator(params) self._coarse_mask_fn = factory.coarsemask_head_generator(params) self._fine_mask_fn = factory.finemask_head_generator(params) # Loss functions. self._cls_loss_fn = losses.RetinanetClassLoss( params.retinanet_loss, params.architecture.num_classes) self._box_loss_fn = losses.RetinanetBoxLoss(params.retinanet_loss) self._box_loss_weight = params.retinanet_loss.box_loss_weight # Mask loss function. self._shapemask_prior_loss_fn = losses.ShapemaskMseLoss() self._shapemask_loss_fn = losses.ShapemaskLoss() self._shape_prior_loss_weight = ( params.shapemask_loss.shape_prior_loss_weight) self._coarse_mask_loss_weight = ( params.shapemask_loss.coarse_mask_loss_weight) self._fine_mask_loss_weight = (params.shapemask_loss.fine_mask_loss_weight) # Predict function. self._generate_detections_fn = postprocess_ops.MultilevelDetectionGenerator( params.architecture.min_level, params.architecture.max_level, params.postprocess) def build_outputs(self, inputs, mode): is_training = mode == mode_keys.TRAIN images = inputs['image'] if 'anchor_boxes' in inputs: anchor_boxes = inputs['anchor_boxes'] else: anchor_boxes = anchor.Anchor( self._params.architecture.min_level, self._params.architecture.max_level, self._params.anchor.num_scales, self._params.anchor.aspect_ratios, self._params.anchor.anchor_size, images.get_shape().as_list()[1:3]).multilevel_boxes batch_size = tf.shape(images)[0] for level in anchor_boxes: anchor_boxes[level] = tf.tile( tf.expand_dims(anchor_boxes[level], 0), [batch_size, 1, 1, 1]) backbone_features = self._backbone_fn(images, is_training=is_training) fpn_features = self._fpn_fn(backbone_features, is_training=is_training) cls_outputs, box_outputs = self._retinanet_head_fn( fpn_features, is_training=is_training) valid_boxes, valid_scores, valid_classes, valid_detections = ( self._generate_detections_fn(box_outputs, cls_outputs, anchor_boxes, inputs['image_info'][:, 1:2, :])) image_size = images.get_shape().as_list()[1:3] valid_outer_boxes = box_utils.compute_outer_boxes( tf.reshape(valid_boxes, [-1, 4]), image_size, scale=self._params.shapemask_parser.outer_box_scale) valid_outer_boxes = tf.reshape(valid_outer_boxes, tf.shape(valid_boxes)) # Wrapping if else code paths into a layer to make the checkpoint loadable # in prediction mode. class SampledBoxesLayer(tf.keras.layers.Layer): """ShapeMask model function.""" def call(self, inputs, val_boxes, val_classes, val_outer_boxes, training): if training: boxes = inputs['mask_boxes'] outer_boxes = inputs['mask_outer_boxes'] classes = inputs['mask_classes'] else: boxes = val_boxes classes = val_classes outer_boxes = val_outer_boxes return boxes, classes, outer_boxes boxes, classes, outer_boxes = SampledBoxesLayer()( inputs, valid_boxes, valid_classes, valid_outer_boxes, training=is_training) instance_features, prior_masks = self._shape_prior_head_fn( fpn_features, boxes, outer_boxes, classes, is_training) coarse_mask_logits = self._coarse_mask_fn(instance_features, prior_masks, classes, is_training) fine_mask_logits = self._fine_mask_fn(instance_features, coarse_mask_logits, classes, is_training) model_outputs = { 'cls_outputs': cls_outputs, 'box_outputs': box_outputs, 'fine_mask_logits': fine_mask_logits, 'coarse_mask_logits': coarse_mask_logits, 'prior_masks': prior_masks, } if not is_training: model_outputs.update({ 'num_detections': valid_detections, 'detection_boxes': valid_boxes, 'detection_outer_boxes': valid_outer_boxes, 'detection_masks': fine_mask_logits, 'detection_classes': valid_classes, 'detection_scores': valid_scores, }) return model_outputs def build_loss_fn(self): if self._keras_model is None: raise ValueError('build_loss_fn() must be called after build_model().') filter_fn = self.make_filter_trainable_variables_fn() trainable_variables = filter_fn(self._keras_model.trainable_variables) def _total_loss_fn(labels, outputs): cls_loss = self._cls_loss_fn(outputs['cls_outputs'], labels['cls_targets'], labels['num_positives']) box_loss = self._box_loss_fn(outputs['box_outputs'], labels['box_targets'], labels['num_positives']) # Adds Shapemask model losses. shape_prior_loss = self._shapemask_prior_loss_fn(outputs['prior_masks'], labels['mask_targets'], labels['mask_is_valid']) coarse_mask_loss = self._shapemask_loss_fn(outputs['coarse_mask_logits'], labels['mask_targets'], labels['mask_is_valid']) fine_mask_loss = self._shapemask_loss_fn(outputs['fine_mask_logits'], labels['fine_mask_targets'], labels['mask_is_valid']) model_loss = ( cls_loss + self._box_loss_weight * box_loss + shape_prior_loss * self._shape_prior_loss_weight + coarse_mask_loss * self._coarse_mask_loss_weight + fine_mask_loss * self._fine_mask_loss_weight) l2_regularization_loss = self.weight_decay_loss(trainable_variables) total_loss = model_loss + l2_regularization_loss shapemask_losses = { 'total_loss': total_loss, 'loss': total_loss, 'retinanet_cls_loss': cls_loss, 'l2_regularization_loss': l2_regularization_loss, 'retinanet_box_loss': box_loss, 'shapemask_prior_loss': shape_prior_loss, 'shapemask_coarse_mask_loss': coarse_mask_loss, 'shapemask_fine_mask_loss': fine_mask_loss, 'model_loss': model_loss, } return shapemask_losses return _total_loss_fn def build_input_layers(self, params, mode): is_training = mode == mode_keys.TRAIN input_shape = ( params.shapemask_parser.output_size + [params.shapemask_parser.num_channels]) if is_training: batch_size = params.train.batch_size input_layer = { 'image': tf.keras.layers.Input( shape=input_shape, batch_size=batch_size, name='image', dtype=tf.bfloat16 if self._use_bfloat16 else tf.float32), 'image_info': tf.keras.layers.Input( shape=[4, 2], batch_size=batch_size, name='image_info'), 'mask_classes': tf.keras.layers.Input( shape=[params.shapemask_parser.num_sampled_masks], batch_size=batch_size, name='mask_classes', dtype=tf.int64), 'mask_outer_boxes': tf.keras.layers.Input( shape=[params.shapemask_parser.num_sampled_masks, 4], batch_size=batch_size, name='mask_outer_boxes', dtype=tf.float32), 'mask_boxes': tf.keras.layers.Input( shape=[params.shapemask_parser.num_sampled_masks, 4], batch_size=batch_size, name='mask_boxes', dtype=tf.float32), } else: batch_size = params.eval.batch_size input_layer = { 'image': tf.keras.layers.Input( shape=input_shape, batch_size=batch_size, name='image', dtype=tf.bfloat16 if self._use_bfloat16 else tf.float32), 'image_info': tf.keras.layers.Input( shape=[4, 2], batch_size=batch_size, name='image_info'), } return input_layer def build_model(self, params, mode): if self._keras_model is None: input_layers = self.build_input_layers(self._params, mode) outputs = self.model_outputs(input_layers, mode) model = tf.keras.models.Model( inputs=input_layers, outputs=outputs, name='shapemask') assert model is not None, 'Fail to build tf.keras.Model.' model.optimizer = self.build_optimizer() self._keras_model = model return self._keras_model def post_processing(self, labels, outputs): required_output_fields = [ 'num_detections', 'detection_boxes', 'detection_classes', 'detection_masks', 'detection_scores' ] for field in required_output_fields: if field not in outputs: raise ValueError( '"{}" is missing in outputs, requried {} found {}'.format( field, required_output_fields, outputs.keys())) required_label_fields = ['image_info'] for field in required_label_fields: if field not in labels: raise ValueError( '"{}" is missing in labels, requried {} found {}'.format( field, required_label_fields, labels.keys())) predictions = { 'image_info': labels['image_info'], 'num_detections': outputs['num_detections'], 'detection_boxes': outputs['detection_boxes'], 'detection_outer_boxes': outputs['detection_outer_boxes'], 'detection_classes': outputs['detection_classes'], 'detection_scores': outputs['detection_scores'], 'detection_masks': outputs['detection_masks'], } if 'groundtruths' in labels: predictions['source_id'] = labels['groundtruths']['source_id'] labels = labels['groundtruths'] return labels, predictions def eval_metrics(self): return eval_factory.evaluator_generator(self._params.eval)
12,110
38.708197
80
py
models
models-master/official/legacy/detection/modeling/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/detection/modeling/architecture/identity.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Identity Fn that forwards the input features.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function class Identity(object): """Identity function that forwards the input features.""" def __call__(self, features, is_training=False): """Only forwards the input features.""" return features
974
32.62069
74
py
models
models-master/official/legacy/detection/modeling/architecture/spinenet.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Implementation of SpineNet model. X. Du, T-Y. Lin, P. Jin, G. Ghiasi, M. Tan, Y. Cui, Q. V. Le, X. Song SpineNet: Learning Scale-Permuted Backbone for Recognition and Localization https://arxiv.org/abs/1912.05027 """ import math from absl import logging import tensorflow as tf from official.legacy.detection.modeling.architecture import nn_blocks from official.modeling import tf_utils layers = tf.keras.layers FILTER_SIZE_MAP = { 1: 32, 2: 64, 3: 128, 4: 256, 5: 256, 6: 256, 7: 256, } # The fixed SpineNet architecture discovered by NAS. # Each element represents a specification of a building block: # (block_level, block_fn, (input_offset0, input_offset1), is_output). SPINENET_BLOCK_SPECS = [ (2, 'bottleneck', (0, 1), False), (4, 'residual', (0, 1), False), (3, 'bottleneck', (2, 3), False), (4, 'bottleneck', (2, 4), False), (6, 'residual', (3, 5), False), (4, 'bottleneck', (3, 5), False), (5, 'residual', (6, 7), False), (7, 'residual', (6, 8), False), (5, 'bottleneck', (8, 9), False), (5, 'bottleneck', (8, 10), False), (4, 'bottleneck', (5, 10), True), (3, 'bottleneck', (4, 10), True), (5, 'bottleneck', (7, 12), True), (7, 'bottleneck', (5, 14), True), (6, 'bottleneck', (12, 14), True), ] SCALING_MAP = { '49S': { 'endpoints_num_filters': 128, 'filter_size_scale': 0.65, 'resample_alpha': 0.5, 'block_repeats': 1, }, '49': { 'endpoints_num_filters': 256, 'filter_size_scale': 1.0, 'resample_alpha': 0.5, 'block_repeats': 1, }, '96': { 'endpoints_num_filters': 256, 'filter_size_scale': 1.0, 'resample_alpha': 0.5, 'block_repeats': 2, }, '143': { 'endpoints_num_filters': 256, 'filter_size_scale': 1.0, 'resample_alpha': 1.0, 'block_repeats': 3, }, '190': { 'endpoints_num_filters': 512, 'filter_size_scale': 1.3, 'resample_alpha': 1.0, 'block_repeats': 4, }, } class BlockSpec(object): """A container class that specifies the block configuration for SpineNet.""" def __init__(self, level, block_fn, input_offsets, is_output): self.level = level self.block_fn = block_fn self.input_offsets = input_offsets self.is_output = is_output def build_block_specs(block_specs=None): """Builds the list of BlockSpec objects for SpineNet.""" if not block_specs: block_specs = SPINENET_BLOCK_SPECS logging.info('Building SpineNet block specs: %s', block_specs) return [BlockSpec(*b) for b in block_specs] class SpineNet(tf.keras.Model): """Class to build SpineNet models.""" def __init__(self, input_specs=tf.keras.layers.InputSpec(shape=[None, 640, 640, 3]), min_level=3, max_level=7, block_specs=None, endpoints_num_filters=256, resample_alpha=0.5, block_repeats=1, filter_size_scale=1.0, kernel_initializer='VarianceScaling', kernel_regularizer=None, bias_regularizer=None, activation='relu', use_sync_bn=False, norm_momentum=0.99, norm_epsilon=0.001, **kwargs): """SpineNet model.""" self._min_level = min_level self._max_level = max_level self._block_specs = ( build_block_specs() if block_specs is None else block_specs ) self._endpoints_num_filters = endpoints_num_filters self._resample_alpha = resample_alpha self._block_repeats = block_repeats self._filter_size_scale = filter_size_scale self._kernel_initializer = kernel_initializer self._kernel_regularizer = kernel_regularizer self._bias_regularizer = bias_regularizer self._use_sync_bn = use_sync_bn self._norm_momentum = norm_momentum self._norm_epsilon = norm_epsilon if activation == 'relu': self._activation = tf.nn.relu elif activation == 'swish': self._activation = tf.nn.swish else: raise ValueError('Activation {} not implemented.'.format(activation)) self._init_block_fn = 'bottleneck' self._num_init_blocks = 2 if use_sync_bn: self._norm = layers.experimental.SyncBatchNormalization else: self._norm = layers.BatchNormalization if tf.keras.backend.image_data_format() == 'channels_last': self._bn_axis = -1 else: self._bn_axis = 1 # Build SpineNet. inputs = tf.keras.Input(shape=input_specs.shape[1:]) net = self._build_stem(inputs=inputs) net = self._build_scale_permuted_network( net=net, input_width=input_specs.shape[1]) net = self._build_endpoints(net=net) super(SpineNet, self).__init__(inputs=inputs, outputs=net) def _block_group(self, inputs, filters, strides, block_fn_cand, block_repeats=1, name='block_group'): """Creates one group of blocks for the SpineNet model.""" block_fn_candidates = { 'bottleneck': nn_blocks.BottleneckBlock, 'residual': nn_blocks.ResidualBlock, } block_fn = block_fn_candidates[block_fn_cand] _, _, _, num_filters = inputs.get_shape().as_list() if block_fn_cand == 'bottleneck': use_projection = not (num_filters == (filters * 4) and strides == 1) else: use_projection = not (num_filters == filters and strides == 1) x = block_fn( filters=filters, strides=strides, use_projection=use_projection, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer, activation=self._activation, use_sync_bn=self._use_sync_bn, norm_momentum=self._norm_momentum, norm_epsilon=self._norm_epsilon)( inputs) for _ in range(1, block_repeats): x = block_fn( filters=filters, strides=1, use_projection=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer, activation=self._activation, use_sync_bn=self._use_sync_bn, norm_momentum=self._norm_momentum, norm_epsilon=self._norm_epsilon)( x) return tf.identity(x, name=name) def _build_stem(self, inputs): """Build SpineNet stem.""" x = layers.Conv2D( filters=64, kernel_size=7, strides=2, use_bias=False, padding='same', kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer)( inputs) x = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon)( x) x = tf_utils.get_activation(self._activation)(x) x = layers.MaxPool2D(pool_size=3, strides=2, padding='same')(x) net = [] # Build the initial level 2 blocks. for i in range(self._num_init_blocks): x = self._block_group( inputs=x, filters=int(FILTER_SIZE_MAP[2] * self._filter_size_scale), strides=1, block_fn_cand=self._init_block_fn, block_repeats=self._block_repeats, name='stem_block_{}'.format(i + 1)) net.append(x) return net def _build_scale_permuted_network(self, net, input_width, weighted_fusion=False): """Build scale-permuted network.""" net_sizes = [int(math.ceil(input_width / 2**2))] * len(net) net_block_fns = [self._init_block_fn] * len(net) num_outgoing_connections = [0] * len(net) endpoints = {} for i, block_spec in enumerate(self._block_specs): # Find out specs for the target block. target_width = int(math.ceil(input_width / 2**block_spec.level)) target_num_filters = int(FILTER_SIZE_MAP[block_spec.level] * self._filter_size_scale) target_block_fn = block_spec.block_fn # Resample then merge input0 and input1. parents = [] input0 = block_spec.input_offsets[0] input1 = block_spec.input_offsets[1] x0 = self._resample_with_alpha( inputs=net[input0], input_width=net_sizes[input0], input_block_fn=net_block_fns[input0], target_width=target_width, target_num_filters=target_num_filters, target_block_fn=target_block_fn, alpha=self._resample_alpha) parents.append(x0) num_outgoing_connections[input0] += 1 x1 = self._resample_with_alpha( inputs=net[input1], input_width=net_sizes[input1], input_block_fn=net_block_fns[input1], target_width=target_width, target_num_filters=target_num_filters, target_block_fn=target_block_fn, alpha=self._resample_alpha) parents.append(x1) num_outgoing_connections[input1] += 1 # Merge 0 outdegree blocks to the output block. if block_spec.is_output: for j, (j_feat, j_connections) in enumerate(zip(net, num_outgoing_connections)): if j_connections == 0 and (j_feat.shape[2] == target_width and j_feat.shape[3] == x0.shape[3]): parents.append(j_feat) num_outgoing_connections[j] += 1 # pylint: disable=g-direct-tensorflow-import if weighted_fusion: dtype = parents[0].dtype parent_weights = [ tf.nn.relu(tf.cast(tf.Variable(1.0, name='block{}_fusion{}'.format( i, j)), dtype=dtype)) for j in range(len(parents))] weights_sum = tf.add_n(parent_weights) parents = [ parents[i] * parent_weights[i] / (weights_sum + 0.0001) for i in range(len(parents)) ] # Fuse all parent nodes then build a new block. x = tf_utils.get_activation(self._activation)(tf.add_n(parents)) x = self._block_group( inputs=x, filters=target_num_filters, strides=1, block_fn_cand=target_block_fn, block_repeats=self._block_repeats, name='scale_permuted_block_{}'.format(i + 1)) net.append(x) net_sizes.append(target_width) net_block_fns.append(target_block_fn) num_outgoing_connections.append(0) # Save output feats. if block_spec.is_output: if block_spec.level in endpoints: raise ValueError('Duplicate feats found for output level {}.'.format( block_spec.level)) if (block_spec.level < self._min_level or block_spec.level > self._max_level): raise ValueError('Output level is out of range [{}, {}]'.format( self._min_level, self._max_level)) endpoints[block_spec.level] = x return endpoints def _build_endpoints(self, net): """Match filter size for endpoints before sharing conv layers.""" endpoints = {} for level in range(self._min_level, self._max_level + 1): x = layers.Conv2D( filters=self._endpoints_num_filters, kernel_size=1, strides=1, use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer)( net[level]) x = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon)( x) x = tf_utils.get_activation(self._activation)(x) endpoints[level] = x return endpoints def _resample_with_alpha(self, inputs, input_width, input_block_fn, target_width, target_num_filters, target_block_fn, alpha=0.5): """Match resolution and feature dimension.""" _, _, _, input_num_filters = inputs.get_shape().as_list() if input_block_fn == 'bottleneck': input_num_filters /= 4 new_num_filters = int(input_num_filters * alpha) x = layers.Conv2D( filters=new_num_filters, kernel_size=1, strides=1, use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer)( inputs) x = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon)( x) x = tf_utils.get_activation(self._activation)(x) # Spatial resampling. if input_width > target_width: x = layers.Conv2D( filters=new_num_filters, kernel_size=3, strides=2, padding='SAME', use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer)( x) x = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon)( x) x = tf_utils.get_activation(self._activation)(x) input_width /= 2 while input_width > target_width: x = layers.MaxPool2D(pool_size=3, strides=2, padding='SAME')(x) input_width /= 2 elif input_width < target_width: scale = target_width // input_width x = layers.UpSampling2D(size=(scale, scale))(x) # Last 1x1 conv to match filter size. if target_block_fn == 'bottleneck': target_num_filters *= 4 x = layers.Conv2D( filters=target_num_filters, kernel_size=1, strides=1, use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer)( x) x = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon)( x) return x class SpineNetBuilder(object): """SpineNet builder.""" def __init__(self, model_id, input_specs=tf.keras.layers.InputSpec(shape=[None, 640, 640, 3]), min_level=3, max_level=7, block_specs=None, kernel_initializer='VarianceScaling', kernel_regularizer=None, bias_regularizer=None, activation='relu', use_sync_bn=False, norm_momentum=0.99, norm_epsilon=0.001): if model_id not in SCALING_MAP: raise ValueError( 'SpineNet {} is not a valid architecture.'.format(model_id)) scaling_params = SCALING_MAP[model_id] self._input_specs = input_specs self._min_level = min_level self._max_level = max_level self._block_specs = block_specs or build_block_specs() self._endpoints_num_filters = scaling_params['endpoints_num_filters'] self._resample_alpha = scaling_params['resample_alpha'] self._block_repeats = scaling_params['block_repeats'] self._filter_size_scale = scaling_params['filter_size_scale'] self._kernel_initializer = kernel_initializer self._kernel_regularizer = kernel_regularizer self._bias_regularizer = bias_regularizer self._activation = activation self._use_sync_bn = use_sync_bn self._norm_momentum = norm_momentum self._norm_epsilon = norm_epsilon def __call__(self, inputs, is_training=None): model = SpineNet( input_specs=self._input_specs, min_level=self._min_level, max_level=self._max_level, block_specs=self._block_specs, endpoints_num_filters=self._endpoints_num_filters, resample_alpha=self._resample_alpha, block_repeats=self._block_repeats, filter_size_scale=self._filter_size_scale, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer, activation=self._activation, use_sync_bn=self._use_sync_bn, norm_momentum=self._norm_momentum, norm_epsilon=self._norm_epsilon) return model(inputs)
17,327
33.312871
80
py
models
models-master/official/legacy/detection/modeling/architecture/resnet.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Contains definitions for the post-activation form of Residual Networks. Residual networks (ResNets) were proposed in: [1] Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun Deep Residual Learning for Image Recognition. arXiv:1512.03385 """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from official.legacy.detection.modeling.architecture import nn_ops # TODO(b/140112644): Refactor the code with Keras style, i.e. build and call. class Resnet(object): """Class to build ResNet family model.""" def __init__( self, resnet_depth, activation='relu', norm_activation=nn_ops.norm_activation_builder(activation='relu'), data_format='channels_last'): """ResNet initialization function. Args: resnet_depth: `int` depth of ResNet backbone model. activation: the activation function. norm_activation: an operation that includes a normalization layer followed by an optional activation layer. data_format: `str` either "channels_first" for `[batch, channels, height, width]` or "channels_last for `[batch, height, width, channels]`. """ self._resnet_depth = resnet_depth if activation == 'relu': self._activation_op = tf.nn.relu elif activation == 'swish': self._activation_op = tf.nn.swish else: raise ValueError('Unsupported activation `{}`.'.format(activation)) self._norm_activation = norm_activation self._data_format = data_format model_params = { 10: { 'block': self.residual_block, 'layers': [1, 1, 1, 1] }, 18: { 'block': self.residual_block, 'layers': [2, 2, 2, 2] }, 34: { 'block': self.residual_block, 'layers': [3, 4, 6, 3] }, 50: { 'block': self.bottleneck_block, 'layers': [3, 4, 6, 3] }, 101: { 'block': self.bottleneck_block, 'layers': [3, 4, 23, 3] }, 152: { 'block': self.bottleneck_block, 'layers': [3, 8, 36, 3] }, 200: { 'block': self.bottleneck_block, 'layers': [3, 24, 36, 3] } } if resnet_depth not in model_params: valid_resnet_depths = ', '.join( [str(depth) for depth in sorted(model_params.keys())]) raise ValueError( 'The resnet_depth should be in [%s]. Not a valid resnet_depth:' % (valid_resnet_depths), self._resnet_depth) params = model_params[resnet_depth] self._resnet_fn = self.resnet_v1_generator(params['block'], params['layers']) def __call__(self, inputs, is_training=None): """Returns the ResNet model for a given size and number of output classes. Args: inputs: a `Tesnor` with shape [batch_size, height, width, 3] representing a batch of images. is_training: `bool` if True, the model is in training mode. Returns: a `dict` containing `int` keys for continuous feature levels [2, 3, 4, 5]. The values are corresponding feature hierarchy in ResNet with shape [batch_size, height_l, width_l, num_filters]. """ with tf.name_scope('resnet%s' % self._resnet_depth): return self._resnet_fn(inputs, is_training) def fixed_padding(self, inputs, kernel_size): """Pads the input along the spatial dimensions independently of input size. Args: inputs: `Tensor` of size `[batch, channels, height, width]` or `[batch, height, width, channels]` depending on `data_format`. kernel_size: `int` kernel size to be used for `conv2d` or max_pool2d` operations. Should be a positive integer. Returns: A padded `Tensor` of the same `data_format` with size either intact (if `kernel_size == 1`) or padded (if `kernel_size > 1`). """ pad_total = kernel_size - 1 pad_beg = pad_total // 2 pad_end = pad_total - pad_beg if self._data_format == 'channels_first': padded_inputs = tf.pad( tensor=inputs, paddings=[[0, 0], [0, 0], [pad_beg, pad_end], [pad_beg, pad_end]]) else: padded_inputs = tf.pad( tensor=inputs, paddings=[[0, 0], [pad_beg, pad_end], [pad_beg, pad_end], [0, 0]]) return padded_inputs def conv2d_fixed_padding(self, inputs, filters, kernel_size, strides): """Strided 2-D convolution with explicit padding. The padding is consistent and is based only on `kernel_size`, not on the dimensions of `inputs` (as opposed to using `tf.layers.conv2d` alone). Args: inputs: `Tensor` of size `[batch, channels, height_in, width_in]`. filters: `int` number of filters in the convolution. kernel_size: `int` size of the kernel to be used in the convolution. strides: `int` strides of the convolution. Returns: A `Tensor` of shape `[batch, filters, height_out, width_out]`. """ if strides > 1: inputs = self.fixed_padding(inputs, kernel_size) return tf.keras.layers.Conv2D( filters=filters, kernel_size=kernel_size, strides=strides, padding=('SAME' if strides == 1 else 'VALID'), use_bias=False, kernel_initializer=tf.initializers.VarianceScaling(), data_format=self._data_format)( inputs=inputs) def residual_block(self, inputs, filters, strides, use_projection=False, is_training=None): """Standard building block for residual networks with BN after convolutions. Args: inputs: `Tensor` of size `[batch, channels, height, width]`. filters: `int` number of filters for the first two convolutions. Note that the third and final convolution will use 4 times as many filters. strides: `int` block stride. If greater than 1, this block will ultimately downsample the input. use_projection: `bool` for whether this block should use a projection shortcut (versus the default identity shortcut). This is usually `True` for the first block of a block group, which may change the number of filters and the resolution. is_training: `bool` if True, the model is in training mode. Returns: The output `Tensor` of the block. """ shortcut = inputs if use_projection: # Projection shortcut in first layer to match filters and strides shortcut = self.conv2d_fixed_padding( inputs=inputs, filters=filters, kernel_size=1, strides=strides) shortcut = self._norm_activation(use_activation=False)( shortcut, is_training=is_training) inputs = self.conv2d_fixed_padding( inputs=inputs, filters=filters, kernel_size=3, strides=strides) inputs = self._norm_activation()(inputs, is_training=is_training) inputs = self.conv2d_fixed_padding( inputs=inputs, filters=filters, kernel_size=3, strides=1) inputs = self._norm_activation( use_activation=False, init_zero=True)( inputs, is_training=is_training) return self._activation_op(inputs + shortcut) def bottleneck_block(self, inputs, filters, strides, use_projection=False, is_training=None): """Bottleneck block variant for residual networks with BN after convolutions. Args: inputs: `Tensor` of size `[batch, channels, height, width]`. filters: `int` number of filters for the first two convolutions. Note that the third and final convolution will use 4 times as many filters. strides: `int` block stride. If greater than 1, this block will ultimately downsample the input. use_projection: `bool` for whether this block should use a projection shortcut (versus the default identity shortcut). This is usually `True` for the first block of a block group, which may change the number of filters and the resolution. is_training: `bool` if True, the model is in training mode. Returns: The output `Tensor` of the block. """ shortcut = inputs if use_projection: # Projection shortcut only in first block within a group. Bottleneck # blocks end with 4 times the number of filters. filters_out = 4 * filters shortcut = self.conv2d_fixed_padding( inputs=inputs, filters=filters_out, kernel_size=1, strides=strides) shortcut = self._norm_activation(use_activation=False)( shortcut, is_training=is_training) inputs = self.conv2d_fixed_padding( inputs=inputs, filters=filters, kernel_size=1, strides=1) inputs = self._norm_activation()(inputs, is_training=is_training) inputs = self.conv2d_fixed_padding( inputs=inputs, filters=filters, kernel_size=3, strides=strides) inputs = self._norm_activation()(inputs, is_training=is_training) inputs = self.conv2d_fixed_padding( inputs=inputs, filters=4 * filters, kernel_size=1, strides=1) inputs = self._norm_activation( use_activation=False, init_zero=True)( inputs, is_training=is_training) return self._activation_op(inputs + shortcut) def block_group(self, inputs, filters, block_fn, blocks, strides, name, is_training): """Creates one group of blocks for the ResNet model. Args: inputs: `Tensor` of size `[batch, channels, height, width]`. filters: `int` number of filters for the first convolution of the layer. block_fn: `function` for the block to use within the model blocks: `int` number of blocks contained in the layer. strides: `int` stride to use for the first convolution of the layer. If greater than 1, this layer will downsample the input. name: `str`name for the Tensor output of the block layer. is_training: `bool` if True, the model is in training mode. Returns: The output `Tensor` of the block layer. """ # Only the first block per block_group uses projection shortcut and strides. inputs = block_fn( inputs, filters, strides, use_projection=True, is_training=is_training) for _ in range(1, blocks): inputs = block_fn(inputs, filters, 1, is_training=is_training) return tf.identity(inputs, name) def resnet_v1_generator(self, block_fn, layers): """Generator for ResNet v1 models. Args: block_fn: `function` for the block to use within the model. Either `residual_block` or `bottleneck_block`. layers: list of 4 `int`s denoting the number of blocks to include in each of the 4 block groups. Each group consists of blocks that take inputs of the same resolution. Returns: Model `function` that takes in `inputs` and `is_training` and returns the output `Tensor` of the ResNet model. """ def model(inputs, is_training=None): """Creation of the model graph.""" inputs = self.conv2d_fixed_padding( inputs=inputs, filters=64, kernel_size=7, strides=2) inputs = tf.identity(inputs, 'initial_conv') inputs = self._norm_activation()(inputs, is_training=is_training) inputs = tf.keras.layers.MaxPool2D( pool_size=3, strides=2, padding='SAME', data_format=self._data_format)( inputs) inputs = tf.identity(inputs, 'initial_max_pool') c2 = self.block_group( inputs=inputs, filters=64, block_fn=block_fn, blocks=layers[0], strides=1, name='block_group1', is_training=is_training) c3 = self.block_group( inputs=c2, filters=128, block_fn=block_fn, blocks=layers[1], strides=2, name='block_group2', is_training=is_training) c4 = self.block_group( inputs=c3, filters=256, block_fn=block_fn, blocks=layers[2], strides=2, name='block_group3', is_training=is_training) c5 = self.block_group( inputs=c4, filters=512, block_fn=block_fn, blocks=layers[3], strides=2, name='block_group4', is_training=is_training) return {2: c2, 3: c3, 4: c4, 5: c5} return model
13,159
36.280453
81
py
models
models-master/official/legacy/detection/modeling/architecture/nn_ops.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Neural network operations commonly shared by the architectures.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import functools import tensorflow as tf class NormActivation(tf.keras.layers.Layer): """Combined Normalization and Activation layers.""" def __init__(self, momentum=0.997, epsilon=1e-4, trainable=True, init_zero=False, use_activation=True, activation='relu', fused=True, name=None): """A class to construct layers for a batch normalization followed by a ReLU. Args: momentum: momentum for the moving average. epsilon: small float added to variance to avoid dividing by zero. trainable: `bool`, if True also add variables to the graph collection GraphKeys.TRAINABLE_VARIABLES. If False, freeze batch normalization layer. init_zero: `bool` if True, initializes scale parameter of batch normalization with 0. If False, initialize it with 1. use_activation: `bool`, whether to add the optional activation layer after the batch normalization layer. activation: 'string', the type of the activation layer. Currently support `relu` and `swish`. fused: `bool` fused option in batch normalziation. name: `str` name for the operation. """ super(NormActivation, self).__init__(trainable=trainable) if init_zero: gamma_initializer = tf.keras.initializers.Zeros() else: gamma_initializer = tf.keras.initializers.Ones() self._normalization_op = tf.keras.layers.BatchNormalization( momentum=momentum, epsilon=epsilon, center=True, scale=True, trainable=trainable, fused=fused, gamma_initializer=gamma_initializer, name=name) self._use_activation = use_activation if activation == 'relu': self._activation_op = tf.nn.relu elif activation == 'swish': self._activation_op = tf.nn.swish else: raise ValueError('Unsupported activation `{}`.'.format(activation)) def __call__(self, inputs, is_training=None): """Builds the normalization layer followed by an optional activation layer. Args: inputs: `Tensor` of shape `[batch, channels, ...]`. is_training: `boolean`, if True if model is in training mode. Returns: A normalized `Tensor` with the same `data_format`. """ # We will need to keep training=None by default, so that it can be inherit # from keras.Model.training if is_training and self.trainable: is_training = True inputs = self._normalization_op(inputs, training=is_training) if self._use_activation: inputs = self._activation_op(inputs) return inputs def norm_activation_builder(momentum=0.997, epsilon=1e-4, trainable=True, activation='relu', **kwargs): return functools.partial( NormActivation, momentum=momentum, epsilon=epsilon, trainable=trainable, activation=activation, **kwargs)
3,853
34.036364
80
py
models
models-master/official/legacy/detection/modeling/architecture/heads.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Classes to build various prediction heads in all supported models.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import functools import numpy as np import tensorflow as tf from official.legacy.detection.modeling.architecture import nn_ops from official.legacy.detection.ops import spatial_transform_ops class RpnHead(tf.keras.layers.Layer): """Region Proposal Network head.""" def __init__( self, min_level, max_level, anchors_per_location, num_convs=2, num_filters=256, use_separable_conv=False, activation='relu', use_batch_norm=True, norm_activation=nn_ops.norm_activation_builder(activation='relu')): """Initialize params to build Region Proposal Network head. Args: min_level: `int` number of minimum feature level. max_level: `int` number of maximum feature level. anchors_per_location: `int` number of number of anchors per pixel location. num_convs: `int` number that represents the number of the intermediate conv layers before the prediction. num_filters: `int` number that represents the number of filters of the intermediate conv layers. use_separable_conv: `bool`, indicating whether the separable conv layers is used. activation: activation function. Support 'relu' and 'swish'. use_batch_norm: 'bool', indicating whether batchnorm layers are added. norm_activation: an operation that includes a normalization layer followed by an optional activation layer. """ super().__init__(autocast=False) self._min_level = min_level self._max_level = max_level self._anchors_per_location = anchors_per_location if activation == 'relu': self._activation_op = tf.nn.relu elif activation == 'swish': self._activation_op = tf.nn.swish else: raise ValueError('Unsupported activation `{}`.'.format(activation)) self._use_batch_norm = use_batch_norm if use_separable_conv: self._conv2d_op = functools.partial( tf.keras.layers.SeparableConv2D, depth_multiplier=1, bias_initializer=tf.zeros_initializer()) else: self._conv2d_op = functools.partial( tf.keras.layers.Conv2D, kernel_initializer=tf.keras.initializers.RandomNormal(stddev=0.01), bias_initializer=tf.zeros_initializer()) self._rpn_conv = self._conv2d_op( num_filters, kernel_size=(3, 3), strides=(1, 1), activation=(None if self._use_batch_norm else self._activation_op), padding='same', name='rpn') self._rpn_class_conv = self._conv2d_op( anchors_per_location, kernel_size=(1, 1), strides=(1, 1), padding='valid', name='rpn-class') self._rpn_box_conv = self._conv2d_op( 4 * anchors_per_location, kernel_size=(1, 1), strides=(1, 1), padding='valid', name='rpn-box') self._norm_activations = {} if self._use_batch_norm: for level in range(self._min_level, self._max_level + 1): self._norm_activations[level] = norm_activation(name='rpn-l%d-bn' % level) def _shared_rpn_heads(self, features, anchors_per_location, level, is_training): """Shared RPN heads.""" features = self._rpn_conv(features) if self._use_batch_norm: # The batch normalization layers are not shared between levels. features = self._norm_activations[level]( features, is_training=is_training) # Proposal classification scores scores = self._rpn_class_conv(features) # Proposal bbox regression deltas bboxes = self._rpn_box_conv(features) return scores, bboxes def call(self, features, is_training=None): scores_outputs = {} box_outputs = {} with tf.name_scope('rpn_head'): for level in range(self._min_level, self._max_level + 1): scores_output, box_output = self._shared_rpn_heads( features[level], self._anchors_per_location, level, is_training) scores_outputs[level] = scores_output box_outputs[level] = box_output return scores_outputs, box_outputs class OlnRpnHead(tf.keras.layers.Layer): """Region Proposal Network for Object Localization Network (OLN).""" def __init__( self, min_level, max_level, anchors_per_location, num_convs=2, num_filters=256, use_separable_conv=False, activation='relu', use_batch_norm=True, norm_activation=nn_ops.norm_activation_builder(activation='relu')): """Initialize params to build Region Proposal Network head. Args: min_level: `int` number of minimum feature level. max_level: `int` number of maximum feature level. anchors_per_location: `int` number of number of anchors per pixel location. num_convs: `int` number that represents the number of the intermediate conv layers before the prediction. num_filters: `int` number that represents the number of filters of the intermediate conv layers. use_separable_conv: `bool`, indicating whether the separable conv layers is used. activation: activation function. Support 'relu' and 'swish'. use_batch_norm: 'bool', indicating whether batchnorm layers are added. norm_activation: an operation that includes a normalization layer followed by an optional activation layer. """ self._min_level = min_level self._max_level = max_level self._anchors_per_location = anchors_per_location if activation == 'relu': self._activation_op = tf.nn.relu elif activation == 'swish': self._activation_op = tf.nn.swish else: raise ValueError('Unsupported activation `{}`.'.format(activation)) self._use_batch_norm = use_batch_norm if use_separable_conv: self._conv2d_op = functools.partial( tf.keras.layers.SeparableConv2D, depth_multiplier=1, bias_initializer=tf.zeros_initializer()) else: self._conv2d_op = functools.partial( tf.keras.layers.Conv2D, kernel_initializer=tf.keras.initializers.RandomNormal(stddev=0.01), bias_initializer=tf.zeros_initializer()) self._rpn_conv = self._conv2d_op( num_filters, kernel_size=(3, 3), strides=(1, 1), activation=(None if self._use_batch_norm else self._activation_op), padding='same', name='rpn') self._rpn_class_conv = self._conv2d_op( anchors_per_location, kernel_size=(1, 1), strides=(1, 1), padding='valid', name='rpn-class') self._rpn_box_conv = self._conv2d_op( 4 * anchors_per_location, kernel_size=(1, 1), strides=(1, 1), padding='valid', name='rpn-box-lrtb') self._rpn_center_conv = self._conv2d_op( anchors_per_location, kernel_size=(1, 1), strides=(1, 1), padding='valid', name='rpn-centerness') self._norm_activations = {} if self._use_batch_norm: for level in range(self._min_level, self._max_level + 1): self._norm_activations[level] = norm_activation(name='rpn-l%d-bn' % level) def _shared_rpn_heads(self, features, anchors_per_location, level, is_training): """Shared RPN heads.""" features = self._rpn_conv(features) if self._use_batch_norm: # The batch normalization layers are not shared between levels. features = self._norm_activations[level]( features, is_training=is_training) # Feature L2 normalization for training stability features = tf.math.l2_normalize( features, axis=-1, name='rpn-norm',) # Proposal classification scores scores = self._rpn_class_conv(features) # Proposal bbox regression deltas bboxes = self._rpn_box_conv(features) # Proposal centerness scores centers = self._rpn_center_conv(features) return scores, bboxes, centers def __call__(self, features, is_training=None): scores_outputs = {} box_outputs = {} center_outputs = {} with tf.name_scope('rpn_head'): for level in range(self._min_level, self._max_level + 1): scores_output, box_output, center_output = self._shared_rpn_heads( features[level], self._anchors_per_location, level, is_training) scores_outputs[level] = scores_output box_outputs[level] = box_output center_outputs[level] = center_output return scores_outputs, box_outputs, center_outputs class FastrcnnHead(tf.keras.layers.Layer): """Fast R-CNN box head.""" def __init__( self, num_classes, num_convs=0, num_filters=256, use_separable_conv=False, num_fcs=2, fc_dims=1024, activation='relu', use_batch_norm=True, norm_activation=nn_ops.norm_activation_builder(activation='relu')): """Initialize params to build Fast R-CNN box head. Args: num_classes: a integer for the number of classes. num_convs: `int` number that represents the number of the intermediate conv layers before the FC layers. num_filters: `int` number that represents the number of filters of the intermediate conv layers. use_separable_conv: `bool`, indicating whether the separable conv layers is used. num_fcs: `int` number that represents the number of FC layers before the predictions. fc_dims: `int` number that represents the number of dimension of the FC layers. activation: activation function. Support 'relu' and 'swish'. use_batch_norm: 'bool', indicating whether batchnorm layers are added. norm_activation: an operation that includes a normalization layer followed by an optional activation layer. """ super(FastrcnnHead, self).__init__(autocast=False) self._num_classes = num_classes self._num_convs = num_convs self._num_filters = num_filters if use_separable_conv: self._conv2d_op = functools.partial( tf.keras.layers.SeparableConv2D, depth_multiplier=1, bias_initializer=tf.zeros_initializer()) else: self._conv2d_op = functools.partial( tf.keras.layers.Conv2D, kernel_initializer=tf.keras.initializers.VarianceScaling( scale=2, mode='fan_out', distribution='untruncated_normal'), bias_initializer=tf.zeros_initializer()) self._num_fcs = num_fcs self._fc_dims = fc_dims if activation == 'relu': self._activation_op = tf.nn.relu elif activation == 'swish': self._activation_op = tf.nn.swish else: raise ValueError('Unsupported activation `{}`.'.format(activation)) self._use_batch_norm = use_batch_norm self._norm_activation = norm_activation self._conv_ops = [] self._conv_bn_ops = [] for i in range(self._num_convs): self._conv_ops.append( self._conv2d_op( self._num_filters, kernel_size=(3, 3), strides=(1, 1), padding='same', dilation_rate=(1, 1), activation=(None if self._use_batch_norm else self._activation_op), name='conv_{}'.format(i))) if self._use_batch_norm: self._conv_bn_ops.append(self._norm_activation()) self._fc_ops = [] self._fc_bn_ops = [] for i in range(self._num_fcs): self._fc_ops.append( tf.keras.layers.Dense( units=self._fc_dims, activation=(None if self._use_batch_norm else self._activation_op), name='fc{}'.format(i))) if self._use_batch_norm: self._fc_bn_ops.append(self._norm_activation(fused=False)) self._class_predict = tf.keras.layers.Dense( self._num_classes, kernel_initializer=tf.keras.initializers.RandomNormal(stddev=0.01), bias_initializer=tf.zeros_initializer(), name='class-predict') self._box_predict = tf.keras.layers.Dense( self._num_classes * 4, kernel_initializer=tf.keras.initializers.RandomNormal(stddev=0.001), bias_initializer=tf.zeros_initializer(), name='box-predict') def call(self, roi_features, is_training=None): """Box and class branches for the Mask-RCNN model. Args: roi_features: A ROI feature tensor of shape [batch_size, num_rois, height_l, width_l, num_filters]. is_training: `boolean`, if True if model is in training mode. Returns: class_outputs: a tensor with a shape of [batch_size, num_rois, num_classes], representing the class predictions. box_outputs: a tensor with a shape of [batch_size, num_rois, num_classes * 4], representing the box predictions. """ with tf.name_scope( 'fast_rcnn_head'): # reshape inputs beofre FC. _, num_rois, height, width, filters = roi_features.get_shape().as_list() net = tf.reshape(roi_features, [-1, height, width, filters]) for i in range(self._num_convs): net = self._conv_ops[i](net) if self._use_batch_norm: net = self._conv_bn_ops[i](net, is_training=is_training) filters = self._num_filters if self._num_convs > 0 else filters net = tf.reshape(net, [-1, num_rois, height * width * filters]) for i in range(self._num_fcs): net = self._fc_ops[i](net) if self._use_batch_norm: net = self._fc_bn_ops[i](net, is_training=is_training) class_outputs = self._class_predict(net) box_outputs = self._box_predict(net) return class_outputs, box_outputs class OlnBoxScoreHead(tf.keras.layers.Layer): """Box head of Object Localization Network (OLN).""" def __init__( self, num_classes, num_convs=0, num_filters=256, use_separable_conv=False, num_fcs=2, fc_dims=1024, activation='relu', use_batch_norm=True, norm_activation=nn_ops.norm_activation_builder(activation='relu')): """Initialize params to build OLN box head. Args: num_classes: a integer for the number of classes. num_convs: `int` number that represents the number of the intermediate conv layers before the FC layers. num_filters: `int` number that represents the number of filters of the intermediate conv layers. use_separable_conv: `bool`, indicating whether the separable conv layers is used. num_fcs: `int` number that represents the number of FC layers before the predictions. fc_dims: `int` number that represents the number of dimension of the FC layers. activation: activation function. Support 'relu' and 'swish'. use_batch_norm: 'bool', indicating whether batchnorm layers are added. norm_activation: an operation that includes a normalization layer followed by an optional activation layer. """ self._num_classes = num_classes self._num_convs = num_convs self._num_filters = num_filters if use_separable_conv: self._conv2d_op = functools.partial( tf.keras.layers.SeparableConv2D, depth_multiplier=1, bias_initializer=tf.zeros_initializer()) else: self._conv2d_op = functools.partial( tf.keras.layers.Conv2D, kernel_initializer=tf.keras.initializers.VarianceScaling( scale=2, mode='fan_out', distribution='untruncated_normal'), bias_initializer=tf.zeros_initializer()) self._num_fcs = num_fcs self._fc_dims = fc_dims if activation == 'relu': self._activation_op = tf.nn.relu elif activation == 'swish': self._activation_op = tf.nn.swish else: raise ValueError('Unsupported activation `{}`.'.format(activation)) self._use_batch_norm = use_batch_norm self._norm_activation = norm_activation self._conv_ops = [] self._conv_bn_ops = [] for i in range(self._num_convs): self._conv_ops.append( self._conv2d_op( self._num_filters, kernel_size=(3, 3), strides=(1, 1), padding='same', dilation_rate=(1, 1), activation=(None if self._use_batch_norm else self._activation_op), name='conv_{}'.format(i))) if self._use_batch_norm: self._conv_bn_ops.append(self._norm_activation()) self._fc_ops = [] self._fc_bn_ops = [] for i in range(self._num_fcs): self._fc_ops.append( tf.keras.layers.Dense( units=self._fc_dims, activation=(None if self._use_batch_norm else self._activation_op), name='fc{}'.format(i))) if self._use_batch_norm: self._fc_bn_ops.append(self._norm_activation(fused=False)) self._class_predict = tf.keras.layers.Dense( self._num_classes, kernel_initializer=tf.keras.initializers.RandomNormal(stddev=0.01), bias_initializer=tf.zeros_initializer(), name='class-predict') self._box_predict = tf.keras.layers.Dense( self._num_classes * 4, kernel_initializer=tf.keras.initializers.RandomNormal(stddev=0.001), bias_initializer=tf.zeros_initializer(), name='box-predict') self._score_predict = tf.keras.layers.Dense( 1, kernel_initializer=tf.keras.initializers.RandomNormal(stddev=0.01), bias_initializer=tf.zeros_initializer(), name='score-predict') def __call__(self, roi_features, is_training=None): """Box and class branches for the Mask-RCNN model. Args: roi_features: A ROI feature tensor of shape [batch_size, num_rois, height_l, width_l, num_filters]. is_training: `boolean`, if True if model is in training mode. Returns: class_outputs: a tensor with a shape of [batch_size, num_rois, num_classes], representing the class predictions. box_outputs: a tensor with a shape of [batch_size, num_rois, num_classes * 4], representing the box predictions. """ with tf.name_scope('fast_rcnn_head'): # reshape inputs beofre FC. _, num_rois, height, width, filters = roi_features.get_shape().as_list() net = tf.reshape(roi_features, [-1, height, width, filters]) for i in range(self._num_convs): net = self._conv_ops[i](net) if self._use_batch_norm: net = self._conv_bn_ops[i](net, is_training=is_training) filters = self._num_filters if self._num_convs > 0 else filters net = tf.reshape(net, [-1, num_rois, height * width * filters]) for i in range(self._num_fcs): net = self._fc_ops[i](net) if self._use_batch_norm: net = self._fc_bn_ops[i](net, is_training=is_training) class_outputs = self._class_predict(net) box_outputs = self._box_predict(net) score_outputs = self._score_predict(net) return class_outputs, box_outputs, score_outputs class MaskrcnnHead(tf.keras.layers.Layer): """Mask R-CNN head.""" def __init__( self, num_classes, mask_target_size, num_convs=4, num_filters=256, use_separable_conv=False, activation='relu', use_batch_norm=True, norm_activation=nn_ops.norm_activation_builder(activation='relu')): """Initialize params to build Fast R-CNN head. Args: num_classes: a integer for the number of classes. mask_target_size: a integer that is the resolution of masks. num_convs: `int` number that represents the number of the intermediate conv layers before the prediction. num_filters: `int` number that represents the number of filters of the intermediate conv layers. use_separable_conv: `bool`, indicating whether the separable conv layers is used. activation: activation function. Support 'relu' and 'swish'. use_batch_norm: 'bool', indicating whether batchnorm layers are added. norm_activation: an operation that includes a normalization layer followed by an optional activation layer. """ super(MaskrcnnHead, self).__init__(autocast=False) self._num_classes = num_classes self._mask_target_size = mask_target_size self._num_convs = num_convs self._num_filters = num_filters if use_separable_conv: self._conv2d_op = functools.partial( tf.keras.layers.SeparableConv2D, depth_multiplier=1, bias_initializer=tf.zeros_initializer()) else: self._conv2d_op = functools.partial( tf.keras.layers.Conv2D, kernel_initializer=tf.keras.initializers.VarianceScaling( scale=2, mode='fan_out', distribution='untruncated_normal'), bias_initializer=tf.zeros_initializer()) if activation == 'relu': self._activation_op = tf.nn.relu elif activation == 'swish': self._activation_op = tf.nn.swish else: raise ValueError('Unsupported activation `{}`.'.format(activation)) self._use_batch_norm = use_batch_norm self._norm_activation = norm_activation self._conv2d_ops = [] for i in range(self._num_convs): self._conv2d_ops.append( self._conv2d_op( self._num_filters, kernel_size=(3, 3), strides=(1, 1), padding='same', dilation_rate=(1, 1), activation=(None if self._use_batch_norm else self._activation_op), name='mask-conv-l%d' % i)) self._mask_conv_transpose = tf.keras.layers.Conv2DTranspose( self._num_filters, kernel_size=(2, 2), strides=(2, 2), padding='valid', activation=(None if self._use_batch_norm else self._activation_op), kernel_initializer=tf.keras.initializers.VarianceScaling( scale=2, mode='fan_out', distribution='untruncated_normal'), bias_initializer=tf.zeros_initializer(), name='conv5-mask') with tf.name_scope('mask_head'): self._mask_conv2d_op = self._conv2d_op( self._num_classes, kernel_size=(1, 1), strides=(1, 1), padding='valid', name='mask_fcn_logits') def call(self, roi_features, class_indices, is_training=None): """Mask branch for the Mask-RCNN model. Args: roi_features: A ROI feature tensor of shape [batch_size, num_rois, height_l, width_l, num_filters]. class_indices: a Tensor of shape [batch_size, num_rois], indicating which class the ROI is. is_training: `boolean`, if True if model is in training mode. Returns: mask_outputs: a tensor with a shape of [batch_size, num_masks, mask_height, mask_width, num_classes], representing the mask predictions. fg_gather_indices: a tensor with a shape of [batch_size, num_masks, 2], representing the fg mask targets. Raises: ValueError: If boxes is not a rank-3 tensor or the last dimension of boxes is not 4. """ with tf.name_scope('mask_head'): _, num_rois, height, width, filters = roi_features.get_shape().as_list() net = tf.reshape(roi_features, [-1, height, width, filters]) for i in range(self._num_convs): net = self._conv2d_ops[i](net) if self._use_batch_norm: net = self._norm_activation()(net, is_training=is_training) net = self._mask_conv_transpose(net) if self._use_batch_norm: net = self._norm_activation()(net, is_training=is_training) mask_outputs = self._mask_conv2d_op(net) mask_outputs = tf.reshape(mask_outputs, [ -1, num_rois, self._mask_target_size, self._mask_target_size, self._num_classes ]) with tf.name_scope('masks_post_processing'): mask_outputs = tf.gather( mask_outputs, tf.cast(class_indices, tf.int32), axis=-1, batch_dims=2, ) return mask_outputs class RetinanetHead(object): """RetinaNet head.""" def __init__( self, min_level, max_level, num_classes, anchors_per_location, num_convs=4, num_filters=256, use_separable_conv=False, norm_activation=nn_ops.norm_activation_builder(activation='relu')): """Initialize params to build RetinaNet head. Args: min_level: `int` number of minimum feature level. max_level: `int` number of maximum feature level. num_classes: `int` number of classification categories. anchors_per_location: `int` number of anchors per pixel location. num_convs: `int` number of stacked convolution before the last prediction layer. num_filters: `int` number of filters used in the head architecture. use_separable_conv: `bool` to indicate whether to use separable convoluation. norm_activation: an operation that includes a normalization layer followed by an optional activation layer. """ self._min_level = min_level self._max_level = max_level self._num_classes = num_classes self._anchors_per_location = anchors_per_location self._num_convs = num_convs self._num_filters = num_filters self._use_separable_conv = use_separable_conv with tf.name_scope('class_net') as scope_name: self._class_name_scope = tf.name_scope(scope_name) with tf.name_scope('box_net') as scope_name: self._box_name_scope = tf.name_scope(scope_name) self._build_class_net_layers(norm_activation) self._build_box_net_layers(norm_activation) def _class_net_batch_norm_name(self, i, level): return 'class-%d-%d' % (i, level) def _box_net_batch_norm_name(self, i, level): return 'box-%d-%d' % (i, level) def _build_class_net_layers(self, norm_activation): """Build re-usable layers for class prediction network.""" if self._use_separable_conv: self._class_predict = tf.keras.layers.SeparableConv2D( self._num_classes * self._anchors_per_location, kernel_size=(3, 3), bias_initializer=tf.constant_initializer(-np.log((1 - 0.01) / 0.01)), padding='same', name='class-predict') else: self._class_predict = tf.keras.layers.Conv2D( self._num_classes * self._anchors_per_location, kernel_size=(3, 3), bias_initializer=tf.constant_initializer(-np.log((1 - 0.01) / 0.01)), kernel_initializer=tf.keras.initializers.RandomNormal(stddev=1e-5), padding='same', name='class-predict') self._class_conv = [] self._class_norm_activation = {} for i in range(self._num_convs): if self._use_separable_conv: self._class_conv.append( tf.keras.layers.SeparableConv2D( self._num_filters, kernel_size=(3, 3), bias_initializer=tf.zeros_initializer(), activation=None, padding='same', name='class-' + str(i))) else: self._class_conv.append( tf.keras.layers.Conv2D( self._num_filters, kernel_size=(3, 3), bias_initializer=tf.zeros_initializer(), kernel_initializer=tf.keras.initializers.RandomNormal( stddev=0.01), activation=None, padding='same', name='class-' + str(i))) for level in range(self._min_level, self._max_level + 1): name = self._class_net_batch_norm_name(i, level) self._class_norm_activation[name] = norm_activation(name=name) def _build_box_net_layers(self, norm_activation): """Build re-usable layers for box prediction network.""" if self._use_separable_conv: self._box_predict = tf.keras.layers.SeparableConv2D( 4 * self._anchors_per_location, kernel_size=(3, 3), bias_initializer=tf.zeros_initializer(), padding='same', name='box-predict') else: self._box_predict = tf.keras.layers.Conv2D( 4 * self._anchors_per_location, kernel_size=(3, 3), bias_initializer=tf.zeros_initializer(), kernel_initializer=tf.keras.initializers.RandomNormal(stddev=1e-5), padding='same', name='box-predict') self._box_conv = [] self._box_norm_activation = {} for i in range(self._num_convs): if self._use_separable_conv: self._box_conv.append( tf.keras.layers.SeparableConv2D( self._num_filters, kernel_size=(3, 3), activation=None, bias_initializer=tf.zeros_initializer(), padding='same', name='box-' + str(i))) else: self._box_conv.append( tf.keras.layers.Conv2D( self._num_filters, kernel_size=(3, 3), activation=None, bias_initializer=tf.zeros_initializer(), kernel_initializer=tf.keras.initializers.RandomNormal( stddev=0.01), padding='same', name='box-' + str(i))) for level in range(self._min_level, self._max_level + 1): name = self._box_net_batch_norm_name(i, level) self._box_norm_activation[name] = norm_activation(name=name) def __call__(self, fpn_features, is_training=None): """Returns outputs of RetinaNet head.""" class_outputs = {} box_outputs = {} with tf.name_scope('retinanet_head'): for level in range(self._min_level, self._max_level + 1): features = fpn_features[level] class_outputs[level] = self.class_net( features, level, is_training=is_training) box_outputs[level] = self.box_net( features, level, is_training=is_training) return class_outputs, box_outputs def class_net(self, features, level, is_training): """Class prediction network for RetinaNet.""" with self._class_name_scope: for i in range(self._num_convs): features = self._class_conv[i](features) # The convolution layers in the class net are shared among all levels, # but each level has its batch normlization to capture the statistical # difference among different levels. name = self._class_net_batch_norm_name(i, level) features = self._class_norm_activation[name]( features, is_training=is_training) classes = self._class_predict(features) return classes def box_net(self, features, level, is_training=None): """Box regression network for RetinaNet.""" with self._box_name_scope: for i in range(self._num_convs): features = self._box_conv[i](features) # The convolution layers in the box net are shared among all levels, but # each level has its batch normlization to capture the statistical # difference among different levels. name = self._box_net_batch_norm_name(i, level) features = self._box_norm_activation[name]( features, is_training=is_training) boxes = self._box_predict(features) return boxes # TODO(yeqing): Refactor this class when it is ready for var_scope reuse. class ShapemaskPriorHead(object): """ShapeMask Prior head.""" def __init__(self, num_classes, num_downsample_channels, mask_crop_size, use_category_for_mask, shape_prior_path): """Initialize params to build RetinaNet head. Args: num_classes: Number of output classes. num_downsample_channels: number of channels in mask branch. mask_crop_size: feature crop size. use_category_for_mask: use class information in mask branch. shape_prior_path: the path to load shape priors. """ self._mask_num_classes = num_classes if use_category_for_mask else 1 self._num_downsample_channels = num_downsample_channels self._mask_crop_size = mask_crop_size self._shape_prior_path = shape_prior_path self._use_category_for_mask = use_category_for_mask self._shape_prior_fc = tf.keras.layers.Dense( self._num_downsample_channels, name='shape-prior-fc') def __call__(self, fpn_features, boxes, outer_boxes, classes, is_training): """Generate the detection priors from the box detections and FPN features. This corresponds to the Fig. 4 of the ShapeMask paper at https://arxiv.org/pdf/1904.03239.pdf Args: fpn_features: a dictionary of FPN features. boxes: a float tensor of shape [batch_size, num_instances, 4] representing the tight gt boxes from dataloader/detection. outer_boxes: a float tensor of shape [batch_size, num_instances, 4] representing the loose gt boxes from dataloader/detection. classes: a int Tensor of shape [batch_size, num_instances] of instance classes. is_training: training mode or not. Returns: instance_features: a float Tensor of shape [batch_size * num_instances, mask_crop_size, mask_crop_size, num_downsample_channels]. This is the instance feature crop. detection_priors: A float Tensor of shape [batch_size * num_instances, mask_size, mask_size, 1]. """ with tf.name_scope('prior_mask'): batch_size, num_instances, _ = boxes.get_shape().as_list() outer_boxes = tf.cast(outer_boxes, tf.float32) boxes = tf.cast(boxes, tf.float32) instance_features = spatial_transform_ops.multilevel_crop_and_resize( fpn_features, outer_boxes, output_size=self._mask_crop_size) instance_features = self._shape_prior_fc(instance_features) shape_priors = self._get_priors() # Get uniform priors for each outer box. uniform_priors = tf.ones([ batch_size, num_instances, self._mask_crop_size, self._mask_crop_size ]) uniform_priors = spatial_transform_ops.crop_mask_in_target_box( uniform_priors, boxes, outer_boxes, self._mask_crop_size) # Classify shape priors using uniform priors + instance features. prior_distribution = self._classify_shape_priors( tf.cast(instance_features, tf.float32), uniform_priors, classes) instance_priors = tf.gather(shape_priors, classes) instance_priors *= tf.expand_dims( tf.expand_dims(tf.cast(prior_distribution, tf.float32), axis=-1), axis=-1) instance_priors = tf.reduce_sum(instance_priors, axis=2) detection_priors = spatial_transform_ops.crop_mask_in_target_box( instance_priors, boxes, outer_boxes, self._mask_crop_size) return instance_features, detection_priors def _get_priors(self): """Load shape priors from file.""" # loads class specific or agnostic shape priors if self._shape_prior_path: # Priors are loaded into shape [mask_num_classes, num_clusters, 32, 32]. priors = np.load(tf.io.gfile.GFile(self._shape_prior_path, 'rb')) priors = tf.convert_to_tensor(priors, dtype=tf.float32) self._num_clusters = priors.get_shape().as_list()[1] else: # If prior path does not exist, do not use priors, i.e., pirors equal to # uniform empty 32x32 patch. self._num_clusters = 1 priors = tf.zeros([ self._mask_num_classes, self._num_clusters, self._mask_crop_size, self._mask_crop_size ]) return priors def _classify_shape_priors(self, features, uniform_priors, classes): """Classify the uniform prior by predicting the shape modes. Classify the object crop features into K modes of the clusters for each category. Args: features: A float Tensor of shape [batch_size, num_instances, mask_size, mask_size, num_channels]. uniform_priors: A float Tensor of shape [batch_size, num_instances, mask_size, mask_size] representing the uniform detection priors. classes: A int Tensor of shape [batch_size, num_instances] of detection class ids. Returns: prior_distribution: A float Tensor of shape [batch_size, num_instances, num_clusters] representing the classifier output probability over all possible shapes. """ batch_size, num_instances, _, _, _ = features.get_shape().as_list() features *= tf.expand_dims(uniform_priors, axis=-1) # Reduce spatial dimension of features. The features have shape # [batch_size, num_instances, num_channels]. features = tf.reduce_mean(features, axis=(2, 3)) logits = tf.keras.layers.Dense( self._mask_num_classes * self._num_clusters, kernel_initializer=tf.random_normal_initializer(stddev=0.01), name='classify-shape-prior-fc')(features) logits = tf.reshape( logits, [batch_size, num_instances, self._mask_num_classes, self._num_clusters]) if self._use_category_for_mask: logits = tf.gather(logits, tf.expand_dims(classes, axis=-1), batch_dims=2) logits = tf.squeeze(logits, axis=2) else: logits = logits[:, :, 0, :] distribution = tf.nn.softmax(logits, name='shape_prior_weights') return distribution class ShapemaskCoarsemaskHead(object): """ShapemaskCoarsemaskHead head.""" def __init__(self, num_classes, num_downsample_channels, mask_crop_size, use_category_for_mask, num_convs, norm_activation=nn_ops.norm_activation_builder()): """Initialize params to build ShapeMask coarse and fine prediction head. Args: num_classes: `int` number of mask classification categories. num_downsample_channels: `int` number of filters at mask head. mask_crop_size: feature crop size. use_category_for_mask: use class information in mask branch. num_convs: `int` number of stacked convolution before the last prediction layer. norm_activation: an operation that includes a normalization layer followed by an optional activation layer. """ self._mask_num_classes = num_classes if use_category_for_mask else 1 self._use_category_for_mask = use_category_for_mask self._num_downsample_channels = num_downsample_channels self._mask_crop_size = mask_crop_size self._num_convs = num_convs self._norm_activation = norm_activation self._coarse_mask_fc = tf.keras.layers.Dense( self._num_downsample_channels, name='coarse-mask-fc') self._class_conv = [] self._class_norm_activation = [] for i in range(self._num_convs): self._class_conv.append( tf.keras.layers.Conv2D( self._num_downsample_channels, kernel_size=(3, 3), bias_initializer=tf.zeros_initializer(), kernel_initializer=tf.keras.initializers.RandomNormal( stddev=0.01), padding='same', name='coarse-mask-class-%d' % i)) self._class_norm_activation.append( norm_activation(name='coarse-mask-class-%d-bn' % i)) self._class_predict = tf.keras.layers.Conv2D( self._mask_num_classes, kernel_size=(1, 1), # Focal loss bias initialization to have foreground 0.01 probability. bias_initializer=tf.constant_initializer(-np.log((1 - 0.01) / 0.01)), kernel_initializer=tf.keras.initializers.RandomNormal(stddev=0.01), padding='same', name='coarse-mask-class-predict') def __call__(self, features, detection_priors, classes, is_training): """Generate instance masks from FPN features and detection priors. This corresponds to the Fig. 5-6 of the ShapeMask paper at https://arxiv.org/pdf/1904.03239.pdf Args: features: a float Tensor of shape [batch_size, num_instances, mask_crop_size, mask_crop_size, num_downsample_channels]. This is the instance feature crop. detection_priors: a float Tensor of shape [batch_size, num_instances, mask_crop_size, mask_crop_size, 1]. This is the detection prior for the instance. classes: a int Tensor of shape [batch_size, num_instances] of instance classes. is_training: a bool indicating whether in training mode. Returns: mask_outputs: instance mask prediction as a float Tensor of shape [batch_size, num_instances, mask_size, mask_size]. """ with tf.name_scope('coarse_mask'): # Transform detection priors to have the same dimension as features. detection_priors = tf.expand_dims(detection_priors, axis=-1) detection_priors = self._coarse_mask_fc(detection_priors) features += detection_priors mask_logits = self.decoder_net(features, is_training) # Gather the logits with right input class. if self._use_category_for_mask: mask_logits = tf.transpose(mask_logits, [0, 1, 4, 2, 3]) mask_logits = tf.gather( mask_logits, tf.expand_dims(classes, -1), batch_dims=2) mask_logits = tf.squeeze(mask_logits, axis=2) else: mask_logits = mask_logits[..., 0] return mask_logits def decoder_net(self, features, is_training=False): """Coarse mask decoder network architecture. Args: features: A tensor of size [batch, height_in, width_in, channels_in]. is_training: Whether batch_norm layers are in training mode. Returns: images: A feature tensor of size [batch, output_size, output_size, num_channels] """ (batch_size, num_instances, height, width, num_channels) = features.get_shape().as_list() features = tf.reshape( features, [batch_size * num_instances, height, width, num_channels]) for i in range(self._num_convs): features = self._class_conv[i](features) features = self._class_norm_activation[i]( features, is_training=is_training) mask_logits = self._class_predict(features) mask_logits = tf.reshape( mask_logits, [batch_size, num_instances, height, width, self._mask_num_classes]) return mask_logits class ShapemaskFinemaskHead(object): """ShapemaskFinemaskHead head.""" def __init__(self, num_classes, num_downsample_channels, mask_crop_size, use_category_for_mask, num_convs, upsample_factor, norm_activation=nn_ops.norm_activation_builder()): """Initialize params to build ShapeMask coarse and fine prediction head. Args: num_classes: `int` number of mask classification categories. num_downsample_channels: `int` number of filters at mask head. mask_crop_size: feature crop size. use_category_for_mask: use class information in mask branch. num_convs: `int` number of stacked convolution before the last prediction layer. upsample_factor: `int` number of fine mask upsampling factor. norm_activation: an operation that includes a batch normalization layer followed by a relu layer(optional). """ self._use_category_for_mask = use_category_for_mask self._mask_num_classes = num_classes if use_category_for_mask else 1 self._num_downsample_channels = num_downsample_channels self._mask_crop_size = mask_crop_size self._num_convs = num_convs self.up_sample_factor = upsample_factor self._fine_mask_fc = tf.keras.layers.Dense( self._num_downsample_channels, name='fine-mask-fc') self._upsample_conv = tf.keras.layers.Conv2DTranspose( self._num_downsample_channels, (self.up_sample_factor, self.up_sample_factor), (self.up_sample_factor, self.up_sample_factor), name='fine-mask-conv2d-tran') self._fine_class_conv = [] self._fine_class_bn = [] for i in range(self._num_convs): self._fine_class_conv.append( tf.keras.layers.Conv2D( self._num_downsample_channels, kernel_size=(3, 3), bias_initializer=tf.zeros_initializer(), kernel_initializer=tf.keras.initializers.RandomNormal( stddev=0.01), activation=None, padding='same', name='fine-mask-class-%d' % i)) self._fine_class_bn.append( norm_activation(name='fine-mask-class-%d-bn' % i)) self._class_predict_conv = tf.keras.layers.Conv2D( self._mask_num_classes, kernel_size=(1, 1), # Focal loss bias initialization to have foreground 0.01 probability. bias_initializer=tf.constant_initializer(-np.log((1 - 0.01) / 0.01)), kernel_initializer=tf.keras.initializers.RandomNormal(stddev=0.01), padding='same', name='fine-mask-class-predict') def __call__(self, features, mask_logits, classes, is_training): """Generate instance masks from FPN features and detection priors. This corresponds to the Fig. 5-6 of the ShapeMask paper at https://arxiv.org/pdf/1904.03239.pdf Args: features: a float Tensor of shape [batch_size, num_instances, mask_crop_size, mask_crop_size, num_downsample_channels]. This is the instance feature crop. mask_logits: a float Tensor of shape [batch_size, num_instances, mask_crop_size, mask_crop_size] indicating predicted mask logits. classes: a int Tensor of shape [batch_size, num_instances] of instance classes. is_training: a bool indicating whether in training mode. Returns: mask_outputs: instance mask prediction as a float Tensor of shape [batch_size, num_instances, mask_size, mask_size]. """ # Extract the foreground mean features # with tf.variable_scope('fine_mask', reuse=tf.AUTO_REUSE): with tf.name_scope('fine_mask'): mask_probs = tf.nn.sigmoid(mask_logits) # Compute instance embedding for hard average. binary_mask = tf.cast(tf.greater(mask_probs, 0.5), features.dtype) instance_embedding = tf.reduce_sum( features * tf.expand_dims(binary_mask, axis=-1), axis=(2, 3)) instance_embedding /= tf.expand_dims( tf.reduce_sum(binary_mask, axis=(2, 3)) + 1e-20, axis=-1) # Take the difference between crop features and mean instance features. features -= tf.expand_dims( tf.expand_dims(instance_embedding, axis=2), axis=2) features += self._fine_mask_fc(tf.expand_dims(mask_probs, axis=-1)) # Decoder to generate upsampled segmentation mask. mask_logits = self.decoder_net(features, is_training) if self._use_category_for_mask: mask_logits = tf.transpose(mask_logits, [0, 1, 4, 2, 3]) mask_logits = tf.gather( mask_logits, tf.expand_dims(classes, -1), batch_dims=2) mask_logits = tf.squeeze(mask_logits, axis=2) else: mask_logits = mask_logits[..., 0] return mask_logits def decoder_net(self, features, is_training=False): """Fine mask decoder network architecture. Args: features: A tensor of size [batch, height_in, width_in, channels_in]. is_training: Whether batch_norm layers are in training mode. Returns: images: A feature tensor of size [batch, output_size, output_size, num_channels], where output size is self._gt_upsample_scale times that of input. """ (batch_size, num_instances, height, width, num_channels) = features.get_shape().as_list() features = tf.reshape( features, [batch_size * num_instances, height, width, num_channels]) for i in range(self._num_convs): features = self._fine_class_conv[i](features) features = self._fine_class_bn[i](features, is_training=is_training) if self.up_sample_factor > 1: features = self._upsample_conv(features) # Predict per-class instance masks. mask_logits = self._class_predict_conv(features) mask_logits = tf.reshape(mask_logits, [ batch_size, num_instances, height * self.up_sample_factor, width * self.up_sample_factor, self._mask_num_classes ]) return mask_logits
48,977
37.44427
80
py
models
models-master/official/legacy/detection/modeling/architecture/factory.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Model architecture factory.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from official.legacy.detection.modeling.architecture import fpn from official.legacy.detection.modeling.architecture import heads from official.legacy.detection.modeling.architecture import identity from official.legacy.detection.modeling.architecture import nn_ops from official.legacy.detection.modeling.architecture import resnet from official.legacy.detection.modeling.architecture import spinenet def norm_activation_generator(params): return nn_ops.norm_activation_builder( momentum=params.batch_norm_momentum, epsilon=params.batch_norm_epsilon, trainable=params.batch_norm_trainable, activation=params.activation) def backbone_generator(params): """Generator function for various backbone models.""" if params.architecture.backbone == 'resnet': resnet_params = params.resnet backbone_fn = resnet.Resnet( resnet_depth=resnet_params.resnet_depth, activation=params.norm_activation.activation, norm_activation=norm_activation_generator( params.norm_activation)) elif params.architecture.backbone == 'spinenet': spinenet_params = params.spinenet backbone_fn = spinenet.SpineNetBuilder(model_id=spinenet_params.model_id) else: raise ValueError('Backbone model `{}` is not supported.' .format(params.architecture.backbone)) return backbone_fn def multilevel_features_generator(params): """Generator function for various FPN models.""" if params.architecture.multilevel_features == 'fpn': fpn_params = params.fpn fpn_fn = fpn.Fpn( min_level=params.architecture.min_level, max_level=params.architecture.max_level, fpn_feat_dims=fpn_params.fpn_feat_dims, use_separable_conv=fpn_params.use_separable_conv, activation=params.norm_activation.activation, use_batch_norm=fpn_params.use_batch_norm, norm_activation=norm_activation_generator( params.norm_activation)) elif params.architecture.multilevel_features == 'identity': fpn_fn = identity.Identity() else: raise ValueError('The multi-level feature model `{}` is not supported.' .format(params.architecture.multilevel_features)) return fpn_fn def retinanet_head_generator(params): """Generator function for RetinaNet head architecture.""" head_params = params.retinanet_head anchors_per_location = params.anchor.num_scales * len( params.anchor.aspect_ratios) return heads.RetinanetHead( params.architecture.min_level, params.architecture.max_level, params.architecture.num_classes, anchors_per_location, head_params.num_convs, head_params.num_filters, head_params.use_separable_conv, norm_activation=norm_activation_generator(params.norm_activation)) def rpn_head_generator(params): """Generator function for RPN head architecture.""" head_params = params.rpn_head anchors_per_location = params.anchor.num_scales * len( params.anchor.aspect_ratios) return heads.RpnHead( params.architecture.min_level, params.architecture.max_level, anchors_per_location, head_params.num_convs, head_params.num_filters, head_params.use_separable_conv, params.norm_activation.activation, head_params.use_batch_norm, norm_activation=norm_activation_generator(params.norm_activation)) def oln_rpn_head_generator(params): """Generator function for OLN-proposal (OLN-RPN) head architecture.""" head_params = params.rpn_head anchors_per_location = params.anchor.num_scales * len( params.anchor.aspect_ratios) return heads.OlnRpnHead( params.architecture.min_level, params.architecture.max_level, anchors_per_location, head_params.num_convs, head_params.num_filters, head_params.use_separable_conv, params.norm_activation.activation, head_params.use_batch_norm, norm_activation=norm_activation_generator(params.norm_activation)) def fast_rcnn_head_generator(params): """Generator function for Fast R-CNN head architecture.""" head_params = params.frcnn_head return heads.FastrcnnHead( params.architecture.num_classes, head_params.num_convs, head_params.num_filters, head_params.use_separable_conv, head_params.num_fcs, head_params.fc_dims, params.norm_activation.activation, head_params.use_batch_norm, norm_activation=norm_activation_generator(params.norm_activation)) def oln_box_score_head_generator(params): """Generator function for Scoring Fast R-CNN head architecture.""" head_params = params.frcnn_head return heads.OlnBoxScoreHead( params.architecture.num_classes, head_params.num_convs, head_params.num_filters, head_params.use_separable_conv, head_params.num_fcs, head_params.fc_dims, params.norm_activation.activation, head_params.use_batch_norm, norm_activation=norm_activation_generator(params.norm_activation)) def mask_rcnn_head_generator(params): """Generator function for Mask R-CNN head architecture.""" head_params = params.mrcnn_head return heads.MaskrcnnHead( params.architecture.num_classes, params.architecture.mask_target_size, head_params.num_convs, head_params.num_filters, head_params.use_separable_conv, params.norm_activation.activation, head_params.use_batch_norm, norm_activation=norm_activation_generator(params.norm_activation)) def oln_mask_score_head_generator(params): """Generator function for Scoring Mask R-CNN head architecture.""" head_params = params.mrcnn_head return heads.OlnMaskScoreHead( params.architecture.num_classes, params.architecture.mask_target_size, head_params.num_convs, head_params.num_filters, head_params.use_separable_conv, params.norm_activation.activation, head_params.use_batch_norm, norm_activation=norm_activation_generator(params.norm_activation)) def shapeprior_head_generator(params): """Generator function for shape prior head architecture.""" head_params = params.shapemask_head return heads.ShapemaskPriorHead( params.architecture.num_classes, head_params.num_downsample_channels, head_params.mask_crop_size, head_params.use_category_for_mask, head_params.shape_prior_path) def coarsemask_head_generator(params): """Generator function for ShapeMask coarse mask head architecture.""" head_params = params.shapemask_head return heads.ShapemaskCoarsemaskHead( params.architecture.num_classes, head_params.num_downsample_channels, head_params.mask_crop_size, head_params.use_category_for_mask, head_params.num_convs, norm_activation=norm_activation_generator(params.norm_activation)) def finemask_head_generator(params): """Generator function for Shapemask fine mask head architecture.""" head_params = params.shapemask_head return heads.ShapemaskFinemaskHead( params.architecture.num_classes, head_params.num_downsample_channels, head_params.mask_crop_size, head_params.use_category_for_mask, head_params.num_convs, head_params.upsample_factor)
7,994
35.674312
77
py
models
models-master/official/legacy/detection/modeling/architecture/fpn.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Feature Pyramid Networks. Feature Pyramid Networks were proposed in: [1] Tsung-Yi Lin, Piotr Dollar, Ross Girshick, Kaiming He, Bharath Hariharan, , and Serge Belongie Feature Pyramid Networks for Object Detection. CVPR 2017. """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import functools import tensorflow as tf from official.legacy.detection.modeling.architecture import nn_ops from official.legacy.detection.ops import spatial_transform_ops class Fpn(object): """Feature pyramid networks.""" def __init__(self, min_level=3, max_level=7, fpn_feat_dims=256, use_separable_conv=False, activation='relu', use_batch_norm=True, norm_activation=nn_ops.norm_activation_builder( activation='relu')): """FPN initialization function. Args: min_level: `int` minimum level in FPN output feature maps. max_level: `int` maximum level in FPN output feature maps. fpn_feat_dims: `int` number of filters in FPN layers. use_separable_conv: `bool`, if True use separable convolution for convolution in FPN layers. activation: the activation function. use_batch_norm: 'bool', indicating whether batchnorm layers are added. norm_activation: an operation that includes a normalization layer followed by an optional activation layer. """ self._min_level = min_level self._max_level = max_level self._fpn_feat_dims = fpn_feat_dims if use_separable_conv: self._conv2d_op = functools.partial( tf.keras.layers.SeparableConv2D, depth_multiplier=1) else: self._conv2d_op = tf.keras.layers.Conv2D if activation == 'relu': self._activation_op = tf.nn.relu elif activation == 'swish': self._activation_op = tf.nn.swish else: raise ValueError('Unsupported activation `{}`.'.format(activation)) self._use_batch_norm = use_batch_norm self._norm_activation = norm_activation self._norm_activations = {} self._lateral_conv2d_op = {} self._post_hoc_conv2d_op = {} self._coarse_conv2d_op = {} for level in range(self._min_level, self._max_level + 1): if self._use_batch_norm: self._norm_activations[level] = norm_activation( use_activation=False, name='p%d-bn' % level) self._lateral_conv2d_op[level] = self._conv2d_op( filters=self._fpn_feat_dims, kernel_size=(1, 1), padding='same', name='l%d' % level) self._post_hoc_conv2d_op[level] = self._conv2d_op( filters=self._fpn_feat_dims, strides=(1, 1), kernel_size=(3, 3), padding='same', name='post_hoc_d%d' % level) self._coarse_conv2d_op[level] = self._conv2d_op( filters=self._fpn_feat_dims, strides=(2, 2), kernel_size=(3, 3), padding='same', name='p%d' % level) def __call__(self, multilevel_features, is_training=None): """Returns the FPN features for a given multilevel features. Args: multilevel_features: a `dict` containing `int` keys for continuous feature levels, e.g., [2, 3, 4, 5]. The values are corresponding features with shape [batch_size, height_l, width_l, num_filters]. is_training: `bool` if True, the model is in training mode. Returns: a `dict` containing `int` keys for continuous feature levels [min_level, min_level + 1, ..., max_level]. The values are corresponding FPN features with shape [batch_size, height_l, width_l, fpn_feat_dims]. """ input_levels = list(multilevel_features.keys()) if min(input_levels) > self._min_level: raise ValueError( 'The minimum backbone level %d should be '%(min(input_levels)) + 'less or equal to FPN minimum level %d.:'%(self._min_level)) backbone_max_level = min(max(input_levels), self._max_level) with tf.name_scope('fpn'): # Adds lateral connections. feats_lateral = {} for level in range(self._min_level, backbone_max_level + 1): feats_lateral[level] = self._lateral_conv2d_op[level]( multilevel_features[level]) # Adds top-down path. feats = {backbone_max_level: feats_lateral[backbone_max_level]} for level in range(backbone_max_level - 1, self._min_level - 1, -1): feats[level] = spatial_transform_ops.nearest_upsampling( feats[level + 1], 2) + feats_lateral[level] # Adds post-hoc 3x3 convolution kernel. for level in range(self._min_level, backbone_max_level + 1): feats[level] = self._post_hoc_conv2d_op[level](feats[level]) # Adds coarser FPN levels introduced for RetinaNet. for level in range(backbone_max_level + 1, self._max_level + 1): feats_in = feats[level - 1] if level > backbone_max_level + 1: feats_in = self._activation_op(feats_in) feats[level] = self._coarse_conv2d_op[level](feats_in) if self._use_batch_norm: # Adds batch_norm layer. for level in range(self._min_level, self._max_level + 1): feats[level] = self._norm_activations[level]( feats[level], is_training=is_training) return feats
5,976
38.322368
80
py
models
models-master/official/legacy/detection/modeling/architecture/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/detection/modeling/architecture/nn_blocks.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Contains common building blocks for neural networks.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from official.modeling import tf_utils class ResidualBlock(tf.keras.layers.Layer): """A residual block.""" def __init__(self, filters, strides, use_projection=False, kernel_initializer='VarianceScaling', kernel_regularizer=None, bias_regularizer=None, activation='relu', use_sync_bn=False, norm_momentum=0.99, norm_epsilon=0.001, **kwargs): """A residual block with BN after convolutions. Args: filters: `int` number of filters for the first two convolutions. Note that the third and final convolution will use 4 times as many filters. strides: `int` block stride. If greater than 1, this block will ultimately downsample the input. use_projection: `bool` for whether this block should use a projection shortcut (versus the default identity shortcut). This is usually `True` for the first block of a block group, which may change the number of filters and the resolution. kernel_initializer: kernel_initializer for convolutional layers. kernel_regularizer: tf.keras.regularizers.Regularizer object for Conv2D. Default to None. bias_regularizer: tf.keras.regularizers.Regularizer object for Conv2d. Default to None. activation: `str` name of the activation function. use_sync_bn: if True, use synchronized batch normalization. norm_momentum: `float` normalization omentum for the moving average. norm_epsilon: `float` small float added to variance to avoid dividing by zero. **kwargs: keyword arguments to be passed. """ super(ResidualBlock, self).__init__(**kwargs) self._filters = filters self._strides = strides self._use_projection = use_projection self._use_sync_bn = use_sync_bn self._activation = activation self._kernel_initializer = kernel_initializer self._norm_momentum = norm_momentum self._norm_epsilon = norm_epsilon self._kernel_regularizer = kernel_regularizer self._bias_regularizer = bias_regularizer if use_sync_bn: self._norm = tf.keras.layers.experimental.SyncBatchNormalization else: self._norm = tf.keras.layers.BatchNormalization if tf.keras.backend.image_data_format() == 'channels_last': self._bn_axis = -1 else: self._bn_axis = 1 self._activation_fn = tf_utils.get_activation(activation) def build(self, input_shape): if self._use_projection: self._shortcut = tf.keras.layers.Conv2D( filters=self._filters, kernel_size=1, strides=self._strides, use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer) self._norm0 = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon) self._conv1 = tf.keras.layers.Conv2D( filters=self._filters, kernel_size=3, strides=self._strides, padding='same', use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer) self._norm1 = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon) self._conv2 = tf.keras.layers.Conv2D( filters=self._filters, kernel_size=3, strides=1, padding='same', use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer) self._norm2 = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon) super(ResidualBlock, self).build(input_shape) def get_config(self): config = { 'filters': self._filters, 'strides': self._strides, 'use_projection': self._use_projection, 'kernel_initializer': self._kernel_initializer, 'kernel_regularizer': self._kernel_regularizer, 'bias_regularizer': self._bias_regularizer, 'activation': self._activation, 'use_sync_bn': self._use_sync_bn, 'norm_momentum': self._norm_momentum, 'norm_epsilon': self._norm_epsilon } base_config = super(ResidualBlock, self).get_config() return dict(list(base_config.items()) + list(config.items())) def call(self, inputs): shortcut = inputs if self._use_projection: shortcut = self._shortcut(shortcut) shortcut = self._norm0(shortcut) x = self._conv1(inputs) x = self._norm1(x) x = self._activation_fn(x) x = self._conv2(x) x = self._norm2(x) return self._activation_fn(x + shortcut) class BottleneckBlock(tf.keras.layers.Layer): """A standard bottleneck block.""" def __init__(self, filters, strides, use_projection=False, kernel_initializer='VarianceScaling', kernel_regularizer=None, bias_regularizer=None, activation='relu', use_sync_bn=False, norm_momentum=0.99, norm_epsilon=0.001, **kwargs): """A standard bottleneck block with BN after convolutions. Args: filters: `int` number of filters for the first two convolutions. Note that the third and final convolution will use 4 times as many filters. strides: `int` block stride. If greater than 1, this block will ultimately downsample the input. use_projection: `bool` for whether this block should use a projection shortcut (versus the default identity shortcut). This is usually `True` for the first block of a block group, which may change the number of filters and the resolution. kernel_initializer: kernel_initializer for convolutional layers. kernel_regularizer: tf.keras.regularizers.Regularizer object for Conv2D. Default to None. bias_regularizer: tf.keras.regularizers.Regularizer object for Conv2d. Default to None. activation: `str` name of the activation function. use_sync_bn: if True, use synchronized batch normalization. norm_momentum: `float` normalization omentum for the moving average. norm_epsilon: `float` small float added to variance to avoid dividing by zero. **kwargs: keyword arguments to be passed. """ super(BottleneckBlock, self).__init__(**kwargs) self._filters = filters self._strides = strides self._use_projection = use_projection self._use_sync_bn = use_sync_bn self._activation = activation self._kernel_initializer = kernel_initializer self._norm_momentum = norm_momentum self._norm_epsilon = norm_epsilon self._kernel_regularizer = kernel_regularizer self._bias_regularizer = bias_regularizer if use_sync_bn: self._norm = tf.keras.layers.experimental.SyncBatchNormalization else: self._norm = tf.keras.layers.BatchNormalization if tf.keras.backend.image_data_format() == 'channels_last': self._bn_axis = -1 else: self._bn_axis = 1 self._activation_fn = tf_utils.get_activation(activation) def build(self, input_shape): if self._use_projection: self._shortcut = tf.keras.layers.Conv2D( filters=self._filters * 4, kernel_size=1, strides=self._strides, use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer) self._norm0 = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon) self._conv1 = tf.keras.layers.Conv2D( filters=self._filters, kernel_size=1, strides=1, use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer) self._norm1 = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon) self._conv2 = tf.keras.layers.Conv2D( filters=self._filters, kernel_size=3, strides=self._strides, padding='same', use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer) self._norm2 = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon) self._conv3 = tf.keras.layers.Conv2D( filters=self._filters * 4, kernel_size=1, strides=1, use_bias=False, kernel_initializer=self._kernel_initializer, kernel_regularizer=self._kernel_regularizer, bias_regularizer=self._bias_regularizer) self._norm3 = self._norm( axis=self._bn_axis, momentum=self._norm_momentum, epsilon=self._norm_epsilon) super(BottleneckBlock, self).build(input_shape) def get_config(self): config = { 'filters': self._filters, 'strides': self._strides, 'use_projection': self._use_projection, 'kernel_initializer': self._kernel_initializer, 'kernel_regularizer': self._kernel_regularizer, 'bias_regularizer': self._bias_regularizer, 'activation': self._activation, 'use_sync_bn': self._use_sync_bn, 'norm_momentum': self._norm_momentum, 'norm_epsilon': self._norm_epsilon } base_config = super(BottleneckBlock, self).get_config() return dict(list(base_config.items()) + list(config.items())) def call(self, inputs): shortcut = inputs if self._use_projection: shortcut = self._shortcut(shortcut) shortcut = self._norm0(shortcut) x = self._conv1(inputs) x = self._norm1(x) x = self._activation_fn(x) x = self._conv2(x) x = self._norm2(x) x = self._activation_fn(x) x = self._conv3(x) x = self._norm3(x) return self._activation_fn(x + shortcut)
11,219
34.394322
80
py
models
models-master/official/legacy/detection/ops/nms.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tensorflow implementation of non max suppression.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from official.legacy.detection.utils import box_utils NMS_TILE_SIZE = 512 def _self_suppression(iou, _, iou_sum): batch_size = tf.shape(iou)[0] can_suppress_others = tf.cast( tf.reshape(tf.reduce_max(iou, 1) <= 0.5, [batch_size, -1, 1]), iou.dtype) iou_suppressed = tf.reshape( tf.cast(tf.reduce_max(can_suppress_others * iou, 1) <= 0.5, iou.dtype), [batch_size, -1, 1]) * iou iou_sum_new = tf.reduce_sum(iou_suppressed, [1, 2]) return [ iou_suppressed, tf.reduce_any(iou_sum - iou_sum_new > 0.5), iou_sum_new ] def _cross_suppression(boxes, box_slice, iou_threshold, inner_idx): batch_size = tf.shape(boxes)[0] new_slice = tf.slice(boxes, [0, inner_idx * NMS_TILE_SIZE, 0], [batch_size, NMS_TILE_SIZE, 4]) iou = box_utils.bbox_overlap(new_slice, box_slice) ret_slice = tf.expand_dims( tf.cast(tf.reduce_all(iou < iou_threshold, [1]), box_slice.dtype), 2) * box_slice return boxes, ret_slice, iou_threshold, inner_idx + 1 def _suppression_loop_body(boxes, iou_threshold, output_size, idx): """Process boxes in the range [idx*NMS_TILE_SIZE, (idx+1)*NMS_TILE_SIZE). Args: boxes: a tensor with a shape of [batch_size, anchors, 4]. iou_threshold: a float representing the threshold for deciding whether boxes overlap too much with respect to IOU. output_size: an int32 tensor of size [batch_size]. Representing the number of selected boxes for each batch. idx: an integer scalar representing induction variable. Returns: boxes: updated boxes. iou_threshold: pass down iou_threshold to the next iteration. output_size: the updated output_size. idx: the updated induction variable. """ boxes_shape = tf.shape(boxes) num_tiles = boxes_shape[1] // NMS_TILE_SIZE batch_size = boxes_shape[0] # Iterates over tiles that can possibly suppress the current tile. box_slice = tf.slice(boxes, [0, idx * NMS_TILE_SIZE, 0], [batch_size, NMS_TILE_SIZE, 4]) _, box_slice, _, _ = tf.while_loop( lambda _boxes, _box_slice, _threshold, inner_idx: inner_idx < idx, _cross_suppression, [boxes, box_slice, iou_threshold, tf.constant(0)]) # Iterates over the current tile to compute self-suppression. iou = box_utils.bbox_overlap(box_slice, box_slice) mask = tf.expand_dims( tf.reshape(tf.range(NMS_TILE_SIZE), [1, -1]) > tf.reshape( tf.range(NMS_TILE_SIZE), [-1, 1]), 0) iou *= tf.cast(tf.logical_and(mask, iou >= iou_threshold), iou.dtype) suppressed_iou, _, _ = tf.while_loop( lambda _iou, loop_condition, _iou_sum: loop_condition, _self_suppression, [iou, tf.constant(True), tf.reduce_sum(iou, [1, 2])]) suppressed_box = tf.reduce_sum(suppressed_iou, 1) > 0 box_slice *= tf.expand_dims(1.0 - tf.cast(suppressed_box, box_slice.dtype), 2) # Uses box_slice to update the input boxes. mask = tf.reshape( tf.cast(tf.equal(tf.range(num_tiles), idx), boxes.dtype), [1, -1, 1, 1]) boxes = tf.tile(tf.expand_dims( box_slice, [1]), [1, num_tiles, 1, 1]) * mask + tf.reshape( boxes, [batch_size, num_tiles, NMS_TILE_SIZE, 4]) * (1 - mask) boxes = tf.reshape(boxes, boxes_shape) # Updates output_size. output_size += tf.reduce_sum( tf.cast(tf.reduce_any(box_slice > 0, [2]), tf.int32), [1]) return boxes, iou_threshold, output_size, idx + 1 def sorted_non_max_suppression_padded(scores, boxes, max_output_size, iou_threshold): """A wrapper that handles non-maximum suppression. Assumption: * The boxes are sorted by scores unless the box is a dot (all coordinates are zero). * Boxes with higher scores can be used to suppress boxes with lower scores. The overal design of the algorithm is to handle boxes tile-by-tile: boxes = boxes.pad_to_multiply_of(tile_size) num_tiles = len(boxes) // tile_size output_boxes = [] for i in range(num_tiles): box_tile = boxes[i*tile_size : (i+1)*tile_size] for j in range(i - 1): suppressing_tile = boxes[j*tile_size : (j+1)*tile_size] iou = bbox_overlap(box_tile, suppressing_tile) # if the box is suppressed in iou, clear it to a dot box_tile *= _update_boxes(iou) # Iteratively handle the diagnal tile. iou = _box_overlap(box_tile, box_tile) iou_changed = True while iou_changed: # boxes that are not suppressed by anything else suppressing_boxes = _get_suppressing_boxes(iou) # boxes that are suppressed by suppressing_boxes suppressed_boxes = _get_suppressed_boxes(iou, suppressing_boxes) # clear iou to 0 for boxes that are suppressed, as they cannot be used # to suppress other boxes any more new_iou = _clear_iou(iou, suppressed_boxes) iou_changed = (new_iou != iou) iou = new_iou # remaining boxes that can still suppress others, are selected boxes. output_boxes.append(_get_suppressing_boxes(iou)) if len(output_boxes) >= max_output_size: break Args: scores: a tensor with a shape of [batch_size, anchors]. boxes: a tensor with a shape of [batch_size, anchors, 4]. max_output_size: a scalar integer `Tensor` representing the maximum number of boxes to be selected by non max suppression. iou_threshold: a float representing the threshold for deciding whether boxes overlap too much with respect to IOU. Returns: nms_scores: a tensor with a shape of [batch_size, anchors]. It has same dtype as input scores. nms_proposals: a tensor with a shape of [batch_size, anchors, 4]. It has same dtype as input boxes. """ batch_size = tf.shape(boxes)[0] num_boxes = tf.shape(boxes)[1] pad = tf.cast( tf.math.ceil(tf.cast(num_boxes, tf.float32) / NMS_TILE_SIZE), tf.int32) * NMS_TILE_SIZE - num_boxes boxes = tf.pad(tf.cast(boxes, tf.float32), [[0, 0], [0, pad], [0, 0]]) scores = tf.pad( tf.cast(scores, tf.float32), [[0, 0], [0, pad]], constant_values=-1) num_boxes += pad def _loop_cond(unused_boxes, unused_threshold, output_size, idx): return tf.logical_and( tf.reduce_min(output_size) < max_output_size, idx < num_boxes // NMS_TILE_SIZE) selected_boxes, _, output_size, _ = tf.while_loop( _loop_cond, _suppression_loop_body, [boxes, iou_threshold, tf.zeros([batch_size], tf.int32), tf.constant(0)]) idx = num_boxes - tf.cast( tf.nn.top_k( tf.cast(tf.reduce_any(selected_boxes > 0, [2]), tf.int32) * tf.expand_dims(tf.range(num_boxes, 0, -1), 0), max_output_size)[0], tf.int32) idx = tf.minimum(idx, num_boxes - 1) idx = tf.reshape(idx + tf.reshape(tf.range(batch_size) * num_boxes, [-1, 1]), [-1]) boxes = tf.reshape( tf.gather(tf.reshape(boxes, [-1, 4]), idx), [batch_size, max_output_size, 4]) boxes = boxes * tf.cast( tf.reshape(tf.range(max_output_size), [1, -1, 1]) < tf.reshape( output_size, [-1, 1, 1]), boxes.dtype) scores = tf.reshape( tf.gather(tf.reshape(scores, [-1, 1]), idx), [batch_size, max_output_size]) scores = scores * tf.cast( tf.reshape(tf.range(max_output_size), [1, -1]) < tf.reshape( output_size, [-1, 1]), scores.dtype) return scores, boxes
8,141
39.108374
80
py
models
models-master/official/legacy/detection/ops/roi_ops.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ROI-related ops.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from official.legacy.detection.ops import nms from official.legacy.detection.utils import box_utils def multilevel_propose_rois(rpn_boxes, rpn_scores, anchor_boxes, image_shape, rpn_pre_nms_top_k=2000, rpn_post_nms_top_k=1000, rpn_nms_threshold=0.7, rpn_score_threshold=0.0, rpn_min_size_threshold=0.0, decode_boxes=True, clip_boxes=True, use_batched_nms=False, apply_sigmoid_to_score=True): """Proposes RoIs given a group of candidates from different FPN levels. The following describes the steps: 1. For each individual level: a. Apply sigmoid transform if specified. b. Decode boxes if specified. c. Clip boxes if specified. d. Filter small boxes and those fall outside image if specified. e. Apply pre-NMS filtering including pre-NMS top k and score thresholding. f. Apply NMS. 2. Aggregate post-NMS boxes from each level. 3. Apply an overall top k to generate the final selected RoIs. Args: rpn_boxes: a dict with keys representing FPN levels and values representing box tenors of shape [batch_size, feature_h, feature_w, num_anchors * 4]. rpn_scores: a dict with keys representing FPN levels and values representing logit tensors of shape [batch_size, feature_h, feature_w, num_anchors]. anchor_boxes: a dict with keys representing FPN levels and values representing anchor box tensors of shape [batch_size, feature_h, feature_w, num_anchors * 4]. image_shape: a tensor of shape [batch_size, 2] where the last dimension are [height, width] of the scaled image. rpn_pre_nms_top_k: an integer of top scoring RPN proposals *per level* to keep before applying NMS. Default: 2000. rpn_post_nms_top_k: an integer of top scoring RPN proposals *in total* to keep after applying NMS. Default: 1000. rpn_nms_threshold: a float between 0 and 1 representing the IoU threshold used for NMS. If 0.0, no NMS is applied. Default: 0.7. rpn_score_threshold: a float between 0 and 1 representing the minimal box score to keep before applying NMS. This is often used as a pre-filtering step for better performance. If 0, no filtering is applied. Default: 0. rpn_min_size_threshold: a float representing the minimal box size in each side (w.r.t. the scaled image) to keep before applying NMS. This is often used as a pre-filtering step for better performance. If 0, no filtering is applied. Default: 0. decode_boxes: a boolean indicating whether `rpn_boxes` needs to be decoded using `anchor_boxes`. If False, use `rpn_boxes` directly and ignore `anchor_boxes`. Default: True. clip_boxes: a boolean indicating whether boxes are first clipped to the scaled image size before appliying NMS. If False, no clipping is applied and `image_shape` is ignored. Default: True. use_batched_nms: a boolean indicating whether NMS is applied in batch using `tf.image.combined_non_max_suppression`. Currently only available in CPU/GPU. Default: False. apply_sigmoid_to_score: a boolean indicating whether apply sigmoid to `rpn_scores` before applying NMS. Default: True. Returns: selected_rois: a tensor of shape [batch_size, rpn_post_nms_top_k, 4], representing the box coordinates of the selected proposals w.r.t. the scaled image. selected_roi_scores: a tensor of shape [batch_size, rpn_post_nms_top_k, 1], representing the scores of the selected proposals. """ with tf.name_scope('multilevel_propose_rois'): rois = [] roi_scores = [] image_shape = tf.expand_dims(image_shape, axis=1) for level in sorted(rpn_scores.keys()): with tf.name_scope('level_%d' % level): _, feature_h, feature_w, num_anchors_per_location = ( rpn_scores[level].get_shape().as_list()) num_boxes = feature_h * feature_w * num_anchors_per_location this_level_scores = tf.reshape(rpn_scores[level], [-1, num_boxes]) this_level_boxes = tf.reshape(rpn_boxes[level], [-1, num_boxes, 4]) this_level_anchors = tf.cast( tf.reshape(anchor_boxes[level], [-1, num_boxes, 4]), dtype=this_level_scores.dtype) if apply_sigmoid_to_score: this_level_scores = tf.sigmoid(this_level_scores) if decode_boxes: this_level_boxes = box_utils.decode_boxes(this_level_boxes, this_level_anchors) if clip_boxes: this_level_boxes = box_utils.clip_boxes(this_level_boxes, image_shape) if rpn_min_size_threshold > 0.0: this_level_boxes, this_level_scores = box_utils.filter_boxes( this_level_boxes, this_level_scores, image_shape, rpn_min_size_threshold) this_level_pre_nms_top_k = min(num_boxes, rpn_pre_nms_top_k) this_level_post_nms_top_k = min(num_boxes, rpn_post_nms_top_k) if rpn_nms_threshold > 0.0: if use_batched_nms: this_level_rois, this_level_roi_scores, _, _ = ( tf.image.combined_non_max_suppression( tf.expand_dims(this_level_boxes, axis=2), tf.expand_dims(this_level_scores, axis=-1), max_output_size_per_class=this_level_pre_nms_top_k, max_total_size=this_level_post_nms_top_k, iou_threshold=rpn_nms_threshold, score_threshold=rpn_score_threshold, pad_per_class=False, clip_boxes=False)) else: if rpn_score_threshold > 0.0: this_level_boxes, this_level_scores = ( box_utils.filter_boxes_by_scores(this_level_boxes, this_level_scores, rpn_score_threshold)) this_level_boxes, this_level_scores = box_utils.top_k_boxes( this_level_boxes, this_level_scores, k=this_level_pre_nms_top_k) this_level_roi_scores, this_level_rois = ( nms.sorted_non_max_suppression_padded( this_level_scores, this_level_boxes, max_output_size=this_level_post_nms_top_k, iou_threshold=rpn_nms_threshold)) else: this_level_rois, this_level_roi_scores = box_utils.top_k_boxes( this_level_rois, this_level_scores, k=this_level_post_nms_top_k) rois.append(this_level_rois) roi_scores.append(this_level_roi_scores) all_rois = tf.concat(rois, axis=1) all_roi_scores = tf.concat(roi_scores, axis=1) with tf.name_scope('top_k_rois'): _, num_valid_rois = all_roi_scores.get_shape().as_list() overall_top_k = min(num_valid_rois, rpn_post_nms_top_k) selected_rois, selected_roi_scores = box_utils.top_k_boxes( all_rois, all_roi_scores, k=overall_top_k) return selected_rois, selected_roi_scores class ROIGenerator(tf.keras.layers.Layer): """Proposes RoIs for the second stage processing.""" def __init__(self, params): self._rpn_pre_nms_top_k = params.rpn_pre_nms_top_k self._rpn_post_nms_top_k = params.rpn_post_nms_top_k self._rpn_nms_threshold = params.rpn_nms_threshold self._rpn_score_threshold = params.rpn_score_threshold self._rpn_min_size_threshold = params.rpn_min_size_threshold self._test_rpn_pre_nms_top_k = params.test_rpn_pre_nms_top_k self._test_rpn_post_nms_top_k = params.test_rpn_post_nms_top_k self._test_rpn_nms_threshold = params.test_rpn_nms_threshold self._test_rpn_score_threshold = params.test_rpn_score_threshold self._test_rpn_min_size_threshold = params.test_rpn_min_size_threshold self._use_batched_nms = params.use_batched_nms super(ROIGenerator, self).__init__(autocast=False) def call(self, boxes, scores, anchor_boxes, image_shape, is_training): """Generates RoI proposals. Args: boxes: a dict with keys representing FPN levels and values representing box tenors of shape [batch_size, feature_h, feature_w, num_anchors * 4]. scores: a dict with keys representing FPN levels and values representing logit tensors of shape [batch_size, feature_h, feature_w, num_anchors]. anchor_boxes: a dict with keys representing FPN levels and values representing anchor box tensors of shape [batch_size, feature_h, feature_w, num_anchors * 4]. image_shape: a tensor of shape [batch_size, 2] where the last dimension are [height, width] of the scaled image. is_training: a bool indicating whether it is in training or inference mode. Returns: proposed_rois: a tensor of shape [batch_size, rpn_post_nms_top_k, 4], representing the box coordinates of the proposed RoIs w.r.t. the scaled image. proposed_roi_scores: a tensor of shape [batch_size, rpn_post_nms_top_k, 1], representing the scores of the proposed RoIs. """ proposed_rois, proposed_roi_scores = multilevel_propose_rois( boxes, scores, anchor_boxes, image_shape, rpn_pre_nms_top_k=(self._rpn_pre_nms_top_k if is_training else self._test_rpn_pre_nms_top_k), rpn_post_nms_top_k=(self._rpn_post_nms_top_k if is_training else self._test_rpn_post_nms_top_k), rpn_nms_threshold=(self._rpn_nms_threshold if is_training else self._test_rpn_nms_threshold), rpn_score_threshold=(self._rpn_score_threshold if is_training else self._test_rpn_score_threshold), rpn_min_size_threshold=(self._rpn_min_size_threshold if is_training else self._test_rpn_min_size_threshold), decode_boxes=True, clip_boxes=True, use_batched_nms=self._use_batched_nms, apply_sigmoid_to_score=True) return proposed_rois, proposed_roi_scores class OlnROIGenerator(ROIGenerator): """Proposes RoIs for the second stage processing.""" def __call__(self, boxes, scores, anchor_boxes, image_shape, is_training, is_box_lrtb=False, object_scores=None): """Generates RoI proposals. Args: boxes: a dict with keys representing FPN levels and values representing box tenors of shape [batch_size, feature_h, feature_w, num_anchors * 4]. scores: a dict with keys representing FPN levels and values representing logit tensors of shape [batch_size, feature_h, feature_w, num_anchors]. anchor_boxes: a dict with keys representing FPN levels and values representing anchor box tensors of shape [batch_size, feature_h, feature_w, num_anchors * 4]. image_shape: a tensor of shape [batch_size, 2] where the last dimension are [height, width] of the scaled image. is_training: a bool indicating whether it is in training or inference mode. is_box_lrtb: a bool indicating whether boxes are in lrtb (=left,right,top, bottom) format. object_scores: another objectness score (e.g., centerness). In OLN, we use object_scores=centerness as a replacement of the scores at each level. A dict with keys representing FPN levels and values representing logit tensors of shape [batch_size, feature_h, feature_w, num_anchors]. Returns: proposed_rois: a tensor of shape [batch_size, rpn_post_nms_top_k, 4], representing the box coordinates of the proposed RoIs w.r.t. the scaled image. proposed_roi_scores: a tensor of shape [batch_size, rpn_post_nms_top_k, 1], representing the scores of the proposed RoIs. """ proposed_rois, proposed_roi_scores = self.oln_multilevel_propose_rois( boxes, scores, anchor_boxes, image_shape, rpn_pre_nms_top_k=(self._rpn_pre_nms_top_k if is_training else self._test_rpn_pre_nms_top_k), rpn_post_nms_top_k=(self._rpn_post_nms_top_k if is_training else self._test_rpn_post_nms_top_k), rpn_nms_threshold=(self._rpn_nms_threshold if is_training else self._test_rpn_nms_threshold), rpn_score_threshold=(self._rpn_score_threshold if is_training else self._test_rpn_score_threshold), rpn_min_size_threshold=(self._rpn_min_size_threshold if is_training else self._test_rpn_min_size_threshold), decode_boxes=True, clip_boxes=True, use_batched_nms=self._use_batched_nms, apply_sigmoid_to_score=True, is_box_lrtb=is_box_lrtb, rpn_object_scores=object_scores,) return proposed_rois, proposed_roi_scores def oln_multilevel_propose_rois(self, rpn_boxes, rpn_scores, anchor_boxes, image_shape, rpn_pre_nms_top_k=2000, rpn_post_nms_top_k=1000, rpn_nms_threshold=0.7, rpn_score_threshold=0.0, rpn_min_size_threshold=0.0, decode_boxes=True, clip_boxes=True, use_batched_nms=False, apply_sigmoid_to_score=True, is_box_lrtb=False, rpn_object_scores=None,): """Proposes RoIs given a group of candidates from different FPN levels. The following describes the steps: 1. For each individual level: a. Adjust scores for each level if specified by rpn_object_scores. b. Apply sigmoid transform if specified. c. Decode boxes (either of xyhw or left-right-top-bottom format) if specified. d. Clip boxes if specified. e. Filter small boxes and those fall outside image if specified. f. Apply pre-NMS filtering including pre-NMS top k and score thresholding. g. Apply NMS. 2. Aggregate post-NMS boxes from each level. 3. Apply an overall top k to generate the final selected RoIs. Args: rpn_boxes: a dict with keys representing FPN levels and values representing box tenors of shape [batch_size, feature_h, feature_w, num_anchors * 4]. rpn_scores: a dict with keys representing FPN levels and values representing logit tensors of shape [batch_size, feature_h, feature_w, num_anchors]. anchor_boxes: a dict with keys representing FPN levels and values representing anchor box tensors of shape [batch_size, feature_h, feature_w, num_anchors * 4]. image_shape: a tensor of shape [batch_size, 2] where the last dimension are [height, width] of the scaled image. rpn_pre_nms_top_k: an integer of top scoring RPN proposals *per level* to keep before applying NMS. Default: 2000. rpn_post_nms_top_k: an integer of top scoring RPN proposals *in total* to keep after applying NMS. Default: 1000. rpn_nms_threshold: a float between 0 and 1 representing the IoU threshold used for NMS. If 0.0, no NMS is applied. Default: 0.7. rpn_score_threshold: a float between 0 and 1 representing the minimal box score to keep before applying NMS. This is often used as a pre-filtering step for better performance. If 0, no filtering is applied. Default: 0. rpn_min_size_threshold: a float representing the minimal box size in each side (w.r.t. the scaled image) to keep before applying NMS. This is often used as a pre-filtering step for better performance. If 0, no filtering is applied. Default: 0. decode_boxes: a boolean indicating whether `rpn_boxes` needs to be decoded using `anchor_boxes`. If False, use `rpn_boxes` directly and ignore `anchor_boxes`. Default: True. clip_boxes: a boolean indicating whether boxes are first clipped to the scaled image size before appliying NMS. If False, no clipping is applied and `image_shape` is ignored. Default: True. use_batched_nms: a boolean indicating whether NMS is applied in batch using `tf.image.combined_non_max_suppression`. Currently only available in CPU/GPU. Default: False. apply_sigmoid_to_score: a boolean indicating whether apply sigmoid to `rpn_scores` before applying NMS. Default: True. is_box_lrtb: a bool indicating whether boxes are in lrtb (=left,right,top, bottom) format. rpn_object_scores: a predicted objectness score (e.g., centerness). In OLN, we use object_scores=centerness as a replacement of the scores at each level. A dict with keys representing FPN levels and values representing logit tensors of shape [batch_size, feature_h, feature_w, num_anchors]. Returns: selected_rois: a tensor of shape [batch_size, rpn_post_nms_top_k, 4], representing the box coordinates of the selected proposals w.r.t. the scaled image. selected_roi_scores: a tensor of shape [batch_size, rpn_post_nms_top_k, 1],representing the scores of the selected proposals. """ with tf.name_scope('multilevel_propose_rois'): rois = [] roi_scores = [] image_shape = tf.expand_dims(image_shape, axis=1) for level in sorted(rpn_scores.keys()): with tf.name_scope('level_%d' % level): _, feature_h, feature_w, num_anchors_per_location = ( rpn_scores[level].get_shape().as_list()) num_boxes = feature_h * feature_w * num_anchors_per_location this_level_scores = tf.reshape(rpn_scores[level], [-1, num_boxes]) this_level_boxes = tf.reshape(rpn_boxes[level], [-1, num_boxes, 4]) this_level_anchors = tf.cast( tf.reshape(anchor_boxes[level], [-1, num_boxes, 4]), dtype=this_level_scores.dtype) if rpn_object_scores: this_level_object_scores = rpn_object_scores[level] this_level_object_scores = tf.reshape(this_level_object_scores, [-1, num_boxes]) this_level_object_scores = tf.cast(this_level_object_scores, this_level_scores.dtype) this_level_scores = this_level_object_scores if apply_sigmoid_to_score: this_level_scores = tf.sigmoid(this_level_scores) if decode_boxes: if is_box_lrtb: # Box in left-right-top-bottom format. this_level_boxes = box_utils.decode_boxes_lrtb( this_level_boxes, this_level_anchors) else: # Box in standard x-y-h-w format. this_level_boxes = box_utils.decode_boxes( this_level_boxes, this_level_anchors) if clip_boxes: this_level_boxes = box_utils.clip_boxes( this_level_boxes, image_shape) if rpn_min_size_threshold > 0.0: this_level_boxes, this_level_scores = box_utils.filter_boxes( this_level_boxes, this_level_scores, image_shape, rpn_min_size_threshold) this_level_pre_nms_top_k = min(num_boxes, rpn_pre_nms_top_k) this_level_post_nms_top_k = min(num_boxes, rpn_post_nms_top_k) if rpn_nms_threshold > 0.0: if use_batched_nms: this_level_rois, this_level_roi_scores, _, _ = ( tf.image.combined_non_max_suppression( tf.expand_dims(this_level_boxes, axis=2), tf.expand_dims(this_level_scores, axis=-1), max_output_size_per_class=this_level_pre_nms_top_k, max_total_size=this_level_post_nms_top_k, iou_threshold=rpn_nms_threshold, score_threshold=rpn_score_threshold, pad_per_class=False, clip_boxes=False)) else: if rpn_score_threshold > 0.0: this_level_boxes, this_level_scores = ( box_utils.filter_boxes_by_scores(this_level_boxes, this_level_scores, rpn_score_threshold)) this_level_boxes, this_level_scores = box_utils.top_k_boxes( this_level_boxes, this_level_scores, k=this_level_pre_nms_top_k) this_level_roi_scores, this_level_rois = ( nms.sorted_non_max_suppression_padded( this_level_scores, this_level_boxes, max_output_size=this_level_post_nms_top_k, iou_threshold=rpn_nms_threshold)) else: this_level_rois, this_level_roi_scores = box_utils.top_k_boxes( this_level_rois, this_level_scores, k=this_level_post_nms_top_k) rois.append(this_level_rois) roi_scores.append(this_level_roi_scores) all_rois = tf.concat(rois, axis=1) all_roi_scores = tf.concat(roi_scores, axis=1) with tf.name_scope('top_k_rois'): _, num_valid_rois = all_roi_scores.get_shape().as_list() overall_top_k = min(num_valid_rois, rpn_post_nms_top_k) selected_rois, selected_roi_scores = box_utils.top_k_boxes( all_rois, all_roi_scores, k=overall_top_k) return selected_rois, selected_roi_scores
22,986
48.012793
80
py
models
models-master/official/legacy/detection/ops/postprocess_ops.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Post-processing model outputs to generate detection.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import functools import tensorflow as tf from official.legacy.detection.ops import nms from official.legacy.detection.utils import box_utils def generate_detections_factory(params): """Factory to select function to generate detection.""" if params.use_batched_nms: func = functools.partial( _generate_detections_batched, max_total_size=params.max_total_size, nms_iou_threshold=params.nms_iou_threshold, score_threshold=params.score_threshold) else: func = functools.partial( _generate_detections, max_total_size=params.max_total_size, nms_iou_threshold=params.nms_iou_threshold, score_threshold=params.score_threshold, pre_nms_num_boxes=params.pre_nms_num_boxes) return func def _select_top_k_scores(scores_in, pre_nms_num_detections): """Select top_k scores and indices for each class. Args: scores_in: a Tensor with shape [batch_size, N, num_classes], which stacks class logit outputs on all feature levels. The N is the number of total anchors on all levels. The num_classes is the number of classes predicted by the model. pre_nms_num_detections: Number of candidates before NMS. Returns: scores and indices: Tensors with shape [batch_size, pre_nms_num_detections, num_classes]. """ batch_size, num_anchors, num_class = scores_in.get_shape().as_list() scores_trans = tf.transpose(scores_in, perm=[0, 2, 1]) scores_trans = tf.reshape(scores_trans, [-1, num_anchors]) top_k_scores, top_k_indices = tf.nn.top_k( scores_trans, k=pre_nms_num_detections, sorted=True) top_k_scores = tf.reshape(top_k_scores, [batch_size, num_class, pre_nms_num_detections]) top_k_indices = tf.reshape(top_k_indices, [batch_size, num_class, pre_nms_num_detections]) return tf.transpose(top_k_scores, [0, 2, 1]), tf.transpose(top_k_indices, [0, 2, 1]) def _generate_detections(boxes, scores, max_total_size=100, nms_iou_threshold=0.3, score_threshold=0.05, pre_nms_num_boxes=5000): """Generate the final detections given the model outputs. This uses classes unrolling with while loop based NMS, could be parralled at batch dimension. Args: boxes: a tensor with shape [batch_size, N, num_classes, 4] or [batch_size, N, 1, 4], which box predictions on all feature levels. The N is the number of total anchors on all levels. scores: a tensor with shape [batch_size, N, num_classes], which stacks class probability on all feature levels. The N is the number of total anchors on all levels. The num_classes is the number of classes predicted by the model. Note that the class_outputs here is the raw score. max_total_size: a scalar representing maximum number of boxes retained over all classes. nms_iou_threshold: a float representing the threshold for deciding whether boxes overlap too much with respect to IOU. score_threshold: a float representing the threshold for deciding when to remove boxes based on score. pre_nms_num_boxes: an int number of top candidate detections per class before NMS. Returns: nms_boxes: `float` Tensor of shape [batch_size, max_total_size, 4] representing top detected boxes in [y1, x1, y2, x2]. nms_scores: `float` Tensor of shape [batch_size, max_total_size] representing sorted confidence scores for detected boxes. The values are between [0, 1]. nms_classes: `int` Tensor of shape [batch_size, max_total_size] representing classes for detected boxes. valid_detections: `int` Tensor of shape [batch_size] only the top `valid_detections` boxes are valid detections. """ with tf.name_scope('generate_detections'): nmsed_boxes = [] nmsed_classes = [] nmsed_scores = [] valid_detections = [] batch_size, _, num_classes_for_box, _ = boxes.get_shape().as_list() _, total_anchors, num_classes = scores.get_shape().as_list() # Selects top pre_nms_num scores and indices before NMS. scores, indices = _select_top_k_scores( scores, min(total_anchors, pre_nms_num_boxes)) for i in range(num_classes): boxes_i = boxes[:, :, min(num_classes_for_box - 1, i), :] scores_i = scores[:, :, i] # Obtains pre_nms_num_boxes before running NMS. boxes_i = tf.gather(boxes_i, indices[:, :, i], batch_dims=1, axis=1) # Filter out scores. boxes_i, scores_i = box_utils.filter_boxes_by_scores( boxes_i, scores_i, min_score_threshold=score_threshold) (nmsed_scores_i, nmsed_boxes_i) = nms.sorted_non_max_suppression_padded( tf.cast(scores_i, tf.float32), tf.cast(boxes_i, tf.float32), max_total_size, iou_threshold=nms_iou_threshold) nmsed_classes_i = tf.fill([batch_size, max_total_size], i) nmsed_boxes.append(nmsed_boxes_i) nmsed_scores.append(nmsed_scores_i) nmsed_classes.append(nmsed_classes_i) nmsed_boxes = tf.concat(nmsed_boxes, axis=1) nmsed_scores = tf.concat(nmsed_scores, axis=1) nmsed_classes = tf.concat(nmsed_classes, axis=1) nmsed_scores, indices = tf.nn.top_k( nmsed_scores, k=max_total_size, sorted=True) nmsed_boxes = tf.gather(nmsed_boxes, indices, batch_dims=1, axis=1) nmsed_classes = tf.gather(nmsed_classes, indices, batch_dims=1) valid_detections = tf.reduce_sum( input_tensor=tf.cast(tf.greater(nmsed_scores, -1), tf.int32), axis=1) return nmsed_boxes, nmsed_scores, nmsed_classes, valid_detections def _generate_detections_per_image(boxes, scores, max_total_size=100, nms_iou_threshold=0.3, score_threshold=0.05, pre_nms_num_boxes=5000): """Generate the final detections per image given the model outputs. Args: boxes: a tensor with shape [N, num_classes, 4] or [N, 1, 4], which box predictions on all feature levels. The N is the number of total anchors on all levels. scores: a tensor with shape [N, num_classes], which stacks class probability on all feature levels. The N is the number of total anchors on all levels. The num_classes is the number of classes predicted by the model. Note that the class_outputs here is the raw score. max_total_size: a scalar representing maximum number of boxes retained over all classes. nms_iou_threshold: a float representing the threshold for deciding whether boxes overlap too much with respect to IOU. score_threshold: a float representing the threshold for deciding when to remove boxes based on score. pre_nms_num_boxes: an int number of top candidate detections per class before NMS. Returns: nms_boxes: `float` Tensor of shape [max_total_size, 4] representing top detected boxes in [y1, x1, y2, x2]. nms_scores: `float` Tensor of shape [max_total_size] representing sorted confidence scores for detected boxes. The values are between [0, 1]. nms_classes: `int` Tensor of shape [max_total_size] representing classes for detected boxes. valid_detections: `int` Tensor of shape [1] only the top `valid_detections` boxes are valid detections. """ nmsed_boxes = [] nmsed_scores = [] nmsed_classes = [] num_classes_for_box = boxes.get_shape().as_list()[1] num_classes = scores.get_shape().as_list()[1] for i in range(num_classes): boxes_i = boxes[:, min(num_classes_for_box - 1, i)] scores_i = scores[:, i] # Obtains pre_nms_num_boxes before running NMS. scores_i, indices = tf.nn.top_k( scores_i, k=tf.minimum(tf.shape(input=scores_i)[-1], pre_nms_num_boxes)) boxes_i = tf.gather(boxes_i, indices) (nmsed_indices_i, nmsed_num_valid_i) = tf.image.non_max_suppression_padded( tf.cast(boxes_i, tf.float32), tf.cast(scores_i, tf.float32), max_total_size, iou_threshold=nms_iou_threshold, score_threshold=score_threshold, pad_to_max_output_size=True, name='nms_detections_' + str(i)) nmsed_boxes_i = tf.gather(boxes_i, nmsed_indices_i) nmsed_scores_i = tf.gather(scores_i, nmsed_indices_i) # Sets scores of invalid boxes to -1. nmsed_scores_i = tf.where( tf.less(tf.range(max_total_size), [nmsed_num_valid_i]), nmsed_scores_i, -tf.ones_like(nmsed_scores_i)) nmsed_classes_i = tf.fill([max_total_size], i) nmsed_boxes.append(nmsed_boxes_i) nmsed_scores.append(nmsed_scores_i) nmsed_classes.append(nmsed_classes_i) # Concats results from all classes and sort them. nmsed_boxes = tf.concat(nmsed_boxes, axis=0) nmsed_scores = tf.concat(nmsed_scores, axis=0) nmsed_classes = tf.concat(nmsed_classes, axis=0) nmsed_scores, indices = tf.nn.top_k( nmsed_scores, k=max_total_size, sorted=True) nmsed_boxes = tf.gather(nmsed_boxes, indices) nmsed_classes = tf.gather(nmsed_classes, indices) valid_detections = tf.reduce_sum( input_tensor=tf.cast(tf.greater(nmsed_scores, -1), tf.int32)) return nmsed_boxes, nmsed_scores, nmsed_classes, valid_detections def _generate_detections_batched(boxes, scores, max_total_size, nms_iou_threshold, score_threshold): """Generates detected boxes with scores and classes for one-stage detector. The function takes output of multi-level ConvNets and anchor boxes and generates detected boxes. Note that this used batched nms, which is not supported on TPU currently. Args: boxes: a tensor with shape [batch_size, N, num_classes, 4] or [batch_size, N, 1, 4], which box predictions on all feature levels. The N is the number of total anchors on all levels. scores: a tensor with shape [batch_size, N, num_classes], which stacks class probability on all feature levels. The N is the number of total anchors on all levels. The num_classes is the number of classes predicted by the model. Note that the class_outputs here is the raw score. max_total_size: a scalar representing maximum number of boxes retained over all classes. nms_iou_threshold: a float representing the threshold for deciding whether boxes overlap too much with respect to IOU. score_threshold: a float representing the threshold for deciding when to remove boxes based on score. Returns: nms_boxes: `float` Tensor of shape [batch_size, max_total_size, 4] representing top detected boxes in [y1, x1, y2, x2]. nms_scores: `float` Tensor of shape [batch_size, max_total_size] representing sorted confidence scores for detected boxes. The values are between [0, 1]. nms_classes: `int` Tensor of shape [batch_size, max_total_size] representing classes for detected boxes. valid_detections: `int` Tensor of shape [batch_size] only the top `valid_detections` boxes are valid detections. """ with tf.name_scope('generate_detections'): # TODO(tsungyi): Removes normalization/denomalization once the # tf.image.combined_non_max_suppression is coordinate system agnostic. # Normalizes maximum box cooridinates to 1. normalizer = tf.reduce_max(boxes) boxes /= normalizer (nmsed_boxes, nmsed_scores, nmsed_classes, valid_detections) = tf.image.combined_non_max_suppression( boxes, scores, max_output_size_per_class=max_total_size, max_total_size=max_total_size, iou_threshold=nms_iou_threshold, score_threshold=score_threshold, pad_per_class=False, ) # De-normalizes box cooridinates. nmsed_boxes *= normalizer nmsed_classes = tf.cast(nmsed_classes, tf.int32) return nmsed_boxes, nmsed_scores, nmsed_classes, valid_detections class MultilevelDetectionGenerator(tf.keras.layers.Layer): """Generates detected boxes with scores and classes for one-stage detector.""" def __init__(self, min_level, max_level, params): self._min_level = min_level self._max_level = max_level self._generate_detections = generate_detections_factory(params) super(MultilevelDetectionGenerator, self).__init__(autocast=False) def call(self, box_outputs, class_outputs, anchor_boxes, image_shape): # Collects outputs from all levels into a list. boxes = [] scores = [] for i in range(self._min_level, self._max_level + 1): box_outputs_i_shape = tf.shape(box_outputs[i]) batch_size = box_outputs_i_shape[0] num_anchors_per_locations = box_outputs_i_shape[-1] // 4 num_classes = tf.shape(class_outputs[i])[-1] // num_anchors_per_locations # Applies score transformation and remove the implicit background class. scores_i = tf.sigmoid( tf.reshape(class_outputs[i], [batch_size, -1, num_classes])) scores_i = tf.slice(scores_i, [0, 0, 1], [-1, -1, -1]) # Box decoding. # The anchor boxes are shared for all data in a batch. # One stage detector only supports class agnostic box regression. anchor_boxes_i = tf.reshape(anchor_boxes[i], [batch_size, -1, 4]) box_outputs_i = tf.reshape(box_outputs[i], [batch_size, -1, 4]) boxes_i = box_utils.decode_boxes(box_outputs_i, anchor_boxes_i) # Box clipping. boxes_i = box_utils.clip_boxes(boxes_i, image_shape) boxes.append(boxes_i) scores.append(scores_i) boxes = tf.concat(boxes, axis=1) scores = tf.concat(scores, axis=1) nmsed_boxes, nmsed_scores, nmsed_classes, valid_detections = ( self._generate_detections(tf.expand_dims(boxes, axis=2), scores)) # Adds 1 to offset the background class which has index 0. nmsed_classes += 1 return nmsed_boxes, nmsed_scores, nmsed_classes, valid_detections class GenericDetectionGenerator(tf.keras.layers.Layer): """Generates the final detected boxes with scores and classes.""" def __init__(self, params): super(GenericDetectionGenerator, self).__init__(autocast=False) self._generate_detections = generate_detections_factory(params) def call(self, box_outputs, class_outputs, anchor_boxes, image_shape): """Generate final detections. Args: box_outputs: a tensor of shape of [batch_size, K, num_classes * 4] representing the class-specific box coordinates relative to anchors. class_outputs: a tensor of shape of [batch_size, K, num_classes] representing the class logits before applying score activiation. anchor_boxes: a tensor of shape of [batch_size, K, 4] representing the corresponding anchor boxes w.r.t `box_outputs`. image_shape: a tensor of shape of [batch_size, 2] storing the image height and width w.r.t. the scaled image, i.e. the same image space as `box_outputs` and `anchor_boxes`. Returns: nms_boxes: `float` Tensor of shape [batch_size, max_total_size, 4] representing top detected boxes in [y1, x1, y2, x2]. nms_scores: `float` Tensor of shape [batch_size, max_total_size] representing sorted confidence scores for detected boxes. The values are between [0, 1]. nms_classes: `int` Tensor of shape [batch_size, max_total_size] representing classes for detected boxes. valid_detections: `int` Tensor of shape [batch_size] only the top `valid_detections` boxes are valid detections. """ class_outputs = tf.nn.softmax(class_outputs, axis=-1) # Removes the background class. class_outputs_shape = tf.shape(class_outputs) batch_size = class_outputs_shape[0] num_locations = class_outputs_shape[1] num_classes = class_outputs_shape[-1] num_detections = num_locations * (num_classes - 1) class_outputs = tf.slice(class_outputs, [0, 0, 1], [-1, -1, -1]) box_outputs = tf.reshape( box_outputs, tf.stack([batch_size, num_locations, num_classes, 4], axis=-1)) box_outputs = tf.slice(box_outputs, [0, 0, 1, 0], [-1, -1, -1, -1]) anchor_boxes = tf.tile( tf.expand_dims(anchor_boxes, axis=2), [1, 1, num_classes - 1, 1]) box_outputs = tf.reshape(box_outputs, tf.stack([batch_size, num_detections, 4], axis=-1)) anchor_boxes = tf.reshape( anchor_boxes, tf.stack([batch_size, num_detections, 4], axis=-1)) # Box decoding. decoded_boxes = box_utils.decode_boxes( box_outputs, anchor_boxes, weights=[10.0, 10.0, 5.0, 5.0]) # Box clipping decoded_boxes = box_utils.clip_boxes(decoded_boxes, image_shape) decoded_boxes = tf.reshape( decoded_boxes, tf.stack([batch_size, num_locations, num_classes - 1, 4], axis=-1)) nmsed_boxes, nmsed_scores, nmsed_classes, valid_detections = ( self._generate_detections(decoded_boxes, class_outputs)) # Adds 1 to offset the background class which has index 0. nmsed_classes += 1 return nmsed_boxes, nmsed_scores, nmsed_classes, valid_detections class OlnDetectionGenerator(GenericDetectionGenerator): """Generates the final detected boxes with scores and classes.""" def __call__(self, box_outputs, class_outputs, anchor_boxes, image_shape, is_single_fg_score=False, keep_nms=True): """Generate final detections for Object Localization Network (OLN). Args: box_outputs: a tensor of shape of [batch_size, K, num_classes * 4] representing the class-specific box coordinates relative to anchors. class_outputs: a tensor of shape of [batch_size, K, num_classes] representing the class logits before applying score activiation. anchor_boxes: a tensor of shape of [batch_size, K, 4] representing the corresponding anchor boxes w.r.t `box_outputs`. image_shape: a tensor of shape of [batch_size, 2] storing the image height and width w.r.t. the scaled image, i.e. the same image space as `box_outputs` and `anchor_boxes`. is_single_fg_score: a Bool indicator of whether class_outputs includes the background scores concatenated or not. By default, class_outputs is a concatenation of both scores for the foreground and background. That is, scores_without_bg=False. keep_nms: a Bool indicator of whether to perform NMS or not. Returns: nms_boxes: `float` Tensor of shape [batch_size, max_total_size, 4] representing top detected boxes in [y1, x1, y2, x2]. nms_scores: `float` Tensor of shape [batch_size, max_total_size] representing sorted confidence scores for detected boxes. The values are between [0, 1]. nms_classes: `int` Tensor of shape [batch_size, max_total_size] representing classes for detected boxes. valid_detections: `int` Tensor of shape [batch_size] only the top `valid_detections` boxes are valid detections. """ if is_single_fg_score: # Concatenates dummy background scores. dummy_bg_scores = tf.zeros_like(class_outputs) class_outputs = tf.stack([dummy_bg_scores, class_outputs], -1) else: class_outputs = tf.nn.softmax(class_outputs, axis=-1) # Removes the background class. class_outputs_shape = tf.shape(class_outputs) batch_size = class_outputs_shape[0] num_locations = class_outputs_shape[1] num_classes = class_outputs_shape[-1] num_detections = num_locations * (num_classes - 1) class_outputs = tf.slice(class_outputs, [0, 0, 1], [-1, -1, -1]) box_outputs = tf.reshape( box_outputs, tf.stack([batch_size, num_locations, num_classes, 4], axis=-1)) box_outputs = tf.slice(box_outputs, [0, 0, 1, 0], [-1, -1, -1, -1]) anchor_boxes = tf.tile( tf.expand_dims(anchor_boxes, axis=2), [1, 1, num_classes - 1, 1]) box_outputs = tf.reshape(box_outputs, tf.stack([batch_size, num_detections, 4], axis=-1)) anchor_boxes = tf.reshape( anchor_boxes, tf.stack([batch_size, num_detections, 4], axis=-1)) # Box decoding. For RPN outputs, box_outputs are all zeros. decoded_boxes = box_utils.decode_boxes( box_outputs, anchor_boxes, weights=[10.0, 10.0, 5.0, 5.0]) # Box clipping decoded_boxes = box_utils.clip_boxes(decoded_boxes, image_shape) decoded_boxes = tf.reshape( decoded_boxes, tf.stack([batch_size, num_locations, num_classes - 1, 4], axis=-1)) if keep_nms: nmsed_boxes, nmsed_scores, nmsed_classes, valid_detections = ( self._generate_detections(decoded_boxes, class_outputs)) # Adds 1 to offset the background class which has index 0. nmsed_classes += 1 else: nmsed_boxes = decoded_boxes[:, :, 0, :] nmsed_scores = class_outputs[:, :, 0] nmsed_classes = tf.cast(tf.ones_like(nmsed_scores), tf.int32) valid_detections = tf.cast( tf.reduce_sum(tf.ones_like(nmsed_scores), axis=-1), tf.int32) return nmsed_boxes, nmsed_scores, nmsed_classes, valid_detections
21,942
43.062249
80
py
models
models-master/official/legacy/detection/ops/target_ops.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Target and sampling related ops.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from official.legacy.detection.ops import spatial_transform_ops from official.legacy.detection.utils import box_utils from official.vision.utils.object_detection import balanced_positive_negative_sampler def box_matching(boxes, gt_boxes, gt_classes): """Match boxes to groundtruth boxes. Given the proposal boxes and the groundtruth boxes and classes, perform the groundtruth matching by taking the argmax of the IoU between boxes and groundtruth boxes. Args: boxes: a tensor of shape of [batch_size, N, 4] representing the box coordiantes to be matched to groundtruth boxes. gt_boxes: a tensor of shape of [batch_size, MAX_INSTANCES, 4] representing the groundtruth box coordinates. It is padded with -1s to indicate the invalid boxes. gt_classes: [batch_size, MAX_INSTANCES] representing the groundtruth box classes. It is padded with -1s to indicate the invalid classes. Returns: matched_gt_boxes: a tensor of shape of [batch_size, N, 4], representing the matched groundtruth box coordinates for each input box. If the box does not overlap with any groundtruth boxes, the matched boxes of it will be set to all 0s. matched_gt_classes: a tensor of shape of [batch_size, N], representing the matched groundtruth classes for each input box. If the box does not overlap with any groundtruth boxes, the matched box classes of it will be set to 0, which corresponds to the background class. matched_gt_indices: a tensor of shape of [batch_size, N], representing the indices of the matched groundtruth boxes in the original gt_boxes tensor. If the box does not overlap with any groundtruth boxes, the index of the matched groundtruth will be set to -1. matched_iou: a tensor of shape of [batch_size, N], representing the IoU between the box and its matched groundtruth box. The matched IoU is the maximum IoU of the box and all the groundtruth boxes. iou: a tensor of shape of [batch_size, N, K], representing the IoU matrix between boxes and the groundtruth boxes. The IoU between a box and the invalid groundtruth boxes whose coordinates are [-1, -1, -1, -1] is -1. """ # Compute IoU between boxes and gt_boxes. # iou <- [batch_size, N, K] iou = box_utils.bbox_overlap(boxes, gt_boxes) # max_iou <- [batch_size, N] # 0.0 -> no match to gt, or -1.0 match to no gt matched_iou = tf.reduce_max(iou, axis=-1) # background_box_mask <- bool, [batch_size, N] background_box_mask = tf.less_equal(matched_iou, 0.0) argmax_iou_indices = tf.argmax(iou, axis=-1, output_type=tf.int32) argmax_iou_indices_shape = tf.shape(argmax_iou_indices) batch_indices = ( tf.expand_dims(tf.range(argmax_iou_indices_shape[0]), axis=-1) * tf.ones([1, argmax_iou_indices_shape[-1]], dtype=tf.int32)) gather_nd_indices = tf.stack([batch_indices, argmax_iou_indices], axis=-1) matched_gt_boxes = tf.gather_nd(gt_boxes, gather_nd_indices) matched_gt_boxes = tf.where( tf.tile(tf.expand_dims(background_box_mask, axis=-1), [1, 1, 4]), tf.zeros_like(matched_gt_boxes, dtype=matched_gt_boxes.dtype), matched_gt_boxes) matched_gt_classes = tf.gather_nd(gt_classes, gather_nd_indices) matched_gt_classes = tf.where(background_box_mask, tf.zeros_like(matched_gt_classes), matched_gt_classes) matched_gt_indices = tf.where(background_box_mask, -tf.ones_like(argmax_iou_indices), argmax_iou_indices) return (matched_gt_boxes, matched_gt_classes, matched_gt_indices, matched_iou, iou) def assign_and_sample_proposals(proposed_boxes, gt_boxes, gt_classes, num_samples_per_image=512, mix_gt_boxes=True, fg_fraction=0.25, fg_iou_thresh=0.5, bg_iou_thresh_hi=0.5, bg_iou_thresh_lo=0.0): """Assigns the proposals with groundtruth classes and performs subsmpling. Given `proposed_boxes`, `gt_boxes`, and `gt_classes`, the function uses the following algorithm to generate the final `num_samples_per_image` RoIs. 1. Calculates the IoU between each proposal box and each gt_boxes. 2. Assigns each proposed box with a groundtruth class and box by choosing the largest IoU overlap. 3. Samples `num_samples_per_image` boxes from all proposed boxes, and returns box_targets, class_targets, and RoIs. Args: proposed_boxes: a tensor of shape of [batch_size, N, 4]. N is the number of proposals before groundtruth assignment. The last dimension is the box coordinates w.r.t. the scaled images in [ymin, xmin, ymax, xmax] format. gt_boxes: a tensor of shape of [batch_size, MAX_NUM_INSTANCES, 4]. The coordinates of gt_boxes are in the pixel coordinates of the scaled image. This tensor might have padding of values -1 indicating the invalid box coordinates. gt_classes: a tensor with a shape of [batch_size, MAX_NUM_INSTANCES]. This tensor might have paddings with values of -1 indicating the invalid classes. num_samples_per_image: a integer represents RoI minibatch size per image. mix_gt_boxes: a bool indicating whether to mix the groundtruth boxes before sampling proposals. fg_fraction: a float represents the target fraction of RoI minibatch that is labeled foreground (i.e., class > 0). fg_iou_thresh: a float represents the IoU overlap threshold for an RoI to be considered foreground (if >= fg_iou_thresh). bg_iou_thresh_hi: a float represents the IoU overlap threshold for an RoI to be considered background (class = 0 if overlap in [LO, HI)). bg_iou_thresh_lo: a float represents the IoU overlap threshold for an RoI to be considered background (class = 0 if overlap in [LO, HI)). Returns: sampled_rois: a tensor of shape of [batch_size, K, 4], representing the coordinates of the sampled RoIs, where K is the number of the sampled RoIs, i.e. K = num_samples_per_image. sampled_gt_boxes: a tensor of shape of [batch_size, K, 4], storing the box coordinates of the matched groundtruth boxes of the samples RoIs. sampled_gt_classes: a tensor of shape of [batch_size, K], storing the classes of the matched groundtruth boxes of the sampled RoIs. sampled_gt_indices: a tensor of shape of [batch_size, K], storing the indices of the sampled groudntruth boxes in the original `gt_boxes` tensor, i.e. gt_boxes[sampled_gt_indices[:, i]] = sampled_gt_boxes[:, i]. """ with tf.name_scope('sample_proposals'): if mix_gt_boxes: boxes = tf.concat([proposed_boxes, gt_boxes], axis=1) else: boxes = proposed_boxes (matched_gt_boxes, matched_gt_classes, matched_gt_indices, matched_iou, _) = box_matching(boxes, gt_boxes, gt_classes) positive_match = tf.greater(matched_iou, fg_iou_thresh) negative_match = tf.logical_and( tf.greater_equal(matched_iou, bg_iou_thresh_lo), tf.less(matched_iou, bg_iou_thresh_hi)) ignored_match = tf.less(matched_iou, 0.0) # re-assign negatively matched boxes to the background class. matched_gt_classes = tf.where(negative_match, tf.zeros_like(matched_gt_classes), matched_gt_classes) matched_gt_indices = tf.where(negative_match, tf.zeros_like(matched_gt_indices), matched_gt_indices) sample_candidates = tf.logical_and( tf.logical_or(positive_match, negative_match), tf.logical_not(ignored_match)) sampler = ( balanced_positive_negative_sampler.BalancedPositiveNegativeSampler( positive_fraction=fg_fraction, is_static=True)) batch_size, _ = sample_candidates.get_shape().as_list() sampled_indicators = [] for i in range(batch_size): sampled_indicator = sampler.subsample(sample_candidates[i], num_samples_per_image, positive_match[i]) sampled_indicators.append(sampled_indicator) sampled_indicators = tf.stack(sampled_indicators) _, sampled_indices = tf.nn.top_k( tf.cast(sampled_indicators, dtype=tf.int32), k=num_samples_per_image, sorted=True) sampled_indices_shape = tf.shape(sampled_indices) batch_indices = ( tf.expand_dims(tf.range(sampled_indices_shape[0]), axis=-1) * tf.ones([1, sampled_indices_shape[-1]], dtype=tf.int32)) gather_nd_indices = tf.stack([batch_indices, sampled_indices], axis=-1) sampled_rois = tf.gather_nd(boxes, gather_nd_indices) sampled_gt_boxes = tf.gather_nd(matched_gt_boxes, gather_nd_indices) sampled_gt_classes = tf.gather_nd(matched_gt_classes, gather_nd_indices) sampled_gt_indices = tf.gather_nd(matched_gt_indices, gather_nd_indices) return (sampled_rois, sampled_gt_boxes, sampled_gt_classes, sampled_gt_indices) def sample_and_crop_foreground_masks(candidate_rois, candidate_gt_boxes, candidate_gt_classes, candidate_gt_indices, gt_masks, num_mask_samples_per_image=128, mask_target_size=28): """Samples and creates cropped foreground masks for training. Args: candidate_rois: a tensor of shape of [batch_size, N, 4], where N is the number of candidate RoIs to be considered for mask sampling. It includes both positive and negative RoIs. The `num_mask_samples_per_image` positive RoIs will be sampled to create mask training targets. candidate_gt_boxes: a tensor of shape of [batch_size, N, 4], storing the corresponding groundtruth boxes to the `candidate_rois`. candidate_gt_classes: a tensor of shape of [batch_size, N], storing the corresponding groundtruth classes to the `candidate_rois`. 0 in the tensor corresponds to the background class, i.e. negative RoIs. candidate_gt_indices: a tensor of shape [batch_size, N], storing the corresponding groundtruth instance indices to the `candidate_gt_boxes`, i.e. gt_boxes[candidate_gt_indices[:, i]] = candidate_gt_boxes[:, i] and gt_boxes which is of shape [batch_size, MAX_INSTANCES, 4], M >= N, is the superset of candidate_gt_boxes. gt_masks: a tensor of [batch_size, MAX_INSTANCES, mask_height, mask_width] containing all the groundtruth masks which sample masks are drawn from. num_mask_samples_per_image: an integer which specifies the number of masks to sample. mask_target_size: an integer which specifies the final cropped mask size after sampling. The output masks are resized w.r.t the sampled RoIs. Returns: foreground_rois: a tensor of shape of [batch_size, K, 4] storing the RoI that corresponds to the sampled foreground masks, where K = num_mask_samples_per_image. foreground_classes: a tensor of shape of [batch_size, K] storing the classes corresponding to the sampled foreground masks. cropoped_foreground_masks: a tensor of shape of [batch_size, K, mask_target_size, mask_target_size] storing the cropped foreground masks used for training. """ with tf.name_scope('sample_and_crop_foreground_masks'): _, fg_instance_indices = tf.nn.top_k( tf.cast(tf.greater(candidate_gt_classes, 0), dtype=tf.int32), k=num_mask_samples_per_image) fg_instance_indices_shape = tf.shape(fg_instance_indices) batch_indices = ( tf.expand_dims(tf.range(fg_instance_indices_shape[0]), axis=-1) * tf.ones([1, fg_instance_indices_shape[-1]], dtype=tf.int32)) gather_nd_instance_indices = tf.stack([batch_indices, fg_instance_indices], axis=-1) foreground_rois = tf.gather_nd(candidate_rois, gather_nd_instance_indices) foreground_boxes = tf.gather_nd(candidate_gt_boxes, gather_nd_instance_indices) foreground_classes = tf.gather_nd(candidate_gt_classes, gather_nd_instance_indices) foreground_gt_indices = tf.gather_nd(candidate_gt_indices, gather_nd_instance_indices) foreground_gt_indices_shape = tf.shape(foreground_gt_indices) batch_indices = ( tf.expand_dims(tf.range(foreground_gt_indices_shape[0]), axis=-1) * tf.ones([1, foreground_gt_indices_shape[-1]], dtype=tf.int32)) gather_nd_gt_indices = tf.stack([batch_indices, foreground_gt_indices], axis=-1) foreground_masks = tf.gather_nd(gt_masks, gather_nd_gt_indices) cropped_foreground_masks = spatial_transform_ops.crop_mask_in_target_box( foreground_masks, foreground_boxes, foreground_rois, mask_target_size, sample_offset=0.5) return foreground_rois, foreground_classes, cropped_foreground_masks class ROISampler(tf.keras.layers.Layer): """Samples RoIs and creates training targets.""" def __init__(self, params): self._num_samples_per_image = params.num_samples_per_image self._fg_fraction = params.fg_fraction self._fg_iou_thresh = params.fg_iou_thresh self._bg_iou_thresh_hi = params.bg_iou_thresh_hi self._bg_iou_thresh_lo = params.bg_iou_thresh_lo self._mix_gt_boxes = params.mix_gt_boxes super(ROISampler, self).__init__(autocast=False) def call(self, rois, gt_boxes, gt_classes): """Sample and assign RoIs for training. Args: rois: a tensor of shape of [batch_size, N, 4]. N is the number of proposals before groundtruth assignment. The last dimension is the box coordinates w.r.t. the scaled images in [ymin, xmin, ymax, xmax] format. gt_boxes: a tensor of shape of [batch_size, MAX_NUM_INSTANCES, 4]. The coordinates of gt_boxes are in the pixel coordinates of the scaled image. This tensor might have padding of values -1 indicating the invalid box coordinates. gt_classes: a tensor with a shape of [batch_size, MAX_NUM_INSTANCES]. This tensor might have paddings with values of -1 indicating the invalid classes. Returns: sampled_rois: a tensor of shape of [batch_size, K, 4], representing the coordinates of the sampled RoIs, where K is the number of the sampled RoIs, i.e. K = num_samples_per_image. sampled_gt_boxes: a tensor of shape of [batch_size, K, 4], storing the box coordinates of the matched groundtruth boxes of the samples RoIs. sampled_gt_classes: a tensor of shape of [batch_size, K], storing the classes of the matched groundtruth boxes of the sampled RoIs. """ sampled_rois, sampled_gt_boxes, sampled_gt_classes, sampled_gt_indices = ( assign_and_sample_proposals( rois, gt_boxes, gt_classes, num_samples_per_image=self._num_samples_per_image, mix_gt_boxes=self._mix_gt_boxes, fg_fraction=self._fg_fraction, fg_iou_thresh=self._fg_iou_thresh, bg_iou_thresh_hi=self._bg_iou_thresh_hi, bg_iou_thresh_lo=self._bg_iou_thresh_lo)) return (sampled_rois, sampled_gt_boxes, sampled_gt_classes, sampled_gt_indices) class ROIScoreSampler(ROISampler): """Samples RoIs, RoI-scores and creates training targets.""" def __call__(self, rois, roi_scores, gt_boxes, gt_classes): """Sample and assign RoIs for training. Args: rois: a tensor of shape of [batch_size, N, 4]. N is the number of proposals before groundtruth assignment. The last dimension is the box coordinates w.r.t. the scaled images in [ymin, xmin, ymax, xmax] format. roi_scores: gt_boxes: a tensor of shape of [batch_size, MAX_NUM_INSTANCES, 4]. The coordinates of gt_boxes are in the pixel coordinates of the scaled image. This tensor might have padding of values -1 indicating the invalid box coordinates. gt_classes: a tensor with a shape of [batch_size, MAX_NUM_INSTANCES]. This tensor might have paddings with values of -1 indicating the invalid classes. Returns: sampled_rois: a tensor of shape of [batch_size, K, 4], representing the coordinates of the sampled RoIs, where K is the number of the sampled RoIs, i.e. K = num_samples_per_image. sampled_roi_scores: sampled_gt_boxes: a tensor of shape of [batch_size, K, 4], storing the box coordinates of the matched groundtruth boxes of the samples RoIs. sampled_gt_classes: a tensor of shape of [batch_size, K], storing the classes of the matched groundtruth boxes of the sampled RoIs. """ (sampled_rois, sampled_roi_scores, sampled_gt_boxes, sampled_gt_classes, sampled_gt_indices) = ( self.assign_and_sample_proposals_and_scores( rois, roi_scores, gt_boxes, gt_classes, num_samples_per_image=self._num_samples_per_image, mix_gt_boxes=self._mix_gt_boxes, fg_fraction=self._fg_fraction, fg_iou_thresh=self._fg_iou_thresh, bg_iou_thresh_hi=self._bg_iou_thresh_hi, bg_iou_thresh_lo=self._bg_iou_thresh_lo)) return (sampled_rois, sampled_roi_scores, sampled_gt_boxes, sampled_gt_classes, sampled_gt_indices) def assign_and_sample_proposals_and_scores(self, proposed_boxes, proposed_scores, gt_boxes, gt_classes, num_samples_per_image=512, mix_gt_boxes=True, fg_fraction=0.25, fg_iou_thresh=0.5, bg_iou_thresh_hi=0.5, bg_iou_thresh_lo=0.0): """Assigns the proposals with groundtruth classes and performs subsmpling. Given `proposed_boxes`, `gt_boxes`, and `gt_classes`, the function uses the following algorithm to generate the final `num_samples_per_image` RoIs. 1. Calculates the IoU between each proposal box and each gt_boxes. 2. Assigns each proposed box with a groundtruth class and box by choosing the largest IoU overlap. 3. Samples `num_samples_per_image` boxes from all proposed boxes, and returns box_targets, class_targets, and RoIs. Args: proposed_boxes: a tensor of shape of [batch_size, N, 4]. N is the number of proposals before groundtruth assignment. The last dimension is the box coordinates w.r.t. the scaled images in [ymin, xmin, ymax, xmax] format. proposed_scores: a tensor of shape of [batch_size, N]. N is the number of proposals before groundtruth assignment. It is the rpn scores for all proposed boxes which can be either their classification or centerness scores. gt_boxes: a tensor of shape of [batch_size, MAX_NUM_INSTANCES, 4]. The coordinates of gt_boxes are in the pixel coordinates of the scaled image. This tensor might have padding of values -1 indicating the invalid box coordinates. gt_classes: a tensor with a shape of [batch_size, MAX_NUM_INSTANCES]. This tensor might have paddings with values of -1 indicating the invalid classes. num_samples_per_image: a integer represents RoI minibatch size per image. mix_gt_boxes: a bool indicating whether to mix the groundtruth boxes before sampling proposals. fg_fraction: a float represents the target fraction of RoI minibatch that is labeled foreground (i.e., class > 0). fg_iou_thresh: a float represents the IoU overlap threshold for an RoI to be considered foreground (if >= fg_iou_thresh). bg_iou_thresh_hi: a float represents the IoU overlap threshold for an RoI to be considered background (class = 0 if overlap in [LO, HI)). bg_iou_thresh_lo: a float represents the IoU overlap threshold for an RoI to be considered background (class = 0 if overlap in [LO, HI)). Returns: sampled_rois: a tensor of shape of [batch_size, K, 4], representing the coordinates of the sampled RoIs, where K is the number of the sampled RoIs, i.e. K = num_samples_per_image. sampled_scores: a tensor of shape of [batch_size, K], representing the confidence score of the sampled RoIs, where K is the number of the sampled RoIs, i.e. K = num_samples_per_image. sampled_gt_boxes: a tensor of shape of [batch_size, K, 4], storing the box coordinates of the matched groundtruth boxes of the samples RoIs. sampled_gt_classes: a tensor of shape of [batch_size, K], storing the classes of the matched groundtruth boxes of the sampled RoIs. sampled_gt_indices: a tensor of shape of [batch_size, K], storing the indices of the sampled groudntruth boxes in the original `gt_boxes` tensor, i.e. gt_boxes[sampled_gt_indices[:, i]] = sampled_gt_boxes[:, i]. """ with tf.name_scope('sample_proposals_and_scores'): if mix_gt_boxes: boxes = tf.concat([proposed_boxes, gt_boxes], axis=1) gt_scores = tf.ones_like(gt_boxes[:, :, 0]) scores = tf.concat([proposed_scores, gt_scores], axis=1) else: boxes = proposed_boxes scores = proposed_scores (matched_gt_boxes, matched_gt_classes, matched_gt_indices, matched_iou, _) = box_matching(boxes, gt_boxes, gt_classes) positive_match = tf.greater(matched_iou, fg_iou_thresh) negative_match = tf.logical_and( tf.greater_equal(matched_iou, bg_iou_thresh_lo), tf.less(matched_iou, bg_iou_thresh_hi)) ignored_match = tf.less(matched_iou, 0.0) # re-assign negatively matched boxes to the background class. matched_gt_classes = tf.where(negative_match, tf.zeros_like(matched_gt_classes), matched_gt_classes) matched_gt_indices = tf.where(negative_match, tf.zeros_like(matched_gt_indices), matched_gt_indices) sample_candidates = tf.logical_and( tf.logical_or(positive_match, negative_match), tf.logical_not(ignored_match)) sampler = ( balanced_positive_negative_sampler.BalancedPositiveNegativeSampler( positive_fraction=fg_fraction, is_static=True)) batch_size, _ = sample_candidates.get_shape().as_list() sampled_indicators = [] for i in range(batch_size): sampled_indicator = sampler.subsample(sample_candidates[i], num_samples_per_image, positive_match[i]) sampled_indicators.append(sampled_indicator) sampled_indicators = tf.stack(sampled_indicators) _, sampled_indices = tf.nn.top_k( tf.cast(sampled_indicators, dtype=tf.int32), k=num_samples_per_image, sorted=True) sampled_indices_shape = tf.shape(sampled_indices) batch_indices = ( tf.expand_dims(tf.range(sampled_indices_shape[0]), axis=-1) * tf.ones([1, sampled_indices_shape[-1]], dtype=tf.int32)) gather_nd_indices = tf.stack([batch_indices, sampled_indices], axis=-1) sampled_rois = tf.gather_nd(boxes, gather_nd_indices) sampled_roi_scores = tf.gather_nd(scores, gather_nd_indices) sampled_gt_boxes = tf.gather_nd(matched_gt_boxes, gather_nd_indices) sampled_gt_classes = tf.gather_nd(matched_gt_classes, gather_nd_indices) sampled_gt_indices = tf.gather_nd(matched_gt_indices, gather_nd_indices) return (sampled_rois, sampled_roi_scores, sampled_gt_boxes, sampled_gt_classes, sampled_gt_indices) class MaskSampler(tf.keras.layers.Layer): """Samples and creates mask training targets.""" def __init__(self, mask_target_size, num_mask_samples_per_image): self._mask_target_size = mask_target_size self._num_mask_samples_per_image = num_mask_samples_per_image super(MaskSampler, self).__init__(autocast=False) def call(self, candidate_rois, candidate_gt_boxes, candidate_gt_classes, candidate_gt_indices, gt_masks): """Sample and create mask targets for training. Args: candidate_rois: a tensor of shape of [batch_size, N, 4], where N is the number of candidate RoIs to be considered for mask sampling. It includes both positive and negative RoIs. The `num_mask_samples_per_image` positive RoIs will be sampled to create mask training targets. candidate_gt_boxes: a tensor of shape of [batch_size, N, 4], storing the corresponding groundtruth boxes to the `candidate_rois`. candidate_gt_classes: a tensor of shape of [batch_size, N], storing the corresponding groundtruth classes to the `candidate_rois`. 0 in the tensor corresponds to the background class, i.e. negative RoIs. candidate_gt_indices: a tensor of shape [batch_size, N], storing the corresponding groundtruth instance indices to the `candidate_gt_boxes`, i.e. gt_boxes[candidate_gt_indices[:, i]] = candidate_gt_boxes[:, i], where gt_boxes which is of shape [batch_size, MAX_INSTANCES, 4], M >= N, is the superset of candidate_gt_boxes. gt_masks: a tensor of [batch_size, MAX_INSTANCES, mask_height, mask_width] containing all the groundtruth masks which sample masks are drawn from. after sampling. The output masks are resized w.r.t the sampled RoIs. Returns: foreground_rois: a tensor of shape of [batch_size, K, 4] storing the RoI that corresponds to the sampled foreground masks, where K = num_mask_samples_per_image. foreground_classes: a tensor of shape of [batch_size, K] storing the classes corresponding to the sampled foreground masks. cropoped_foreground_masks: a tensor of shape of [batch_size, K, mask_target_size, mask_target_size] storing the cropped foreground masks used for training. """ foreground_rois, foreground_classes, cropped_foreground_masks = ( sample_and_crop_foreground_masks(candidate_rois, candidate_gt_boxes, candidate_gt_classes, candidate_gt_indices, gt_masks, self._num_mask_samples_per_image, self._mask_target_size)) return foreground_rois, foreground_classes, cropped_foreground_masks
28,304
48.484266
85
py
models
models-master/official/legacy/detection/ops/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/detection/ops/spatial_transform_ops.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Functions to performa spatial transformation for Tensor.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf _EPSILON = 1e-8 def nearest_upsampling(data, scale): """Nearest neighbor upsampling implementation. Args: data: A tensor with a shape of [batch, height_in, width_in, channels]. scale: An integer multiple to scale resolution of input data. Returns: data_up: A tensor with a shape of [batch, height_in*scale, width_in*scale, channels]. Same dtype as input data. """ with tf.name_scope('nearest_upsampling'): bs, _, _, c = data.get_shape().as_list() shape = tf.shape(input=data) h = shape[1] w = shape[2] bs = -1 if bs is None else bs # Uses reshape to quickly upsample the input. The nearest pixel is selected # implicitly via broadcasting. data = tf.reshape(data, [bs, h, 1, w, 1, c]) * tf.ones( [1, 1, scale, 1, scale, 1], dtype=data.dtype) return tf.reshape(data, [bs, h * scale, w * scale, c]) def feature_bilinear_interpolation(features, kernel_y, kernel_x): """Feature bilinear interpolation. The RoIAlign feature f can be computed by bilinear interpolation of four neighboring feature points f0, f1, f2, and f3. f(y, x) = [hy, ly] * [[f00, f01], * [hx, lx]^T [f10, f11]] f(y, x) = (hy*hx)f00 + (hy*lx)f01 + (ly*hx)f10 + (lx*ly)f11 f(y, x) = w00*f00 + w01*f01 + w10*f10 + w11*f11 kernel_y = [hy, ly] kernel_x = [hx, lx] Args: features: The features are in shape of [batch_size, num_boxes, output_size * 2, output_size * 2, num_filters]. kernel_y: Tensor of size [batch_size, boxes, output_size, 2, 1]. kernel_x: Tensor of size [batch_size, boxes, output_size, 2, 1]. Returns: A 5-D tensor representing feature crop of shape [batch_size, num_boxes, output_size, output_size, num_filters]. """ (batch_size, num_boxes, output_size, _, num_filters) = features.get_shape().as_list() output_size = output_size // 2 kernel_y = tf.reshape(kernel_y, [batch_size, num_boxes, output_size * 2, 1]) kernel_x = tf.reshape(kernel_x, [batch_size, num_boxes, 1, output_size * 2]) # Use implicit broadcast to generate the interpolation kernel. The # multiplier `4` is for avg pooling. interpolation_kernel = kernel_y * kernel_x * 4 # Interpolate the gathered features with computed interpolation kernels. features *= tf.cast( tf.expand_dims(interpolation_kernel, axis=-1), dtype=features.dtype) features = tf.reshape( features, [batch_size * num_boxes, output_size * 2, output_size * 2, num_filters]) features = tf.nn.avg_pool(features, [1, 2, 2, 1], [1, 2, 2, 1], 'VALID') features = tf.reshape( features, [batch_size, num_boxes, output_size, output_size, num_filters]) return features def compute_grid_positions(boxes, boundaries, output_size, sample_offset): """Compute the grid position w.r.t. the corresponding feature map. Args: boxes: a 3-D tensor of shape [batch_size, num_boxes, 4] encoding the information of each box w.r.t. the corresponding feature map. boxes[:, :, 0:2] are the grid position in (y, x) (float) of the top-left corner of each box. boxes[:, :, 2:4] are the box sizes in (h, w) (float) in terms of the number of pixels of the corresponding feature map size. boundaries: a 3-D tensor of shape [batch_size, num_boxes, 2] representing the boundary (in (y, x)) of the corresponding feature map for each box. Any resampled grid points that go beyond the bounary will be clipped. output_size: a scalar indicating the output crop size. sample_offset: a float number in [0, 1] indicates the subpixel sample offset from grid point. Returns: kernel_y: Tensor of size [batch_size, boxes, output_size, 2, 1]. kernel_x: Tensor of size [batch_size, boxes, output_size, 2, 1]. box_grid_y0y1: Tensor of size [batch_size, boxes, output_size, 2] box_grid_x0x1: Tensor of size [batch_size, boxes, output_size, 2] """ batch_size, num_boxes, _ = boxes.get_shape().as_list() box_grid_x = [] box_grid_y = [] for i in range(output_size): box_grid_x.append(boxes[:, :, 1] + (i + sample_offset) * boxes[:, :, 3] / output_size) box_grid_y.append(boxes[:, :, 0] + (i + sample_offset) * boxes[:, :, 2] / output_size) box_grid_x = tf.stack(box_grid_x, axis=2) box_grid_y = tf.stack(box_grid_y, axis=2) box_grid_y0 = tf.floor(box_grid_y) box_grid_x0 = tf.floor(box_grid_x) box_grid_x0 = tf.maximum(0., box_grid_x0) box_grid_y0 = tf.maximum(0., box_grid_y0) box_grid_x0 = tf.minimum(box_grid_x0, tf.expand_dims(boundaries[:, :, 1], -1)) box_grid_x1 = tf.minimum(box_grid_x0 + 1, tf.expand_dims(boundaries[:, :, 1], -1)) box_grid_y0 = tf.minimum(box_grid_y0, tf.expand_dims(boundaries[:, :, 0], -1)) box_grid_y1 = tf.minimum(box_grid_y0 + 1, tf.expand_dims(boundaries[:, :, 0], -1)) box_gridx0x1 = tf.stack([box_grid_x0, box_grid_x1], axis=-1) box_gridy0y1 = tf.stack([box_grid_y0, box_grid_y1], axis=-1) # The RoIAlign feature f can be computed by bilinear interpolation of four # neighboring feature points f0, f1, f2, and f3. # f(y, x) = [hy, ly] * [[f00, f01], * [hx, lx]^T # [f10, f11]] # f(y, x) = (hy*hx)f00 + (hy*lx)f01 + (ly*hx)f10 + (lx*ly)f11 # f(y, x) = w00*f00 + w01*f01 + w10*f10 + w11*f11 ly = box_grid_y - box_grid_y0 lx = box_grid_x - box_grid_x0 hy = 1.0 - ly hx = 1.0 - lx kernel_y = tf.reshape( tf.stack([hy, ly], axis=3), [batch_size, num_boxes, output_size, 2, 1]) kernel_x = tf.reshape( tf.stack([hx, lx], axis=3), [batch_size, num_boxes, output_size, 2, 1]) return kernel_y, kernel_x, box_gridy0y1, box_gridx0x1 def get_grid_one_hot(box_gridy0y1, box_gridx0x1, feature_height, feature_width): """Get grid_one_hot from indices and feature_size.""" (batch_size, num_boxes, output_size, _) = box_gridx0x1.get_shape().as_list() y_indices = tf.cast( tf.reshape(box_gridy0y1, [batch_size, num_boxes, output_size, 2]), dtype=tf.int32) x_indices = tf.cast( tf.reshape(box_gridx0x1, [batch_size, num_boxes, output_size, 2]), dtype=tf.int32) # shape is [batch_size, num_boxes, output_size, 2, height] grid_y_one_hot = tf.one_hot(tf.cast(y_indices, tf.int32), feature_height) # shape is [batch_size, num_boxes, output_size, 2, width] grid_x_one_hot = tf.one_hot(tf.cast(x_indices, tf.int32), feature_width) return grid_y_one_hot, grid_x_one_hot def selective_crop_and_resize(features, boxes, box_levels, boundaries, output_size=7, sample_offset=0.5, use_einsum_gather=False): """Crop and resize boxes on a set of feature maps. Given multiple features maps indexed by different levels, and a set of boxes where each box is mapped to a certain level, it selectively crops and resizes boxes from the corresponding feature maps to generate the box features. We follow the ROIAlign technique (see https://arxiv.org/pdf/1703.06870.pdf, figure 3 for reference). Specifically, for each feature map, we select an (output_size, output_size) set of pixels corresponding to the box location, and then use bilinear interpolation to select the feature value for each pixel. For performance, we perform the gather and interpolation on all layers as a single operation. In this op the multi-level features are first stacked and gathered into [2*output_size, 2*output_size] feature points. Then bilinear interpolation is performed on the gathered feature points to generate [output_size, output_size] RoIAlign feature map. Here is the step-by-step algorithm: 1. The multi-level features are gathered into a [batch_size, num_boxes, output_size*2, output_size*2, num_filters] Tensor. The Tensor contains four neighboring feature points for each vertice in the output grid. 2. Compute the interpolation kernel of shape [batch_size, num_boxes, output_size*2, output_size*2]. The last 2 axis can be seen as stacking 2x2 interpolation kernels for all vertices in the output grid. 3. Element-wise multiply the gathered features and interpolation kernel. Then apply 2x2 average pooling to reduce spatial dimension to output_size. Args: features: a 5-D tensor of shape [batch_size, num_levels, max_height, max_width, num_filters] where cropping and resizing are based. boxes: a 3-D tensor of shape [batch_size, num_boxes, 4] encoding the information of each box w.r.t. the corresponding feature map. boxes[:, :, 0:2] are the grid position in (y, x) (float) of the top-left corner of each box. boxes[:, :, 2:4] are the box sizes in (h, w) (float) in terms of the number of pixels of the corresponding feature map size. box_levels: a 3-D tensor of shape [batch_size, num_boxes, 1] representing the 0-based corresponding feature level index of each box. boundaries: a 3-D tensor of shape [batch_size, num_boxes, 2] representing the boundary (in (y, x)) of the corresponding feature map for each box. Any resampled grid points that go beyond the bounary will be clipped. output_size: a scalar indicating the output crop size. sample_offset: a float number in [0, 1] indicates the subpixel sample offset from grid point. use_einsum_gather: use einsum to replace gather or not. Replacing einsum with gather can improve performance when feature size is not large, einsum is friendly with model partition as well. Gather's performance is better when feature size is very large and there are multiple box levels. Returns: features_per_box: a 5-D tensor of shape [batch_size, num_boxes, output_size, output_size, num_filters] representing the cropped features. """ (batch_size, num_levels, max_feature_height, max_feature_width, num_filters) = features.get_shape().as_list() _, num_boxes, _ = boxes.get_shape().as_list() kernel_y, kernel_x, box_gridy0y1, box_gridx0x1 = compute_grid_positions( boxes, boundaries, output_size, sample_offset) x_indices = tf.cast( tf.reshape(box_gridx0x1, [batch_size, num_boxes, output_size * 2]), dtype=tf.int32) y_indices = tf.cast( tf.reshape(box_gridy0y1, [batch_size, num_boxes, output_size * 2]), dtype=tf.int32) if use_einsum_gather: # Blinear interpolation is done during the last two gathers: # f(y, x) = [hy, ly] * [[f00, f01], * [hx, lx]^T # [f10, f11]] # [[f00, f01], # [f10, f11]] = tf.einsum(tf.einsum(features, y_one_hot), x_one_hot) # where [hy, ly] and [hx, lx] are the bilinear interpolation kernel. # shape is [batch_size, boxes, output_size, 2, 1] grid_y_one_hot, grid_x_one_hot = get_grid_one_hot(box_gridy0y1, box_gridx0x1, max_feature_height, max_feature_width) # shape is [batch_size, num_boxes, output_size, height] grid_y_weight = tf.reduce_sum( tf.multiply(grid_y_one_hot, kernel_y), axis=-2) # shape is [batch_size, num_boxes, output_size, width] grid_x_weight = tf.reduce_sum( tf.multiply(grid_x_one_hot, kernel_x), axis=-2) # Gather for y_axis. # shape is [batch_size, num_boxes, output_size, width, features] features_per_box = tf.einsum('bmhwf,bmoh->bmowf', features, tf.cast(grid_y_weight, features.dtype)) # Gather for x_axis. # shape is [batch_size, num_boxes, output_size, output_size, features] features_per_box = tf.einsum('bmhwf,bmow->bmhof', features_per_box, tf.cast(grid_x_weight, features.dtype)) else: height_dim_offset = max_feature_width level_dim_offset = max_feature_height * height_dim_offset batch_dim_offset = num_levels * level_dim_offset batch_size_offset = tf.tile( tf.reshape( tf.range(batch_size) * batch_dim_offset, [batch_size, 1, 1, 1]), [1, num_boxes, output_size * 2, output_size * 2]) box_levels_offset = tf.tile( tf.reshape(box_levels * level_dim_offset, [batch_size, num_boxes, 1, 1]), [1, 1, output_size * 2, output_size * 2]) y_indices_offset = tf.tile( tf.reshape(y_indices * height_dim_offset, [batch_size, num_boxes, output_size * 2, 1]), [1, 1, 1, output_size * 2]) x_indices_offset = tf.tile( tf.reshape(x_indices, [batch_size, num_boxes, 1, output_size * 2]), [1, 1, output_size * 2, 1]) indices = tf.reshape( batch_size_offset + box_levels_offset + y_indices_offset + x_indices_offset, [-1]) features = tf.reshape(features, [-1, num_filters]) # TODO(wangtao): replace tf.gather with tf.gather_nd and try to get similar # performance. features_per_box = tf.reshape( tf.gather(features, indices), [batch_size, num_boxes, output_size * 2, output_size * 2, num_filters]) features_per_box = feature_bilinear_interpolation(features_per_box, kernel_y, kernel_x) return features_per_box def multilevel_crop_and_resize(features, boxes, output_size=7): """Crop and resize on multilevel feature pyramid. Generate the (output_size, output_size) set of pixels for each input box by first locating the box into the correct feature level, and then cropping and resizing it using the correspoding feature map of that level. Args: features: A dictionary with key as pyramid level and value as features. The features are in shape of [batch_size, height_l, width_l, num_filters]. boxes: A 3-D Tensor of shape [batch_size, num_boxes, 4]. Each row represents a box with [y1, x1, y2, x2] in un-normalized coordinates. output_size: A scalar to indicate the output crop size. Returns: A 5-D tensor representing feature crop of shape [batch_size, num_boxes, output_size, output_size, num_filters]. """ with tf.name_scope('multilevel_crop_and_resize'): levels = list(features.keys()) min_level = min(levels) max_level = max(levels) batch_size, max_feature_height, max_feature_width, num_filters = ( features[min_level].get_shape().as_list()) _, num_boxes, _ = boxes.get_shape().as_list() # Stack feature pyramid into a features_all of shape # [batch_size, levels, height, width, num_filters]. features_all = [] feature_heights = [] feature_widths = [] for level in range(min_level, max_level + 1): shape = features[level].get_shape().as_list() feature_heights.append(shape[1]) feature_widths.append(shape[2]) # Concat tensor of [batch_size, height_l * width_l, num_filters] for each # levels. features_all.append( tf.reshape(features[level], [batch_size, -1, num_filters])) features_r2 = tf.reshape(tf.concat(features_all, 1), [-1, num_filters]) # Calculate height_l * width_l for each level. level_dim_sizes = [ feature_widths[i] * feature_heights[i] for i in range(len(feature_widths)) ] # level_dim_offsets is accumulated sum of level_dim_size. level_dim_offsets = [0] for i in range(len(feature_widths) - 1): level_dim_offsets.append(level_dim_offsets[i] + level_dim_sizes[i]) batch_dim_size = level_dim_offsets[-1] + level_dim_sizes[-1] level_dim_offsets = tf.constant(level_dim_offsets, tf.int32) height_dim_sizes = tf.constant(feature_widths, tf.int32) # Assigns boxes to the right level. box_width = boxes[:, :, 3] - boxes[:, :, 1] box_height = boxes[:, :, 2] - boxes[:, :, 0] areas_sqrt = tf.sqrt(box_height * box_width) levels = tf.cast( tf.math.floordiv( tf.math.log(tf.divide(areas_sqrt, 224.0)), tf.math.log(2.0)) + 4.0, dtype=tf.int32) # Maps levels between [min_level, max_level]. levels = tf.minimum(max_level, tf.maximum(levels, min_level)) # Projects box location and sizes to corresponding feature levels. scale_to_level = tf.cast( tf.pow(tf.constant(2.0), tf.cast(levels, tf.float32)), dtype=boxes.dtype) boxes /= tf.expand_dims(scale_to_level, axis=2) box_width /= scale_to_level box_height /= scale_to_level boxes = tf.concat([ boxes[:, :, 0:2], tf.expand_dims(box_height, -1), tf.expand_dims(box_width, -1) ], axis=-1) # Maps levels to [0, max_level-min_level]. levels -= min_level level_strides = tf.pow([[2.0]], tf.cast(levels, tf.float32)) boundary = tf.cast( tf.concat([ tf.expand_dims( [[tf.cast(max_feature_height, tf.float32)]] / level_strides - 1, axis=-1), tf.expand_dims( [[tf.cast(max_feature_width, tf.float32)]] / level_strides - 1, axis=-1), ], axis=-1), boxes.dtype) # Compute grid positions. kernel_y, kernel_x, box_gridy0y1, box_gridx0x1 = compute_grid_positions( boxes, boundary, output_size, sample_offset=0.5) x_indices = tf.cast( tf.reshape(box_gridx0x1, [batch_size, num_boxes, output_size * 2]), dtype=tf.int32) y_indices = tf.cast( tf.reshape(box_gridy0y1, [batch_size, num_boxes, output_size * 2]), dtype=tf.int32) batch_size_offset = tf.tile( tf.reshape( tf.range(batch_size) * batch_dim_size, [batch_size, 1, 1, 1]), [1, num_boxes, output_size * 2, output_size * 2]) # Get level offset for each box. Each box belongs to one level. levels_offset = tf.tile( tf.reshape( tf.gather(level_dim_offsets, levels), [batch_size, num_boxes, 1, 1]), [1, 1, output_size * 2, output_size * 2]) y_indices_offset = tf.tile( tf.reshape( y_indices * tf.expand_dims(tf.gather(height_dim_sizes, levels), -1), [batch_size, num_boxes, output_size * 2, 1]), [1, 1, 1, output_size * 2]) x_indices_offset = tf.tile( tf.reshape(x_indices, [batch_size, num_boxes, 1, output_size * 2]), [1, 1, output_size * 2, 1]) indices = tf.reshape( batch_size_offset + levels_offset + y_indices_offset + x_indices_offset, [-1]) # TODO(wangtao): replace tf.gather with tf.gather_nd and try to get similar # performance. features_per_box = tf.reshape( tf.gather(features_r2, indices), [batch_size, num_boxes, output_size * 2, output_size * 2, num_filters]) # Bilinear interpolation. features_per_box = feature_bilinear_interpolation(features_per_box, kernel_y, kernel_x) return features_per_box def single_level_feature_crop(features, level_boxes, detection_prior_levels, min_mask_level, mask_crop_size): """Crop the FPN features at the appropriate levels for each detection. Args: features: a float tensor of shape [batch_size, num_levels, max_feature_size, max_feature_size, num_downsample_channels]. level_boxes: a float Tensor of the level boxes to crop from. [batch_size, num_instances, 4]. detection_prior_levels: an int Tensor of instance assigned level of shape [batch_size, num_instances]. min_mask_level: minimum FPN level to crop mask feature from. mask_crop_size: an int of mask crop size. Returns: crop_features: a float Tensor of shape [batch_size * num_instances, mask_crop_size, mask_crop_size, num_downsample_channels]. This is the instance feature crop. """ (batch_size, num_levels, max_feature_size, _, num_downsample_channels) = features.get_shape().as_list() _, num_of_instances, _ = level_boxes.get_shape().as_list() level_boxes = tf.cast(level_boxes, tf.int32) assert num_of_instances == detection_prior_levels.get_shape().as_list()[1] x_start_indices = level_boxes[:, :, 1] y_start_indices = level_boxes[:, :, 0] # generate the full indices (not just the starting index) x_idx_list = [] y_idx_list = [] for i in range(mask_crop_size): x_idx_list.append(x_start_indices + i) y_idx_list.append(y_start_indices + i) x_indices = tf.stack(x_idx_list, axis=2) y_indices = tf.stack(y_idx_list, axis=2) levels = detection_prior_levels - min_mask_level height_dim_size = max_feature_size level_dim_size = max_feature_size * height_dim_size batch_dim_size = num_levels * level_dim_size # TODO(weicheng) change this to gather_nd for better readability. indices = tf.reshape( tf.tile( tf.reshape( tf.range(batch_size) * batch_dim_size, [batch_size, 1, 1, 1]), [1, num_of_instances, mask_crop_size, mask_crop_size]) + tf.tile( tf.reshape(levels * level_dim_size, [batch_size, num_of_instances, 1, 1]), [1, 1, mask_crop_size, mask_crop_size]) + tf.tile( tf.reshape(y_indices * height_dim_size, [batch_size, num_of_instances, mask_crop_size, 1]), [1, 1, 1, mask_crop_size]) + tf.tile( tf.reshape(x_indices, [batch_size, num_of_instances, 1, mask_crop_size]), [1, 1, mask_crop_size, 1]), [-1]) features_r2 = tf.reshape(features, [-1, num_downsample_channels]) crop_features = tf.reshape( tf.gather(features_r2, indices), [ batch_size * num_of_instances, mask_crop_size, mask_crop_size, num_downsample_channels ]) return crop_features def crop_mask_in_target_box(masks, boxes, target_boxes, output_size, sample_offset=0, use_einsum=True): """Crop masks in target boxes. Args: masks: A tensor with a shape of [batch_size, num_masks, height, width]. boxes: a float tensor representing box cooridnates that tightly enclose masks with a shape of [batch_size, num_masks, 4] in un-normalized coordinates. A box is represented by [ymin, xmin, ymax, xmax]. target_boxes: a float tensor representing target box cooridnates for masks with a shape of [batch_size, num_masks, 4] in un-normalized coordinates. A box is represented by [ymin, xmin, ymax, xmax]. output_size: A scalar to indicate the output crop size. It currently only supports to output a square shape outputs. sample_offset: a float number in [0, 1] indicates the subpixel sample offset from grid point. use_einsum: Use einsum to replace gather in selective_crop_and_resize. Returns: A 4-D tensor representing feature crop of shape [batch_size, num_boxes, output_size, output_size]. """ with tf.name_scope('crop_mask_in_target_box'): batch_size, num_masks, height, width = masks.get_shape().as_list() masks = tf.reshape(masks, [batch_size * num_masks, height, width, 1]) # Pad zeros on the boundary of masks. masks = tf.image.pad_to_bounding_box(masks, 2, 2, height + 4, width + 4) masks = tf.reshape(masks, [batch_size, num_masks, height + 4, width + 4, 1]) # Projects target box locations and sizes to corresponding cropped # mask coordinates. gt_y_min, gt_x_min, gt_y_max, gt_x_max = tf.split( value=boxes, num_or_size_splits=4, axis=2) bb_y_min, bb_x_min, bb_y_max, bb_x_max = tf.split( value=target_boxes, num_or_size_splits=4, axis=2) y_transform = (bb_y_min - gt_y_min) * height / (gt_y_max - gt_y_min + _EPSILON) + 2 x_transform = (bb_x_min - gt_x_min) * height / (gt_x_max - gt_x_min + _EPSILON) + 2 h_transform = (bb_y_max - bb_y_min) * width / ( gt_y_max - gt_y_min + _EPSILON) w_transform = (bb_x_max - bb_x_min) * width / ( gt_x_max - gt_x_min + _EPSILON) boundaries = tf.concat([ tf.cast( tf.ones_like(y_transform) * ((height + 4) - 1), dtype=tf.float32), tf.cast( tf.ones_like(x_transform) * ((width + 4) - 1), dtype=tf.float32) ], axis=-1) # Reshape tensors to have the right shape for selective_crop_and_resize. trasnformed_boxes = tf.concat( [y_transform, x_transform, h_transform, w_transform], -1) levels = tf.tile( tf.reshape(tf.range(num_masks), [1, num_masks]), [batch_size, 1]) cropped_masks = selective_crop_and_resize( masks, trasnformed_boxes, levels, boundaries, output_size, sample_offset=sample_offset, use_einsum_gather=use_einsum) cropped_masks = tf.squeeze(cropped_masks, axis=-1) return cropped_masks
25,998
42.044702
80
py
models
models-master/official/legacy/detection/dataloader/retinanet_parser.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Data parser and processing. Parse image and ground truths in a dataset to training targets and package them into (image, labels) tuple for RetinaNet. T.-Y. Lin, P. Goyal, R. Girshick, K. He, and P. Dollar Focal Loss for Dense Object Detection. arXiv:1708.02002 """ import tensorflow as tf from official.legacy.detection.dataloader import anchor from official.legacy.detection.dataloader import mode_keys as ModeKeys from official.legacy.detection.dataloader import tf_example_decoder from official.legacy.detection.utils import box_utils from official.legacy.detection.utils import input_utils def process_source_id(source_id): """Processes source_id to the right format.""" if source_id.dtype == tf.string: source_id = tf.cast(tf.strings.to_number(source_id), tf.int32) with tf.control_dependencies([source_id]): source_id = tf.cond( pred=tf.equal(tf.size(input=source_id), 0), true_fn=lambda: tf.cast(tf.constant(-1), tf.int32), false_fn=lambda: tf.identity(source_id)) return source_id def pad_groundtruths_to_fixed_size(gt, n): """Pads the first dimension of groundtruths labels to the fixed size.""" gt['boxes'] = input_utils.pad_to_fixed_size(gt['boxes'], n, -1) gt['is_crowds'] = input_utils.pad_to_fixed_size(gt['is_crowds'], n, 0) gt['areas'] = input_utils.pad_to_fixed_size(gt['areas'], n, -1) gt['classes'] = input_utils.pad_to_fixed_size(gt['classes'], n, -1) return gt class Parser(object): """Parser to parse an image and its annotations into a dictionary of tensors.""" def __init__(self, output_size, min_level, max_level, num_scales, aspect_ratios, anchor_size, match_threshold=0.5, unmatched_threshold=0.5, aug_rand_hflip=False, aug_scale_min=1.0, aug_scale_max=1.0, use_autoaugment=False, autoaugment_policy_name='v0', skip_crowd_during_training=True, max_num_instances=100, use_bfloat16=True, mode=None): """Initializes parameters for parsing annotations in the dataset. Args: output_size: `Tensor` or `list` for [height, width] of output image. The output_size should be divided by the largest feature stride 2^max_level. min_level: `int` number of minimum level of the output feature pyramid. max_level: `int` number of maximum level of the output feature pyramid. num_scales: `int` number representing intermediate scales added on each level. For instances, num_scales=2 adds one additional intermediate anchor scales [2^0, 2^0.5] on each level. aspect_ratios: `list` of float numbers representing the aspect raito anchors added on each level. The number indicates the ratio of width to height. For instances, aspect_ratios=[1.0, 2.0, 0.5] adds three anchors on each scale level. anchor_size: `float` number representing the scale of size of the base anchor to the feature stride 2^level. match_threshold: `float` number between 0 and 1 representing the lower-bound threshold to assign positive labels for anchors. An anchor with a score over the threshold is labeled positive. unmatched_threshold: `float` number between 0 and 1 representing the upper-bound threshold to assign negative labels for anchors. An anchor with a score below the threshold is labeled negative. aug_rand_hflip: `bool`, if True, augment training with random horizontal flip. aug_scale_min: `float`, the minimum scale applied to `output_size` for data augmentation during training. aug_scale_max: `float`, the maximum scale applied to `output_size` for data augmentation during training. use_autoaugment: `bool`, if True, use the AutoAugment augmentation policy during training. autoaugment_policy_name: `string` that specifies the name of the AutoAugment policy that will be used during training. skip_crowd_during_training: `bool`, if True, skip annotations labeled with `is_crowd` equals to 1. max_num_instances: `int` number of maximum number of instances in an image. The groundtruth data will be padded to `max_num_instances`. use_bfloat16: `bool`, if True, cast output image to tf.bfloat16. mode: a ModeKeys. Specifies if this is training, evaluation, prediction or prediction with groundtruths in the outputs. """ self._mode = mode self._max_num_instances = max_num_instances self._skip_crowd_during_training = skip_crowd_during_training self._is_training = (mode == ModeKeys.TRAIN) self._example_decoder = tf_example_decoder.TfExampleDecoder( include_mask=False) # Anchor. self._output_size = output_size self._min_level = min_level self._max_level = max_level self._num_scales = num_scales self._aspect_ratios = aspect_ratios self._anchor_size = anchor_size self._match_threshold = match_threshold self._unmatched_threshold = unmatched_threshold # Data augmentation. self._aug_rand_hflip = aug_rand_hflip self._aug_scale_min = aug_scale_min self._aug_scale_max = aug_scale_max # Data Augmentation with AutoAugment. self._use_autoaugment = use_autoaugment self._autoaugment_policy_name = autoaugment_policy_name # Device. self._use_bfloat16 = use_bfloat16 # Data is parsed depending on the model Modekey. if mode == ModeKeys.TRAIN: self._parse_fn = self._parse_train_data elif mode == ModeKeys.EVAL: self._parse_fn = self._parse_eval_data elif mode == ModeKeys.PREDICT or mode == ModeKeys.PREDICT_WITH_GT: self._parse_fn = self._parse_predict_data else: raise ValueError('mode is not defined.') def __call__(self, value): """Parses data to an image and associated training labels. Args: value: a string tensor holding a serialized tf.Example proto. Returns: image: image tensor that is preproessed to have normalized value and dimension [output_size[0], output_size[1], 3] labels: cls_targets: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, anchors_per_location]. The height_l and width_l represent the dimension of class logits at l-th level. box_targets: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, anchors_per_location * 4]. The height_l and width_l represent the dimension of bounding box regression output at l-th level. num_positives: number of positive anchors in the image. anchor_boxes: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, 4] representing anchor boxes at each level. image_info: a 2D `Tensor` that encodes the information of the image and the applied preprocessing. It is in the format of [[original_height, original_width], [scaled_height, scaled_width], [y_scale, x_scale], [y_offset, x_offset]]. groundtruths: source_id: source image id. Default value -1 if the source id is empty in the groundtruth annotation. boxes: groundtruth bounding box annotations. The box is represented in [y1, x1, y2, x2] format. The tennsor is padded with -1 to the fixed dimension [self._max_num_instances, 4]. classes: groundtruth classes annotations. The tennsor is padded with -1 to the fixed dimension [self._max_num_instances]. areas: groundtruth areas annotations. The tennsor is padded with -1 to the fixed dimension [self._max_num_instances]. is_crowds: groundtruth annotations to indicate if an annotation represents a group of instances by value {0, 1}. The tennsor is padded with 0 to the fixed dimension [self._max_num_instances]. """ with tf.name_scope('parser'): data = self._example_decoder.decode(value) return self._parse_fn(data) def _parse_train_data(self, data): """Parses data for training and evaluation.""" classes = data['groundtruth_classes'] boxes = data['groundtruth_boxes'] is_crowds = data['groundtruth_is_crowd'] # Skips annotations with `is_crowd` = True. if self._skip_crowd_during_training and self._is_training: num_groundtrtuhs = tf.shape(input=classes)[0] with tf.control_dependencies([num_groundtrtuhs, is_crowds]): indices = tf.cond( pred=tf.greater(tf.size(input=is_crowds), 0), true_fn=lambda: tf.where(tf.logical_not(is_crowds))[:, 0], false_fn=lambda: tf.cast(tf.range(num_groundtrtuhs), tf.int64)) classes = tf.gather(classes, indices) boxes = tf.gather(boxes, indices) # Gets original image and its size. image = data['image'] image_shape = tf.shape(input=image)[0:2] # Normalizes image with mean and std pixel values. image = input_utils.normalize_image(image) # Flips image randomly during training. if self._aug_rand_hflip: image, boxes = input_utils.random_horizontal_flip(image, boxes) # Converts boxes from normalized coordinates to pixel coordinates. boxes = box_utils.denormalize_boxes(boxes, image_shape) # Resizes and crops image. image, image_info = input_utils.resize_and_crop_image( image, self._output_size, padded_size=input_utils.compute_padded_size(self._output_size, 2**self._max_level), aug_scale_min=self._aug_scale_min, aug_scale_max=self._aug_scale_max) image_height, image_width, _ = image.get_shape().as_list() # Resizes and crops boxes. image_scale = image_info[2, :] offset = image_info[3, :] boxes = input_utils.resize_and_crop_boxes(boxes, image_scale, image_info[1, :], offset) # Filters out ground truth boxes that are all zeros. indices = box_utils.get_non_empty_box_indices(boxes) boxes = tf.gather(boxes, indices) classes = tf.gather(classes, indices) # Assigns anchors. input_anchor = anchor.Anchor(self._min_level, self._max_level, self._num_scales, self._aspect_ratios, self._anchor_size, (image_height, image_width)) anchor_labeler = anchor.AnchorLabeler(input_anchor, self._match_threshold, self._unmatched_threshold) (cls_targets, box_targets, num_positives) = anchor_labeler.label_anchors( boxes, tf.cast(tf.expand_dims(classes, axis=1), tf.float32)) # If bfloat16 is used, casts input image to tf.bfloat16. if self._use_bfloat16: image = tf.cast(image, dtype=tf.bfloat16) # Packs labels for model_fn outputs. labels = { 'cls_targets': cls_targets, 'box_targets': box_targets, 'anchor_boxes': input_anchor.multilevel_boxes, 'num_positives': num_positives, 'image_info': image_info, } return image, labels def _parse_eval_data(self, data): """Parses data for training and evaluation.""" groundtruths = {} classes = data['groundtruth_classes'] boxes = data['groundtruth_boxes'] # Gets original image and its size. image = data['image'] image_shape = tf.shape(input=image)[0:2] # Normalizes image with mean and std pixel values. image = input_utils.normalize_image(image) # Converts boxes from normalized coordinates to pixel coordinates. boxes = box_utils.denormalize_boxes(boxes, image_shape) # Resizes and crops image. image, image_info = input_utils.resize_and_crop_image( image, self._output_size, padded_size=input_utils.compute_padded_size(self._output_size, 2**self._max_level), aug_scale_min=1.0, aug_scale_max=1.0) image_height, image_width, _ = image.get_shape().as_list() # Resizes and crops boxes. image_scale = image_info[2, :] offset = image_info[3, :] boxes = input_utils.resize_and_crop_boxes(boxes, image_scale, image_info[1, :], offset) # Filters out ground truth boxes that are all zeros. indices = box_utils.get_non_empty_box_indices(boxes) boxes = tf.gather(boxes, indices) classes = tf.gather(classes, indices) # Assigns anchors. input_anchor = anchor.Anchor(self._min_level, self._max_level, self._num_scales, self._aspect_ratios, self._anchor_size, (image_height, image_width)) anchor_labeler = anchor.AnchorLabeler(input_anchor, self._match_threshold, self._unmatched_threshold) (cls_targets, box_targets, num_positives) = anchor_labeler.label_anchors( boxes, tf.cast(tf.expand_dims(classes, axis=1), tf.float32)) # If bfloat16 is used, casts input image to tf.bfloat16. if self._use_bfloat16: image = tf.cast(image, dtype=tf.bfloat16) # Sets up groundtruth data for evaluation. groundtruths = { 'source_id': data['source_id'], 'num_groundtrtuhs': tf.shape(data['groundtruth_classes']), 'image_info': image_info, 'boxes': box_utils.denormalize_boxes(data['groundtruth_boxes'], image_shape), 'classes': data['groundtruth_classes'], 'areas': data['groundtruth_area'], 'is_crowds': tf.cast(data['groundtruth_is_crowd'], tf.int32), } groundtruths['source_id'] = process_source_id(groundtruths['source_id']) groundtruths = pad_groundtruths_to_fixed_size(groundtruths, self._max_num_instances) # Packs labels for model_fn outputs. labels = { 'cls_targets': cls_targets, 'box_targets': box_targets, 'anchor_boxes': input_anchor.multilevel_boxes, 'num_positives': num_positives, 'image_info': image_info, 'groundtruths': groundtruths, } return image, labels def _parse_predict_data(self, data): """Parses data for prediction.""" # Gets original image and its size. image = data['image'] image_shape = tf.shape(input=image)[0:2] # Normalizes image with mean and std pixel values. image = input_utils.normalize_image(image) # Resizes and crops image. image, image_info = input_utils.resize_and_crop_image( image, self._output_size, padded_size=input_utils.compute_padded_size(self._output_size, 2**self._max_level), aug_scale_min=1.0, aug_scale_max=1.0) image_height, image_width, _ = image.get_shape().as_list() # If bfloat16 is used, casts input image to tf.bfloat16. if self._use_bfloat16: image = tf.cast(image, dtype=tf.bfloat16) # Compute Anchor boxes. input_anchor = anchor.Anchor(self._min_level, self._max_level, self._num_scales, self._aspect_ratios, self._anchor_size, (image_height, image_width)) labels = { 'anchor_boxes': input_anchor.multilevel_boxes, 'image_info': image_info, } # If mode is PREDICT_WITH_GT, returns groundtruths and training targets # in labels. if self._mode == ModeKeys.PREDICT_WITH_GT: # Converts boxes from normalized coordinates to pixel coordinates. boxes = box_utils.denormalize_boxes(data['groundtruth_boxes'], image_shape) groundtruths = { 'source_id': data['source_id'], 'num_detections': tf.shape(data['groundtruth_classes']), 'boxes': boxes, 'classes': data['groundtruth_classes'], 'areas': data['groundtruth_area'], 'is_crowds': tf.cast(data['groundtruth_is_crowd'], tf.int32), } groundtruths['source_id'] = process_source_id(groundtruths['source_id']) groundtruths = pad_groundtruths_to_fixed_size(groundtruths, self._max_num_instances) labels['groundtruths'] = groundtruths # Computes training objective for evaluation loss. classes = data['groundtruth_classes'] image_scale = image_info[2, :] offset = image_info[3, :] boxes = input_utils.resize_and_crop_boxes(boxes, image_scale, image_info[1, :], offset) # Filters out ground truth boxes that are all zeros. indices = box_utils.get_non_empty_box_indices(boxes) boxes = tf.gather(boxes, indices) # Assigns anchors. anchor_labeler = anchor.AnchorLabeler(input_anchor, self._match_threshold, self._unmatched_threshold) (cls_targets, box_targets, num_positives) = anchor_labeler.label_anchors( boxes, tf.cast(tf.expand_dims(classes, axis=1), tf.float32)) labels['cls_targets'] = cls_targets labels['box_targets'] = box_targets labels['num_positives'] = num_positives return image, labels
18,283
41.920188
82
py
models
models-master/official/legacy/detection/dataloader/shapemask_parser.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Data parser and processing. Parse image and ground truths in a dataset to training targets and package them into (image, labels) tuple for ShapeMask. Weicheng Kuo, Anelia Angelova, Jitendra Malik, Tsung-Yi Lin ShapeMask: Learning to Segment Novel Objects by Refining Shape Priors. arXiv:1904.03239. """ import tensorflow as tf from official.legacy.detection.dataloader import anchor from official.legacy.detection.dataloader import mode_keys as ModeKeys from official.legacy.detection.dataloader import tf_example_decoder from official.legacy.detection.utils import box_utils from official.legacy.detection.utils import class_utils from official.legacy.detection.utils import dataloader_utils from official.legacy.detection.utils import input_utils def pad_to_size(input_tensor, size): """Pads data with zeros to a given length at the first dimension if needed. Args: input_tensor: `Tensor` with any dimension. size: `int` number for the first dimension of output Tensor. Returns: `Tensor` with the first dimension padded to `size` if the first diemsion is less than `size`, otherwise no padding. """ input_shape = tf.shape(input_tensor) padding_shape = [] # Computes the padding length on the first dimension. padding_length = tf.maximum(0, size - tf.shape(input_tensor)[0]) assert_length = tf.Assert( tf.greater_equal(padding_length, 0), [padding_length]) with tf.control_dependencies([assert_length]): padding_shape.append(padding_length) # Copies shapes of the rest of input shape dimensions. for i in range(1, len(input_shape)): padding_shape.append(tf.shape(input=input_tensor)[i]) # Pads input tensor to the fixed first dimension. paddings = tf.cast(tf.zeros(padding_shape), input_tensor.dtype) padded_tensor = tf.concat([input_tensor, paddings], axis=0) return padded_tensor class Parser(object): """ShapeMask Parser to parse an image and its annotations into a dictionary of tensors.""" def __init__(self, output_size, min_level, max_level, num_scales, aspect_ratios, anchor_size, use_category=True, outer_box_scale=1.0, box_jitter_scale=0.025, num_sampled_masks=8, mask_crop_size=32, mask_min_level=3, mask_max_level=5, upsample_factor=4, match_threshold=0.5, unmatched_threshold=0.5, aug_rand_hflip=False, aug_scale_min=1.0, aug_scale_max=1.0, skip_crowd_during_training=True, max_num_instances=100, use_bfloat16=True, mask_train_class='all', mode=None): """Initializes parameters for parsing annotations in the dataset. Args: output_size: `Tensor` or `list` for [height, width] of output image. The output_size should be divided by the largest feature stride 2^max_level. min_level: `int` number of minimum level of the output feature pyramid. max_level: `int` number of maximum level of the output feature pyramid. num_scales: `int` number representing intermediate scales added on each level. For instances, num_scales=2 adds one additional intermediate anchor scales [2^0, 2^0.5] on each level. aspect_ratios: `list` of float numbers representing the aspect raito anchors added on each level. The number indicates the ratio of width to height. For instances, aspect_ratios=[1.0, 2.0, 0.5] adds three anchors on each scale level. anchor_size: `float` number representing the scale of size of the base anchor to the feature stride 2^level. use_category: if `False`, treat all object in all classes in one foreground category. outer_box_scale: `float` number in a range of [1.0, inf) representing the scale from object box to outer box. The mask branch predicts instance mask enclosed in outer box. box_jitter_scale: `float` number representing the noise magnitude to jitter the training groundtruth boxes for mask branch. num_sampled_masks: `int` number of sampled masks for training. mask_crop_size: `list` for [height, width] of output training masks. mask_min_level: `int` number indicating the minimum feature level to obtain instance features. mask_max_level: `int` number indicating the maximum feature level to obtain instance features. upsample_factor: `int` factor of upsampling the fine mask predictions. match_threshold: `float` number between 0 and 1 representing the lower-bound threshold to assign positive labels for anchors. An anchor with a score over the threshold is labeled positive. unmatched_threshold: `float` number between 0 and 1 representing the upper-bound threshold to assign negative labels for anchors. An anchor with a score below the threshold is labeled negative. aug_rand_hflip: `bool`, if True, augment training with random horizontal flip. aug_scale_min: `float`, the minimum scale applied to `output_size` for data augmentation during training. aug_scale_max: `float`, the maximum scale applied to `output_size` for data augmentation during training. skip_crowd_during_training: `bool`, if True, skip annotations labeled with `is_crowd` equals to 1. max_num_instances: `int` number of maximum number of instances in an image. The groundtruth data will be padded to `max_num_instances`. use_bfloat16: `bool`, if True, cast output image to tf.bfloat16. mask_train_class: a string of experiment mode: `all`, `voc` or `nonvoc`. mode: a ModeKeys. Specifies if this is training, evaluation, prediction or prediction with groundtruths in the outputs. """ self._mode = mode self._mask_train_class = mask_train_class self._max_num_instances = max_num_instances self._skip_crowd_during_training = skip_crowd_during_training self._is_training = (mode == ModeKeys.TRAIN) self._example_decoder = tf_example_decoder.TfExampleDecoder( include_mask=True) # Anchor. self._output_size = output_size self._min_level = min_level self._max_level = max_level self._num_scales = num_scales self._aspect_ratios = aspect_ratios self._anchor_size = anchor_size self._match_threshold = match_threshold self._unmatched_threshold = unmatched_threshold # Data augmentation. self._aug_rand_hflip = aug_rand_hflip self._aug_scale_min = aug_scale_min self._aug_scale_max = aug_scale_max # Device. self._use_bfloat16 = use_bfloat16 # ShapeMask specific. # Control of which category to use. self._use_category = use_category self._num_sampled_masks = num_sampled_masks self._mask_crop_size = mask_crop_size self._mask_min_level = mask_min_level self._mask_max_level = mask_max_level self._outer_box_scale = outer_box_scale self._box_jitter_scale = box_jitter_scale self._up_sample_factor = upsample_factor # Data is parsed depending on the model Modekey. if mode == ModeKeys.TRAIN: self._parse_fn = self._parse_train_data elif mode == ModeKeys.EVAL: self._parse_fn = self._parse_eval_data elif mode == ModeKeys.PREDICT or mode == ModeKeys.PREDICT_WITH_GT: self._parse_fn = self._parse_predict_data else: raise ValueError('mode is not defined.') def __call__(self, value): """Parses data to an image and associated training labels. Args: value: a string tensor holding a serialized tf.Example proto. Returns: inputs: image: image tensor that is preproessed to have normalized value and dimension [output_size[0], output_size[1], 3] mask_boxes: sampled boxes that tightly enclose the training masks. The box is represented in [y1, x1, y2, x2] format. The tensor is sampled to the fixed dimension [self._num_sampled_masks, 4]. mask_outer_boxes: loose box that enclose sampled tight box. The box is represented in [y1, x1, y2, x2] format. The tensor is sampled to the fixed dimension [self._num_sampled_masks, 4]. mask_classes: the class ids of sampled training masks. The tensor has shape [self._num_sampled_masks]. labels: cls_targets: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, anchors_per_location]. The height_l and width_l represent the dimension of class logits at l-th level. box_targets: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, anchors_per_location * 4]. The height_l and width_l represent the dimension of bounding box regression output at l-th level. num_positives: number of positive anchors in the image. anchor_boxes: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, 4] representing anchor boxes at each level. image_scale: 2D float `Tensor` representing scale factors that apply to [height, width] of input image. mask_targets: training binary mask targets. The tensor has shape [self._num_sampled_masks, self._mask_crop_size, self._mask_crop_size]. mask_is_valid: the binary tensor to indicate if the sampled masks are valide. The sampled masks are invalid when no mask annotations are included in the image. The tensor has shape [1]. groundtruths: source_id: source image id. Default value -1 if the source id is empty in the groundtruth annotation. boxes: groundtruth bounding box annotations. The box is represented in [y1, x1, y2, x2] format. The tensor is padded with -1 to the fixed dimension [self._max_num_instances, 4]. classes: groundtruth classes annotations. The tensor is padded with -1 to the fixed dimension [self._max_num_instances]. areas: groundtruth areas annotations. The tensor is padded with -1 to the fixed dimension [self._max_num_instances]. is_crowds: groundtruth annotations to indicate if an annotation represents a group of instances by value {0, 1}. The tensor is padded with 0 to the fixed dimension [self._max_num_instances]. """ with tf.name_scope('parser'): data = self._example_decoder.decode(value) return self._parse_fn(data) def _parse_train_data(self, data): """Parse data for ShapeMask training.""" classes = data['groundtruth_classes'] boxes = data['groundtruth_boxes'] masks = data['groundtruth_instance_masks'] is_crowds = data['groundtruth_is_crowd'] # Skips annotations with `is_crowd` = True. if self._skip_crowd_during_training and self._is_training: num_groundtrtuhs = tf.shape(classes)[0] with tf.control_dependencies([num_groundtrtuhs, is_crowds]): indices = tf.cond( tf.greater(tf.size(is_crowds), 0), lambda: tf.where(tf.logical_not(is_crowds))[:, 0], lambda: tf.cast(tf.range(num_groundtrtuhs), tf.int64)) classes = tf.gather(classes, indices) boxes = tf.gather(boxes, indices) masks = tf.gather(masks, indices) # Gets original image and its size. image = data['image'] image_shape = tf.shape(image)[0:2] # If not using category, makes all categories with id = 0. if not self._use_category: classes = tf.cast(tf.greater(classes, 0), dtype=tf.float32) # Normalizes image with mean and std pixel values. image = input_utils.normalize_image(image) # Flips image randomly during training. if self._aug_rand_hflip: image, boxes, masks = input_utils.random_horizontal_flip( image, boxes, masks) # Converts boxes from normalized coordinates to pixel coordinates. boxes = box_utils.denormalize_boxes(boxes, image_shape) # Resizes and crops image. image, image_info = input_utils.resize_and_crop_image( image, self._output_size, self._output_size, aug_scale_min=self._aug_scale_min, aug_scale_max=self._aug_scale_max) image_scale = image_info[2, :] offset = image_info[3, :] # Resizes and crops boxes and masks. boxes = input_utils.resize_and_crop_boxes( boxes, image_scale, image_info[1, :], offset) # Filters out ground truth boxes that are all zeros. indices = box_utils.get_non_empty_box_indices(boxes) boxes = tf.gather(boxes, indices) classes = tf.gather(classes, indices) masks = tf.gather(masks, indices) # Assigns anchors. input_anchor = anchor.Anchor( self._min_level, self._max_level, self._num_scales, self._aspect_ratios, self._anchor_size, self._output_size) anchor_labeler = anchor.AnchorLabeler( input_anchor, self._match_threshold, self._unmatched_threshold) (cls_targets, box_targets, num_positives) = anchor_labeler.label_anchors( boxes, tf.cast(tf.expand_dims(classes, axis=1), tf.float32)) # Sample groundtruth masks/boxes/classes for mask branch. num_masks = tf.shape(masks)[0] mask_shape = tf.shape(masks)[1:3] # Pad sampled boxes/masks/classes to a constant batch size. padded_boxes = pad_to_size(boxes, self._num_sampled_masks) padded_classes = pad_to_size(classes, self._num_sampled_masks) padded_masks = pad_to_size(masks, self._num_sampled_masks) # Randomly sample groundtruth masks for mask branch training. For the image # without groundtruth masks, it will sample the dummy padded tensors. rand_indices = tf.random.shuffle( tf.range(tf.maximum(num_masks, self._num_sampled_masks))) rand_indices = tf.math.mod(rand_indices, tf.maximum(num_masks, 1)) rand_indices = rand_indices[0:self._num_sampled_masks] rand_indices = tf.reshape(rand_indices, [self._num_sampled_masks]) sampled_boxes = tf.gather(padded_boxes, rand_indices) sampled_classes = tf.gather(padded_classes, rand_indices) sampled_masks = tf.gather(padded_masks, rand_indices) # Jitter the sampled boxes to mimic the noisy detections. sampled_boxes = box_utils.jitter_boxes( sampled_boxes, noise_scale=self._box_jitter_scale) sampled_boxes = box_utils.clip_boxes(sampled_boxes, self._output_size) # Compute mask targets in feature crop. A feature crop fully contains a # sampled box. mask_outer_boxes = box_utils.compute_outer_boxes( sampled_boxes, tf.shape(image)[0:2], scale=self._outer_box_scale) mask_outer_boxes = box_utils.clip_boxes(mask_outer_boxes, self._output_size) # Compensate the offset of mask_outer_boxes to map it back to original image # scale. mask_outer_boxes_ori = mask_outer_boxes mask_outer_boxes_ori += tf.tile(tf.expand_dims(offset, axis=0), [1, 2]) mask_outer_boxes_ori /= tf.tile(tf.expand_dims(image_scale, axis=0), [1, 2]) norm_mask_outer_boxes_ori = box_utils.normalize_boxes( mask_outer_boxes_ori, mask_shape) # Set sampled_masks shape to [batch_size, height, width, 1]. sampled_masks = tf.cast(tf.expand_dims(sampled_masks, axis=-1), tf.float32) mask_targets = tf.image.crop_and_resize( sampled_masks, norm_mask_outer_boxes_ori, box_indices=tf.range(self._num_sampled_masks), crop_size=[self._mask_crop_size, self._mask_crop_size], method='bilinear', extrapolation_value=0, name='train_mask_targets') mask_targets = tf.where(tf.greater_equal(mask_targets, 0.5), tf.ones_like(mask_targets), tf.zeros_like(mask_targets)) mask_targets = tf.squeeze(mask_targets, axis=-1) if self._up_sample_factor > 1: fine_mask_targets = tf.image.crop_and_resize( sampled_masks, norm_mask_outer_boxes_ori, box_indices=tf.range(self._num_sampled_masks), crop_size=[ self._mask_crop_size * self._up_sample_factor, self._mask_crop_size * self._up_sample_factor ], method='bilinear', extrapolation_value=0, name='train_mask_targets') fine_mask_targets = tf.where( tf.greater_equal(fine_mask_targets, 0.5), tf.ones_like(fine_mask_targets), tf.zeros_like(fine_mask_targets)) fine_mask_targets = tf.squeeze(fine_mask_targets, axis=-1) else: fine_mask_targets = mask_targets # If bfloat16 is used, casts input image to tf.bfloat16. if self._use_bfloat16: image = tf.cast(image, dtype=tf.bfloat16) valid_image = tf.cast(tf.not_equal(num_masks, 0), tf.int32) if self._mask_train_class == 'all': mask_is_valid = valid_image * tf.ones_like(sampled_classes, tf.int32) else: # Get the intersection of sampled classes with training splits. mask_valid_classes = tf.cast( tf.expand_dims( class_utils.coco_split_class_ids(self._mask_train_class), 1), sampled_classes.dtype) match = tf.reduce_any( tf.equal(tf.expand_dims(sampled_classes, 0), mask_valid_classes), 0) mask_is_valid = valid_image * tf.cast(match, tf.int32) # Packs labels for model_fn outputs. labels = { 'cls_targets': cls_targets, 'box_targets': box_targets, 'anchor_boxes': input_anchor.multilevel_boxes, 'num_positives': num_positives, 'image_info': image_info, # For ShapeMask. 'mask_targets': mask_targets, 'fine_mask_targets': fine_mask_targets, 'mask_is_valid': mask_is_valid, } inputs = { 'image': image, 'image_info': image_info, 'mask_boxes': sampled_boxes, 'mask_outer_boxes': mask_outer_boxes, 'mask_classes': sampled_classes, } return inputs, labels def _parse_predict_data(self, data): """Parse data for ShapeMask training.""" classes = data['groundtruth_classes'] boxes = data['groundtruth_boxes'] masks = data['groundtruth_instance_masks'] # Gets original image and its size. image = data['image'] image_shape = tf.shape(image)[0:2] # If not using category, makes all categories with id = 0. if not self._use_category: classes = tf.cast(tf.greater(classes, 0), dtype=tf.float32) # Normalizes image with mean and std pixel values. image = input_utils.normalize_image(image) # Converts boxes from normalized coordinates to pixel coordinates. boxes = box_utils.denormalize_boxes(boxes, image_shape) # Resizes and crops image. image, image_info = input_utils.resize_and_crop_image( image, self._output_size, self._output_size, aug_scale_min=1.0, aug_scale_max=1.0) image_scale = image_info[2, :] offset = image_info[3, :] # Resizes and crops boxes and masks. boxes = input_utils.resize_and_crop_boxes( boxes, image_scale, image_info[1, :], offset) masks = input_utils.resize_and_crop_masks( tf.expand_dims(masks, axis=-1), image_scale, self._output_size, offset) # Filters out ground truth boxes that are all zeros. indices = box_utils.get_non_empty_box_indices(boxes) boxes = tf.gather(boxes, indices) classes = tf.gather(classes, indices) # Assigns anchors. input_anchor = anchor.Anchor( self._min_level, self._max_level, self._num_scales, self._aspect_ratios, self._anchor_size, self._output_size) anchor_labeler = anchor.AnchorLabeler( input_anchor, self._match_threshold, self._unmatched_threshold) # If bfloat16 is used, casts input image to tf.bfloat16. if self._use_bfloat16: image = tf.cast(image, dtype=tf.bfloat16) labels = { 'anchor_boxes': input_anchor.multilevel_boxes, 'image_info': image_info, } if self._mode == ModeKeys.PREDICT_WITH_GT: # Converts boxes from normalized coordinates to pixel coordinates. groundtruths = { 'source_id': data['source_id'], 'height': data['height'], 'width': data['width'], 'num_detections': tf.shape(data['groundtruth_classes']), 'boxes': box_utils.denormalize_boxes( data['groundtruth_boxes'], image_shape), 'classes': data['groundtruth_classes'], # 'masks': tf.squeeze(masks, axis=-1), 'areas': data['groundtruth_area'], 'is_crowds': tf.cast(data['groundtruth_is_crowd'], tf.int32), } groundtruths['source_id'] = dataloader_utils.process_source_id( groundtruths['source_id']) groundtruths = dataloader_utils.pad_groundtruths_to_fixed_size( groundtruths, self._max_num_instances) # Computes training labels. (cls_targets, box_targets, num_positives) = anchor_labeler.label_anchors( boxes, tf.cast(tf.expand_dims(classes, axis=1), tf.float32)) # Packs labels for model_fn outputs. labels.update({ 'cls_targets': cls_targets, 'box_targets': box_targets, 'num_positives': num_positives, 'groundtruths': groundtruths, }) inputs = { 'image': image, 'image_info': image_info, } return inputs, labels
22,331
41.781609
92
py
models
models-master/official/legacy/detection/dataloader/maskrcnn_parser.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Data parser and processing for Mask R-CNN.""" import tensorflow as tf from official.legacy.detection.dataloader import anchor from official.legacy.detection.dataloader import mode_keys as ModeKeys from official.legacy.detection.dataloader import tf_example_decoder from official.legacy.detection.utils import box_utils from official.legacy.detection.utils import dataloader_utils from official.legacy.detection.utils import input_utils class Parser(object): """Parser to parse an image and its annotations into a dictionary of tensors.""" def __init__(self, output_size, min_level, max_level, num_scales, aspect_ratios, anchor_size, rpn_match_threshold=0.7, rpn_unmatched_threshold=0.3, rpn_batch_size_per_im=256, rpn_fg_fraction=0.5, aug_rand_hflip=False, aug_scale_min=1.0, aug_scale_max=1.0, skip_crowd_during_training=True, max_num_instances=100, include_mask=False, mask_crop_size=112, use_bfloat16=True, mode=None): """Initializes parameters for parsing annotations in the dataset. Args: output_size: `Tensor` or `list` for [height, width] of output image. The output_size should be divided by the largest feature stride 2^max_level. min_level: `int` number of minimum level of the output feature pyramid. max_level: `int` number of maximum level of the output feature pyramid. num_scales: `int` number representing intermediate scales added on each level. For instances, num_scales=2 adds one additional intermediate anchor scales [2^0, 2^0.5] on each level. aspect_ratios: `list` of float numbers representing the aspect raito anchors added on each level. The number indicates the ratio of width to height. For instances, aspect_ratios=[1.0, 2.0, 0.5] adds three anchors on each scale level. anchor_size: `float` number representing the scale of size of the base anchor to the feature stride 2^level. rpn_match_threshold: rpn_unmatched_threshold: rpn_batch_size_per_im: rpn_fg_fraction: aug_rand_hflip: `bool`, if True, augment training with random horizontal flip. aug_scale_min: `float`, the minimum scale applied to `output_size` for data augmentation during training. aug_scale_max: `float`, the maximum scale applied to `output_size` for data augmentation during training. skip_crowd_during_training: `bool`, if True, skip annotations labeled with `is_crowd` equals to 1. max_num_instances: `int` number of maximum number of instances in an image. The groundtruth data will be padded to `max_num_instances`. include_mask: a bool to indicate whether parse mask groundtruth. mask_crop_size: the size which groundtruth mask is cropped to. use_bfloat16: `bool`, if True, cast output image to tf.bfloat16. mode: a ModeKeys. Specifies if this is training, evaluation, prediction or prediction with groundtruths in the outputs. """ self._mode = mode self._max_num_instances = max_num_instances self._skip_crowd_during_training = skip_crowd_during_training self._is_training = (mode == ModeKeys.TRAIN) self._example_decoder = tf_example_decoder.TfExampleDecoder( include_mask=include_mask) # Anchor. self._output_size = output_size self._min_level = min_level self._max_level = max_level self._num_scales = num_scales self._aspect_ratios = aspect_ratios self._anchor_size = anchor_size # Target assigning. self._rpn_match_threshold = rpn_match_threshold self._rpn_unmatched_threshold = rpn_unmatched_threshold self._rpn_batch_size_per_im = rpn_batch_size_per_im self._rpn_fg_fraction = rpn_fg_fraction # Data augmentation. self._aug_rand_hflip = aug_rand_hflip self._aug_scale_min = aug_scale_min self._aug_scale_max = aug_scale_max # Mask. self._include_mask = include_mask self._mask_crop_size = mask_crop_size # Device. self._use_bfloat16 = use_bfloat16 # Data is parsed depending on the model Modekey. if mode == ModeKeys.TRAIN: self._parse_fn = self._parse_train_data elif mode == ModeKeys.EVAL: self._parse_fn = self._parse_eval_data elif mode == ModeKeys.PREDICT or mode == ModeKeys.PREDICT_WITH_GT: self._parse_fn = self._parse_predict_data else: raise ValueError('mode is not defined.') def __call__(self, value): """Parses data to an image and associated training labels. Args: value: a string tensor holding a serialized tf.Example proto. Returns: image, labels: if mode == ModeKeys.TRAIN. see _parse_train_data. {'images': image, 'labels': labels}: if mode == ModeKeys.PREDICT or ModeKeys.PREDICT_WITH_GT. """ with tf.name_scope('parser'): data = self._example_decoder.decode(value) return self._parse_fn(data) def _parse_train_data(self, data): """Parses data for training. Args: data: the decoded tensor dictionary from TfExampleDecoder. Returns: image: image tensor that is preproessed to have normalized value and dimension [output_size[0], output_size[1], 3] labels: a dictionary of tensors used for training. The following describes {key: value} pairs in the dictionary. image_info: a 2D `Tensor` that encodes the information of the image and the applied preprocessing. It is in the format of [[original_height, original_width], [scaled_height, scaled_width], anchor_boxes: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, 4] representing anchor boxes at each level. rpn_score_targets: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, anchors_per_location]. The height_l and width_l represent the dimension of class logits at l-th level. rpn_box_targets: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, anchors_per_location * 4]. The height_l and width_l represent the dimension of bounding box regression output at l-th level. gt_boxes: Groundtruth bounding box annotations. The box is represented in [y1, x1, y2, x2] format. The coordinates are w.r.t the scaled image that is fed to the network. The tennsor is padded with -1 to the fixed dimension [self._max_num_instances, 4]. gt_classes: Groundtruth classes annotations. The tennsor is padded with -1 to the fixed dimension [self._max_num_instances]. gt_masks: groundtrugh masks cropped by the bounding box and resized to a fixed size determined by mask_crop_size. """ classes = data['groundtruth_classes'] boxes = data['groundtruth_boxes'] if self._include_mask: masks = data['groundtruth_instance_masks'] is_crowds = data['groundtruth_is_crowd'] # Skips annotations with `is_crowd` = True. if self._skip_crowd_during_training and self._is_training: num_groundtruths = tf.shape(classes)[0] with tf.control_dependencies([num_groundtruths, is_crowds]): indices = tf.cond( tf.greater(tf.size(is_crowds), 0), lambda: tf.where(tf.logical_not(is_crowds))[:, 0], lambda: tf.cast(tf.range(num_groundtruths), tf.int64)) classes = tf.gather(classes, indices) boxes = tf.gather(boxes, indices) if self._include_mask: masks = tf.gather(masks, indices) # Gets original image and its size. image = data['image'] image_shape = tf.shape(image)[0:2] # Normalizes image with mean and std pixel values. image = input_utils.normalize_image(image) # Flips image randomly during training. if self._aug_rand_hflip: if self._include_mask: image, boxes, masks = input_utils.random_horizontal_flip( image, boxes, masks) else: image, boxes = input_utils.random_horizontal_flip( image, boxes) # Converts boxes from normalized coordinates to pixel coordinates. # Now the coordinates of boxes are w.r.t. the original image. boxes = box_utils.denormalize_boxes(boxes, image_shape) # Resizes and crops image. image, image_info = input_utils.resize_and_crop_image( image, self._output_size, padded_size=input_utils.compute_padded_size( self._output_size, 2 ** self._max_level), aug_scale_min=self._aug_scale_min, aug_scale_max=self._aug_scale_max) image_height, image_width, _ = image.get_shape().as_list() # Resizes and crops boxes. # Now the coordinates of boxes are w.r.t the scaled image. image_scale = image_info[2, :] offset = image_info[3, :] boxes = input_utils.resize_and_crop_boxes( boxes, image_scale, image_info[1, :], offset) # Filters out ground truth boxes that are all zeros. indices = box_utils.get_non_empty_box_indices(boxes) boxes = tf.gather(boxes, indices) classes = tf.gather(classes, indices) if self._include_mask: masks = tf.gather(masks, indices) # Transfer boxes to the original image space and do normalization. cropped_boxes = boxes + tf.tile(tf.expand_dims(offset, axis=0), [1, 2]) cropped_boxes /= tf.tile(tf.expand_dims(image_scale, axis=0), [1, 2]) cropped_boxes = box_utils.normalize_boxes(cropped_boxes, image_shape) num_masks = tf.shape(masks)[0] masks = tf.image.crop_and_resize( tf.expand_dims(masks, axis=-1), cropped_boxes, box_indices=tf.range(num_masks, dtype=tf.int32), crop_size=[self._mask_crop_size, self._mask_crop_size], method='bilinear') masks = tf.squeeze(masks, axis=-1) # Assigns anchor targets. # Note that after the target assignment, box targets are absolute pixel # offsets w.r.t. the scaled image. input_anchor = anchor.Anchor( self._min_level, self._max_level, self._num_scales, self._aspect_ratios, self._anchor_size, (image_height, image_width)) anchor_labeler = anchor.RpnAnchorLabeler( input_anchor, self._rpn_match_threshold, self._rpn_unmatched_threshold, self._rpn_batch_size_per_im, self._rpn_fg_fraction) rpn_score_targets, rpn_box_targets = anchor_labeler.label_anchors( boxes, tf.cast(tf.expand_dims(classes, axis=-1), dtype=tf.float32)) # If bfloat16 is used, casts input image to tf.bfloat16. if self._use_bfloat16: image = tf.cast(image, dtype=tf.bfloat16) inputs = { 'image': image, 'image_info': image_info, } # Packs labels for model_fn outputs. labels = { 'anchor_boxes': input_anchor.multilevel_boxes, 'image_info': image_info, 'rpn_score_targets': rpn_score_targets, 'rpn_box_targets': rpn_box_targets, } inputs['gt_boxes'] = input_utils.pad_to_fixed_size(boxes, self._max_num_instances, -1) inputs['gt_classes'] = input_utils.pad_to_fixed_size( classes, self._max_num_instances, -1) if self._include_mask: inputs['gt_masks'] = input_utils.pad_to_fixed_size( masks, self._max_num_instances, -1) return inputs, labels def _parse_eval_data(self, data): """Parses data for evaluation.""" raise NotImplementedError('Not implemented!') def _parse_predict_data(self, data): """Parses data for prediction. Args: data: the decoded tensor dictionary from TfExampleDecoder. Returns: A dictionary of {'images': image, 'labels': labels} where image: image tensor that is preproessed to have normalized value and dimension [output_size[0], output_size[1], 3] labels: a dictionary of tensors used for training. The following describes {key: value} pairs in the dictionary. source_ids: Source image id. Default value -1 if the source id is empty in the groundtruth annotation. image_info: a 2D `Tensor` that encodes the information of the image and the applied preprocessing. It is in the format of [[original_height, original_width], [scaled_height, scaled_width], anchor_boxes: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, 4] representing anchor boxes at each level. """ # Gets original image and its size. image = data['image'] image_shape = tf.shape(image)[0:2] # Normalizes image with mean and std pixel values. image = input_utils.normalize_image(image) # Resizes and crops image. image, image_info = input_utils.resize_and_crop_image( image, self._output_size, padded_size=input_utils.compute_padded_size( self._output_size, 2 ** self._max_level), aug_scale_min=1.0, aug_scale_max=1.0) image_height, image_width, _ = image.get_shape().as_list() # If bfloat16 is used, casts input image to tf.bfloat16. if self._use_bfloat16: image = tf.cast(image, dtype=tf.bfloat16) # Compute Anchor boxes. _ = anchor.Anchor(self._min_level, self._max_level, self._num_scales, self._aspect_ratios, self._anchor_size, (image_height, image_width)) labels = { 'image_info': image_info, } if self._mode == ModeKeys.PREDICT_WITH_GT: # Converts boxes from normalized coordinates to pixel coordinates. boxes = box_utils.denormalize_boxes( data['groundtruth_boxes'], image_shape) groundtruths = { 'source_id': data['source_id'], 'height': data['height'], 'width': data['width'], 'num_detections': tf.shape(data['groundtruth_classes']), 'boxes': boxes, 'classes': data['groundtruth_classes'], 'areas': data['groundtruth_area'], 'is_crowds': tf.cast(data['groundtruth_is_crowd'], tf.int32), } groundtruths['source_id'] = dataloader_utils.process_source_id( groundtruths['source_id']) groundtruths = dataloader_utils.pad_groundtruths_to_fixed_size( groundtruths, self._max_num_instances) # TODO(yeqing): Remove the `groundtrtuh` layer key (no longer needed). labels['groundtruths'] = groundtruths inputs = { 'image': image, 'image_info': image_info, } return inputs, labels
15,731
40.183246
82
py
models
models-master/official/legacy/detection/dataloader/input_reader.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Data loader and input processing.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from typing import Optional, Text import tensorflow as tf from official.legacy.detection.dataloader import factory from official.legacy.detection.dataloader import mode_keys as ModeKeys from official.modeling.hyperparams import params_dict class InputFn(object): """Input function that creates dataset from files.""" def __init__(self, file_pattern: Text, params: params_dict.ParamsDict, mode: Text, batch_size: int, num_examples: Optional[int] = -1): """Initialize. Args: file_pattern: the file pattern for the data example (TFRecords). params: the parameter object for constructing example parser and model. mode: ModeKeys.TRAIN or ModeKeys.Eval batch_size: the data batch size. num_examples: If positive, only takes this number of examples and raise tf.errors.OutOfRangeError after that. If non-positive, it will be ignored. """ assert file_pattern is not None assert mode is not None assert batch_size is not None self._file_pattern = file_pattern self._mode = mode self._is_training = (mode == ModeKeys.TRAIN) self._batch_size = batch_size self._num_examples = num_examples self._parser_fn = factory.parser_generator(params, mode) self._dataset_fn = tf.data.TFRecordDataset self._input_sharding = (not self._is_training) try: if self._is_training: self._input_sharding = params.train.input_sharding else: self._input_sharding = params.eval.input_sharding except AttributeError: pass def __call__(self, ctx=None, batch_size: int = None): """Provides tf.data.Dataset object. Args: ctx: context object. batch_size: expected batch size input data. Returns: tf.data.Dataset object. """ if not batch_size: batch_size = self._batch_size assert batch_size is not None dataset = tf.data.Dataset.list_files( self._file_pattern, shuffle=self._is_training) if self._input_sharding and ctx and ctx.num_input_pipelines > 1: dataset = dataset.shard(ctx.num_input_pipelines, ctx.input_pipeline_id) dataset = dataset.cache() if self._is_training: dataset = dataset.repeat() dataset = dataset.interleave( map_func=self._dataset_fn, cycle_length=32, num_parallel_calls=tf.data.experimental.AUTOTUNE) if self._is_training: dataset = dataset.shuffle(1000) if self._num_examples > 0: dataset = dataset.take(self._num_examples) # Parses the fetched records to input tensors for model function. dataset = dataset.map( self._parser_fn, num_parallel_calls=tf.data.experimental.AUTOTUNE) dataset = dataset.batch(batch_size, drop_remainder=True) dataset = dataset.prefetch(tf.data.experimental.AUTOTUNE) return dataset
3,653
33.471698
77
py
models
models-master/official/legacy/detection/dataloader/tf_example_decoder.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tensorflow Example proto decoder for object detection. A decoder to decode string tensors containing serialized tensorflow.Example protos for object detection. """ import tensorflow as tf class TfExampleDecoder(object): """Tensorflow Example proto decoder.""" def __init__(self, include_mask=False): self._include_mask = include_mask self._keys_to_features = { 'image/encoded': tf.io.FixedLenFeature((), tf.string), 'image/source_id': tf.io.FixedLenFeature((), tf.string), 'image/height': tf.io.FixedLenFeature((), tf.int64), 'image/width': tf.io.FixedLenFeature((), tf.int64), 'image/object/bbox/xmin': tf.io.VarLenFeature(tf.float32), 'image/object/bbox/xmax': tf.io.VarLenFeature(tf.float32), 'image/object/bbox/ymin': tf.io.VarLenFeature(tf.float32), 'image/object/bbox/ymax': tf.io.VarLenFeature(tf.float32), 'image/object/class/label': tf.io.VarLenFeature(tf.int64), 'image/object/area': tf.io.VarLenFeature(tf.float32), 'image/object/is_crowd': tf.io.VarLenFeature(tf.int64), } if include_mask: self._keys_to_features.update({ 'image/object/mask': tf.io.VarLenFeature(tf.string), }) def _decode_image(self, parsed_tensors): """Decodes the image and set its static shape.""" image = tf.io.decode_image(parsed_tensors['image/encoded'], channels=3) image.set_shape([None, None, 3]) return image def _decode_boxes(self, parsed_tensors): """Concat box coordinates in the format of [ymin, xmin, ymax, xmax].""" xmin = parsed_tensors['image/object/bbox/xmin'] xmax = parsed_tensors['image/object/bbox/xmax'] ymin = parsed_tensors['image/object/bbox/ymin'] ymax = parsed_tensors['image/object/bbox/ymax'] return tf.stack([ymin, xmin, ymax, xmax], axis=-1) def _decode_masks(self, parsed_tensors): """Decode a set of PNG masks to the tf.float32 tensors.""" def _decode_png_mask(png_bytes): mask = tf.squeeze( tf.io.decode_png(png_bytes, channels=1, dtype=tf.uint8), axis=-1) mask = tf.cast(mask, dtype=tf.float32) mask.set_shape([None, None]) return mask height = parsed_tensors['image/height'] width = parsed_tensors['image/width'] masks = parsed_tensors['image/object/mask'] return tf.cond( pred=tf.greater(tf.size(input=masks), 0), true_fn=lambda: tf.map_fn(_decode_png_mask, masks, dtype=tf.float32), false_fn=lambda: tf.zeros([0, height, width], dtype=tf.float32)) def _decode_areas(self, parsed_tensors): xmin = parsed_tensors['image/object/bbox/xmin'] xmax = parsed_tensors['image/object/bbox/xmax'] ymin = parsed_tensors['image/object/bbox/ymin'] ymax = parsed_tensors['image/object/bbox/ymax'] return tf.cond( tf.greater(tf.shape(parsed_tensors['image/object/area'])[0], 0), lambda: parsed_tensors['image/object/area'], lambda: (xmax - xmin) * (ymax - ymin)) def decode(self, serialized_example): """Decode the serialized example. Args: serialized_example: a single serialized tf.Example string. Returns: decoded_tensors: a dictionary of tensors with the following fields: - image: a uint8 tensor of shape [None, None, 3]. - source_id: a string scalar tensor. - height: an integer scalar tensor. - width: an integer scalar tensor. - groundtruth_classes: a int64 tensor of shape [None]. - groundtruth_is_crowd: a bool tensor of shape [None]. - groundtruth_area: a float32 tensor of shape [None]. - groundtruth_boxes: a float32 tensor of shape [None, 4]. - groundtruth_instance_masks: a float32 tensor of shape [None, None, None]. - groundtruth_instance_masks_png: a string tensor of shape [None]. """ parsed_tensors = tf.io.parse_single_example( serialized=serialized_example, features=self._keys_to_features) for k in parsed_tensors: if isinstance(parsed_tensors[k], tf.SparseTensor): if parsed_tensors[k].dtype == tf.string: parsed_tensors[k] = tf.sparse.to_dense( parsed_tensors[k], default_value='') else: parsed_tensors[k] = tf.sparse.to_dense( parsed_tensors[k], default_value=0) image = self._decode_image(parsed_tensors) boxes = self._decode_boxes(parsed_tensors) areas = self._decode_areas(parsed_tensors) is_crowds = tf.cond( tf.greater(tf.shape(parsed_tensors['image/object/is_crowd'])[0], 0), lambda: tf.cast(parsed_tensors['image/object/is_crowd'], dtype=tf.bool), lambda: tf.zeros_like(parsed_tensors['image/object/class/label'], dtype=tf.bool)) # pylint: disable=line-too-long if self._include_mask: masks = self._decode_masks(parsed_tensors) decoded_tensors = { 'image': image, 'source_id': parsed_tensors['image/source_id'], 'height': parsed_tensors['image/height'], 'width': parsed_tensors['image/width'], 'groundtruth_classes': parsed_tensors['image/object/class/label'], 'groundtruth_is_crowd': is_crowds, 'groundtruth_area': areas, 'groundtruth_boxes': boxes, } if self._include_mask: decoded_tensors.update({ 'groundtruth_instance_masks': masks, 'groundtruth_instance_masks_png': parsed_tensors['image/object/mask'], }) return decoded_tensors
6,219
38.617834
122
py
models
models-master/official/legacy/detection/dataloader/factory.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Model architecture factory.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from official.legacy.detection.dataloader import maskrcnn_parser from official.legacy.detection.dataloader import olnmask_parser from official.legacy.detection.dataloader import retinanet_parser from official.legacy.detection.dataloader import shapemask_parser def parser_generator(params, mode): """Generator function for various dataset parser.""" if params.architecture.parser == 'retinanet_parser': anchor_params = params.anchor parser_params = params.retinanet_parser parser_fn = retinanet_parser.Parser( output_size=parser_params.output_size, min_level=params.architecture.min_level, max_level=params.architecture.max_level, num_scales=anchor_params.num_scales, aspect_ratios=anchor_params.aspect_ratios, anchor_size=anchor_params.anchor_size, match_threshold=parser_params.match_threshold, unmatched_threshold=parser_params.unmatched_threshold, aug_rand_hflip=parser_params.aug_rand_hflip, aug_scale_min=parser_params.aug_scale_min, aug_scale_max=parser_params.aug_scale_max, use_autoaugment=parser_params.use_autoaugment, autoaugment_policy_name=parser_params.autoaugment_policy_name, skip_crowd_during_training=parser_params.skip_crowd_during_training, max_num_instances=parser_params.max_num_instances, use_bfloat16=params.architecture.use_bfloat16, mode=mode) elif params.architecture.parser == 'maskrcnn_parser': anchor_params = params.anchor parser_params = params.maskrcnn_parser parser_fn = maskrcnn_parser.Parser( output_size=parser_params.output_size, min_level=params.architecture.min_level, max_level=params.architecture.max_level, num_scales=anchor_params.num_scales, aspect_ratios=anchor_params.aspect_ratios, anchor_size=anchor_params.anchor_size, rpn_match_threshold=parser_params.rpn_match_threshold, rpn_unmatched_threshold=parser_params.rpn_unmatched_threshold, rpn_batch_size_per_im=parser_params.rpn_batch_size_per_im, rpn_fg_fraction=parser_params.rpn_fg_fraction, aug_rand_hflip=parser_params.aug_rand_hflip, aug_scale_min=parser_params.aug_scale_min, aug_scale_max=parser_params.aug_scale_max, skip_crowd_during_training=parser_params.skip_crowd_during_training, max_num_instances=parser_params.max_num_instances, include_mask=params.architecture.include_mask, mask_crop_size=parser_params.mask_crop_size, use_bfloat16=params.architecture.use_bfloat16, mode=mode) elif params.architecture.parser == 'olnmask_parser': anchor_params = params.anchor parser_params = params.olnmask_parser parser_fn = olnmask_parser.Parser( output_size=parser_params.output_size, min_level=params.architecture.min_level, max_level=params.architecture.max_level, num_scales=anchor_params.num_scales, aspect_ratios=anchor_params.aspect_ratios, anchor_size=anchor_params.anchor_size, rpn_match_threshold=parser_params.rpn_match_threshold, rpn_unmatched_threshold=parser_params.rpn_unmatched_threshold, rpn_batch_size_per_im=parser_params.rpn_batch_size_per_im, rpn_fg_fraction=parser_params.rpn_fg_fraction, aug_rand_hflip=parser_params.aug_rand_hflip, aug_scale_min=parser_params.aug_scale_min, aug_scale_max=parser_params.aug_scale_max, skip_crowd_during_training=parser_params.skip_crowd_during_training, max_num_instances=parser_params.max_num_instances, include_mask=params.architecture.include_mask, mask_crop_size=parser_params.mask_crop_size, use_bfloat16=params.architecture.use_bfloat16, mode=mode, has_centerness=parser_params.has_centerness, rpn_center_match_iou_threshold=( parser_params.rpn_center_match_iou_threshold), rpn_center_unmatched_iou_threshold=( parser_params.rpn_center_unmatched_iou_threshold), rpn_num_center_samples_per_im=( parser_params.rpn_num_center_samples_per_im), class_agnostic=parser_params.class_agnostic, train_class=parser_params.train_class,) elif params.architecture.parser == 'shapemask_parser': anchor_params = params.anchor parser_params = params.shapemask_parser parser_fn = shapemask_parser.Parser( output_size=parser_params.output_size, min_level=params.architecture.min_level, max_level=params.architecture.max_level, num_scales=anchor_params.num_scales, aspect_ratios=anchor_params.aspect_ratios, anchor_size=anchor_params.anchor_size, use_category=parser_params.use_category, outer_box_scale=parser_params.outer_box_scale, box_jitter_scale=parser_params.box_jitter_scale, num_sampled_masks=parser_params.num_sampled_masks, mask_crop_size=parser_params.mask_crop_size, mask_min_level=parser_params.mask_min_level, mask_max_level=parser_params.mask_max_level, upsample_factor=parser_params.upsample_factor, match_threshold=parser_params.match_threshold, unmatched_threshold=parser_params.unmatched_threshold, aug_rand_hflip=parser_params.aug_rand_hflip, aug_scale_min=parser_params.aug_scale_min, aug_scale_max=parser_params.aug_scale_max, skip_crowd_during_training=parser_params.skip_crowd_during_training, max_num_instances=parser_params.max_num_instances, use_bfloat16=params.architecture.use_bfloat16, mask_train_class=parser_params.mask_train_class, mode=mode) else: raise ValueError('Parser %s is not supported.' % params.architecture.parser) return parser_fn
6,563
46.912409
80
py
models
models-master/official/legacy/detection/dataloader/mode_keys.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Standard names for input dataloader modes. The following standard keys are defined: * `TRAIN`: training mode. * `EVAL`: evaluation mode. * `PREDICT`: prediction mode. * `PREDICT_WITH_GT`: prediction mode with groundtruths in returned variables. """ from __future__ import absolute_import from __future__ import division from __future__ import print_function TRAIN = 'train' EVAL = 'eval' PREDICT = 'predict' PREDICT_WITH_GT = 'predict_with_gt'
1,061
30.235294
77
py
models
models-master/official/legacy/detection/dataloader/anchor.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Anchor box and labeler definition.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import collections import tensorflow as tf from official.legacy.detection.utils import box_utils from official.vision.ops import iou_similarity from official.vision.utils.object_detection import argmax_matcher from official.vision.utils.object_detection import balanced_positive_negative_sampler from official.vision.utils.object_detection import box_list from official.vision.utils.object_detection import faster_rcnn_box_coder from official.vision.utils.object_detection import target_assigner class Anchor(object): """Anchor class for anchor-based object detectors.""" def __init__(self, min_level, max_level, num_scales, aspect_ratios, anchor_size, image_size): """Constructs multiscale anchors. Args: min_level: integer number of minimum level of the output feature pyramid. max_level: integer number of maximum level of the output feature pyramid. num_scales: integer number representing intermediate scales added on each level. For instances, num_scales=2 adds one additional intermediate anchor scales [2^0, 2^0.5] on each level. aspect_ratios: list of float numbers representing the aspect ratio anchors added on each level. The number indicates the ratio of width to height. For instances, aspect_ratios=[1.0, 2.0, 0.5] adds three anchors on each scale level. anchor_size: float number representing the scale of size of the base anchor to the feature stride 2^level. image_size: a list of integer numbers or Tensors representing [height, width] of the input image size.The image_size should be divisible by the largest feature stride 2^max_level. """ self.min_level = min_level self.max_level = max_level self.num_scales = num_scales self.aspect_ratios = aspect_ratios self.anchor_size = anchor_size self.image_size = image_size self.boxes = self._generate_boxes() def _generate_boxes(self): """Generates multiscale anchor boxes. Returns: a Tensor of shape [N, 4], represneting anchor boxes of all levels concatenated together. """ boxes_all = [] for level in range(self.min_level, self.max_level + 1): boxes_l = [] for scale in range(self.num_scales): for aspect_ratio in self.aspect_ratios: stride = 2**level intermediate_scale = 2**(scale / float(self.num_scales)) base_anchor_size = self.anchor_size * stride * intermediate_scale aspect_x = aspect_ratio**0.5 aspect_y = aspect_ratio**-0.5 half_anchor_size_x = base_anchor_size * aspect_x / 2.0 half_anchor_size_y = base_anchor_size * aspect_y / 2.0 x = tf.range(stride / 2, self.image_size[1], stride) y = tf.range(stride / 2, self.image_size[0], stride) xv, yv = tf.meshgrid(x, y) xv = tf.cast(tf.reshape(xv, [-1]), dtype=tf.float32) yv = tf.cast(tf.reshape(yv, [-1]), dtype=tf.float32) # Tensor shape Nx4. boxes = tf.stack([ yv - half_anchor_size_y, xv - half_anchor_size_x, yv + half_anchor_size_y, xv + half_anchor_size_x ], axis=1) boxes_l.append(boxes) # Concat anchors on the same level to tensor shape NxAx4. boxes_l = tf.stack(boxes_l, axis=1) boxes_l = tf.reshape(boxes_l, [-1, 4]) boxes_all.append(boxes_l) return tf.concat(boxes_all, axis=0) def unpack_labels(self, labels): """Unpacks an array of labels into multiscales labels.""" unpacked_labels = collections.OrderedDict() count = 0 for level in range(self.min_level, self.max_level + 1): feat_size_y = tf.cast(self.image_size[0] / 2**level, tf.int32) feat_size_x = tf.cast(self.image_size[1] / 2**level, tf.int32) steps = feat_size_y * feat_size_x * self.anchors_per_location unpacked_labels[level] = tf.reshape(labels[count:count + steps], [feat_size_y, feat_size_x, -1]) count += steps return unpacked_labels @property def anchors_per_location(self): return self.num_scales * len(self.aspect_ratios) @property def multilevel_boxes(self): return self.unpack_labels(self.boxes) class AnchorLabeler(object): """Labeler for dense object detector.""" def __init__(self, anchor, match_threshold=0.5, unmatched_threshold=0.5): """Constructs anchor labeler to assign labels to anchors. Args: anchor: an instance of class Anchors. match_threshold: a float number between 0 and 1 representing the lower-bound threshold to assign positive labels for anchors. An anchor with a score over the threshold is labeled positive. unmatched_threshold: a float number between 0 and 1 representing the upper-bound threshold to assign negative labels for anchors. An anchor with a score below the threshold is labeled negative. """ similarity_calc = iou_similarity.IouSimilarity() matcher = argmax_matcher.ArgMaxMatcher( match_threshold, unmatched_threshold=unmatched_threshold, negatives_lower_than_unmatched=True, force_match_for_each_row=True) box_coder = faster_rcnn_box_coder.FasterRcnnBoxCoder() self._target_assigner = target_assigner.TargetAssigner( similarity_calc, matcher, box_coder) self._anchor = anchor self._match_threshold = match_threshold self._unmatched_threshold = unmatched_threshold def label_anchors(self, gt_boxes, gt_labels): """Labels anchors with ground truth inputs. Args: gt_boxes: A float tensor with shape [N, 4] representing groundtruth boxes. For each row, it stores [y0, x0, y1, x1] for four corners of a box. gt_labels: A integer tensor with shape [N, 1] representing groundtruth classes. Returns: cls_targets_dict: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, num_anchors_per_location]. The height_l and width_l represent the dimension of class logits at l-th level. box_targets_dict: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, num_anchors_per_location * 4]. The height_l and width_l represent the dimension of bounding box regression output at l-th level. num_positives: scalar tensor storing number of positives in an image. """ gt_box_list = box_list.BoxList(gt_boxes) anchor_box_list = box_list.BoxList(self._anchor.boxes) # The cls_weights, box_weights are not used. cls_targets, _, box_targets, _, matches = self._target_assigner.assign( anchor_box_list, gt_box_list, gt_labels) # Labels definition in matches.match_results: # (1) match_results[i]>=0, meaning that column i is matched with row # match_results[i]. # (2) match_results[i]=-1, meaning that column i is not matched. # (3) match_results[i]=-2, meaning that column i is ignored. match_results = tf.expand_dims(matches.match_results, axis=1) cls_targets = tf.cast(cls_targets, tf.int32) cls_targets = tf.where( tf.equal(match_results, -1), -tf.ones_like(cls_targets), cls_targets) cls_targets = tf.where( tf.equal(match_results, -2), -2 * tf.ones_like(cls_targets), cls_targets) # Unpacks labels into multi-level representations. cls_targets_dict = self._anchor.unpack_labels(cls_targets) box_targets_dict = self._anchor.unpack_labels(box_targets) num_positives = tf.reduce_sum( input_tensor=tf.cast(tf.greater(matches.match_results, -1), tf.float32)) return cls_targets_dict, box_targets_dict, num_positives class RpnAnchorLabeler(AnchorLabeler): """Labeler for Region Proposal Network.""" def __init__(self, anchor, match_threshold=0.7, unmatched_threshold=0.3, rpn_batch_size_per_im=256, rpn_fg_fraction=0.5): AnchorLabeler.__init__( self, anchor, match_threshold=0.7, unmatched_threshold=0.3) self._rpn_batch_size_per_im = rpn_batch_size_per_im self._rpn_fg_fraction = rpn_fg_fraction def _get_rpn_samples(self, match_results): """Computes anchor labels. This function performs subsampling for foreground (fg) and background (bg) anchors. Args: match_results: A integer tensor with shape [N] representing the matching results of anchors. (1) match_results[i]>=0, meaning that column i is matched with row match_results[i]. (2) match_results[i]=-1, meaning that column i is not matched. (3) match_results[i]=-2, meaning that column i is ignored. Returns: score_targets: a integer tensor with the a shape of [N]. (1) score_targets[i]=1, the anchor is a positive sample. (2) score_targets[i]=0, negative. (3) score_targets[i]=-1, the anchor is don't care (ignore). """ sampler = ( balanced_positive_negative_sampler.BalancedPositiveNegativeSampler( positive_fraction=self._rpn_fg_fraction, is_static=False)) # indicator includes both positive and negative labels. # labels includes only positives labels. # positives = indicator & labels. # negatives = indicator & !labels. # ignore = !indicator. indicator = tf.greater(match_results, -2) labels = tf.greater(match_results, -1) samples = sampler.subsample(indicator, self._rpn_batch_size_per_im, labels) positive_labels = tf.where( tf.logical_and(samples, labels), tf.constant(2, dtype=tf.int32, shape=match_results.shape), tf.constant(0, dtype=tf.int32, shape=match_results.shape)) negative_labels = tf.where( tf.logical_and(samples, tf.logical_not(labels)), tf.constant(1, dtype=tf.int32, shape=match_results.shape), tf.constant(0, dtype=tf.int32, shape=match_results.shape)) ignore_labels = tf.fill(match_results.shape, -1) return (ignore_labels + positive_labels + negative_labels, positive_labels, negative_labels) def label_anchors(self, gt_boxes, gt_labels): """Labels anchors with ground truth inputs. Args: gt_boxes: A float tensor with shape [N, 4] representing groundtruth boxes. For each row, it stores [y0, x0, y1, x1] for four corners of a box. gt_labels: A integer tensor with shape [N, 1] representing groundtruth classes. Returns: score_targets_dict: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, num_anchors]. The height_l and width_l represent the dimension of class logits at l-th level. box_targets_dict: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, num_anchors * 4]. The height_l and width_l represent the dimension of bounding box regression output at l-th level. """ gt_box_list = box_list.BoxList(gt_boxes) anchor_box_list = box_list.BoxList(self._anchor.boxes) # cls_targets, cls_weights, box_weights are not used. _, _, box_targets, _, matches = self._target_assigner.assign( anchor_box_list, gt_box_list, gt_labels) # score_targets contains the subsampled positive and negative anchors. score_targets, _, _ = self._get_rpn_samples(matches.match_results) # Unpacks labels. score_targets_dict = self._anchor.unpack_labels(score_targets) box_targets_dict = self._anchor.unpack_labels(box_targets) return score_targets_dict, box_targets_dict class OlnAnchorLabeler(RpnAnchorLabeler): """Labeler for Region Proposal Network.""" def __init__(self, anchor, match_threshold=0.7, unmatched_threshold=0.3, rpn_batch_size_per_im=256, rpn_fg_fraction=0.5, has_centerness=False, center_match_iou_threshold=0.3, center_unmatched_iou_threshold=0.1, num_center_samples_per_im=256): """Constructs rpn anchor labeler to assign labels and centerness to anchors. Args: anchor: an instance of class Anchors. match_threshold: a float number between 0 and 1 representing the lower-bound threshold to assign positive labels for anchors. An anchor with a score over the threshold is labeled positive. unmatched_threshold: a float number between 0 and 1 representing the upper-bound threshold to assign negative labels for anchors. An anchor with a score below the threshold is labeled negative. rpn_batch_size_per_im: number of anchors that are sampled per image. rpn_fg_fraction: has_centerness: whether to include centerness target creation. An anchor is paired with one centerness score. center_match_iou_threshold: a float number between 0 and 1 representing the lower-bound threshold to sample foreground anchors for centerness regression. An anchor with a score over the threshold is sampled as foreground sample for centerness regression. We sample mostly from the foreground region (255 out of 256 samples). That is, we sample 255 vs 1 (foreground vs background) anchor points to learn centerness regression. center_unmatched_iou_threshold: a float number between 0 and 1 representing the lower-bound threshold to sample background anchors for centerness regression. An anchor with a score over the threshold is sampled as foreground sample for centerness regression. We sample very sparsely from the background region (1 out of 256 samples). That is, we sample 255 vs 1 (foreground vs background) anchor points to learn centerness regression. num_center_samples_per_im: number of anchor points per image that are sampled as centerness targets. """ super(OlnAnchorLabeler, self).__init__( anchor, match_threshold=match_threshold, unmatched_threshold=unmatched_threshold, rpn_batch_size_per_im=rpn_batch_size_per_im, rpn_fg_fraction=rpn_fg_fraction) similarity_calc = iou_similarity.IouSimilarity() matcher = argmax_matcher.ArgMaxMatcher( match_threshold, unmatched_threshold=unmatched_threshold, negatives_lower_than_unmatched=True, force_match_for_each_row=True) box_coder = faster_rcnn_box_coder.FasterRcnnBoxCoder() if has_centerness: center_matcher = argmax_matcher.ArgMaxMatcher( center_match_iou_threshold, unmatched_threshold=center_match_iou_threshold, negatives_lower_than_unmatched=True, force_match_for_each_row=True,) else: center_matcher = None self._target_assigner = target_assigner.OlnTargetAssigner( similarity_calc, matcher, box_coder, center_matcher=center_matcher) self._num_center_samples_per_im = num_center_samples_per_im self._center_unmatched_iou_threshold = center_unmatched_iou_threshold self._rpn_batch_size_per_im = rpn_batch_size_per_im self._rpn_fg_fraction = rpn_fg_fraction def label_anchors_lrtb(self, gt_boxes, gt_labels): """Labels anchors with ground truth inputs. Args: gt_boxes: A float tensor with shape [N, 4] representing groundtruth boxes. For each row, it stores [y0, x0, y1, x1] for four corners of a box. gt_labels: A integer tensor with shape [N, 1] representing groundtruth classes. Returns: score_targets_dict: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, num_anchors]. The height_l and width_l represent the dimension of class logits at l-th level. box_targets_dict: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, num_anchors * 4]. The height_l and width_l represent the dimension of bounding box regression output at l-th level. lrtb_targets_dict: Same strucure to box_target_dict, except the regression targets are converted from xyhw to lrtb format. Ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, num_anchors * 4]. The height_l and width_l represent the dimension of bounding box regression output at l-th level. center_targets_dict: Same structure to score_tragets_dict, except the scores are centerness values ranging from 0 to 1. Ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, num_anchors]. The height_l and width_l represent the dimension of class logits at l-th level. """ gt_box_list = box_list.BoxList(gt_boxes) anchor_box_list = box_list.BoxList(self._anchor.boxes) # cls_targets, cls_weights, box_weights are not used. (_, _, box_targets, _, matches, matched_gt_box_list, matched_anchors_mask, center_matched_gt_box_list, center_matched_anchors_mask, matched_ious) = self._target_assigner.assign( anchor_box_list, gt_box_list, gt_labels) # Box lrtb_targets. lrtb_targets, _ = box_utils.encode_boxes_lrtb( matched_gt_box_list.data['boxes'], anchor_box_list.data['boxes'], weights=[1.0, 1.0, 1.0, 1.0]) lrtb_sanity = tf.logical_and( tf.greater(tf.reduce_min(lrtb_targets, -1), 0.), matched_anchors_mask) # To broadcast lrtb_sanity to the same shape as lrtb_targets. lrtb_sanity = tf.tile(tf.expand_dims(lrtb_sanity, 1), [1, tf.shape(lrtb_targets)[1]]) lrtb_targets = tf.where(lrtb_sanity, lrtb_targets, tf.zeros_like(lrtb_targets)) # RPN anchor-gtbox iou values. iou_targets = tf.where(tf.greater(matched_ious, 0.0), matched_ious, tf.zeros_like(matched_ious)) # Centerness_targets. _, center_targets = box_utils.encode_boxes_lrtb( center_matched_gt_box_list.data['boxes'], anchor_box_list.data['boxes'], weights=[1.0, 1.0, 1.0, 1.0]) # Positive-negative centerness sampler. num_center_samples_per_im = self._num_center_samples_per_im center_pos_neg_sampler = ( balanced_positive_negative_sampler.BalancedPositiveNegativeSampler( positive_fraction=(1.- 1./num_center_samples_per_im), is_static=False)) center_pos_neg_indicator = tf.logical_or( center_matched_anchors_mask, tf.less(iou_targets, self._center_unmatched_iou_threshold)) center_pos_labels = center_matched_anchors_mask center_samples = center_pos_neg_sampler.subsample( center_pos_neg_indicator, num_center_samples_per_im, center_pos_labels) is_valid = center_samples center_targets = tf.where(is_valid, center_targets, (-1) * tf.ones_like(center_targets)) # score_targets contains the subsampled positive and negative anchors. score_targets, _, _ = self._get_rpn_samples(matches.match_results) # Unpacks labels. score_targets_dict = self._anchor.unpack_labels(score_targets) box_targets_dict = self._anchor.unpack_labels(box_targets) lrtb_targets_dict = self._anchor.unpack_labels(lrtb_targets) center_targets_dict = self._anchor.unpack_labels(center_targets) return (score_targets_dict, box_targets_dict, lrtb_targets_dict, center_targets_dict)
20,689
44.076253
85
py
models
models-master/official/legacy/detection/dataloader/__init__.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
609
39.666667
74
py
models
models-master/official/legacy/detection/dataloader/olnmask_parser.py
# Copyright 2023 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Data parser and processing for Mask R-CNN.""" import tensorflow as tf from official.legacy.detection.dataloader import anchor from official.legacy.detection.dataloader.maskrcnn_parser import Parser as MaskrcnnParser from official.legacy.detection.utils import box_utils from official.legacy.detection.utils import class_utils from official.legacy.detection.utils import input_utils class Parser(MaskrcnnParser): """Parser to parse an image and its annotations into a dictionary of tensors.""" def __init__(self, output_size, min_level, max_level, num_scales, aspect_ratios, anchor_size, rpn_match_threshold=0.7, rpn_unmatched_threshold=0.3, rpn_batch_size_per_im=256, rpn_fg_fraction=0.5, aug_rand_hflip=False, aug_scale_min=1.0, aug_scale_max=1.0, skip_crowd_during_training=True, max_num_instances=100, include_mask=False, mask_crop_size=112, use_bfloat16=True, mode=None, # for centerness learning. has_centerness=False, rpn_center_match_iou_threshold=0.3, rpn_center_unmatched_iou_threshold=0.1, rpn_num_center_samples_per_im=256, # for class manipulation. class_agnostic=False, train_class='all', ): """Initializes parameters for parsing annotations in the dataset. Args: output_size: `Tensor` or `list` for [height, width] of output image. The output_size should be divided by the largest feature stride 2^max_level. min_level: `int` number of minimum level of the output feature pyramid. max_level: `int` number of maximum level of the output feature pyramid. num_scales: `int` number representing intermediate scales added on each level. For instances, num_scales=2 adds one additional intermediate anchor scales [2^0, 2^0.5] on each level. aspect_ratios: `list` of float numbers representing the aspect raito anchors added on each level. The number indicates the ratio of width to height. For instances, aspect_ratios=[1.0, 2.0, 0.5] adds three anchors on each scale level. anchor_size: `float` number representing the scale of size of the base anchor to the feature stride 2^level. rpn_match_threshold: rpn_unmatched_threshold: rpn_batch_size_per_im: rpn_fg_fraction: aug_rand_hflip: `bool`, if True, augment training with random horizontal flip. aug_scale_min: `float`, the minimum scale applied to `output_size` for data augmentation during training. aug_scale_max: `float`, the maximum scale applied to `output_size` for data augmentation during training. skip_crowd_during_training: `bool`, if True, skip annotations labeled with `is_crowd` equals to 1. max_num_instances: `int` number of maximum number of instances in an image. The groundtruth data will be padded to `max_num_instances`. include_mask: a bool to indicate whether parse mask groundtruth. mask_crop_size: the size which groundtruth mask is cropped to. use_bfloat16: `bool`, if True, cast output image to tf.bfloat16. mode: a ModeKeys. Specifies if this is training, evaluation, prediction or prediction with groundtruths in the outputs. has_centerness: whether to create centerness targets rpn_center_match_iou_threshold: iou threshold for valid centerness samples ,set to 0.3 by default. rpn_center_unmatched_iou_threshold: iou threshold for invalid centerness samples, set to 0.1 by default. rpn_num_center_samples_per_im: number of centerness samples per image, 256 by default. class_agnostic: whether to merge class ids into one foreground(=1) class, False by default. train_class: 'all' or 'voc' or 'nonvoc', 'all' by default. """ super(Parser, self).__init__( output_size=output_size, min_level=min_level, max_level=max_level, num_scales=num_scales, aspect_ratios=aspect_ratios, anchor_size=anchor_size, rpn_match_threshold=rpn_match_threshold, rpn_unmatched_threshold=rpn_unmatched_threshold, rpn_batch_size_per_im=rpn_batch_size_per_im, rpn_fg_fraction=rpn_fg_fraction, aug_rand_hflip=aug_rand_hflip, aug_scale_min=aug_scale_min, aug_scale_max=aug_scale_max, skip_crowd_during_training=skip_crowd_during_training, max_num_instances=max_num_instances, include_mask=include_mask, mask_crop_size=mask_crop_size, use_bfloat16=use_bfloat16, mode=mode,) # Centerness target assigning. self._has_centerness = has_centerness self._rpn_center_match_iou_threshold = rpn_center_match_iou_threshold self._rpn_center_unmatched_iou_threshold = ( rpn_center_unmatched_iou_threshold) self._rpn_num_center_samples_per_im = rpn_num_center_samples_per_im # Class manipulation. self._class_agnostic = class_agnostic self._train_class = train_class def _parse_train_data(self, data): """Parses data for training. Args: data: the decoded tensor dictionary from TfExampleDecoder. Returns: image: image tensor that is preproessed to have normalized value and dimension [output_size[0], output_size[1], 3] labels: a dictionary of tensors used for training. The following describes {key: value} pairs in the dictionary. image_info: a 2D `Tensor` that encodes the information of the image and the applied preprocessing. It is in the format of [[original_height, original_width], [scaled_height, scaled_width], anchor_boxes: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, 4] representing anchor boxes at each level. rpn_score_targets: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, anchors_per_location]. The height_l and width_l represent the dimension of class logits at l-th level. rpn_box_targets: ordered dictionary with keys [min_level, min_level+1, ..., max_level]. The values are tensor with shape [height_l, width_l, anchors_per_location * 4]. The height_l and width_l represent the dimension of bounding box regression output at l-th level. gt_boxes: Groundtruth bounding box annotations. The box is represented in [y1, x1, y2, x2] format. The coordinates are w.r.t the scaled image that is fed to the network. The tennsor is padded with -1 to the fixed dimension [self._max_num_instances, 4]. gt_classes: Groundtruth classes annotations. The tennsor is padded with -1 to the fixed dimension [self._max_num_instances]. gt_masks: groundtrugh masks cropped by the bounding box and resized to a fixed size determined by mask_crop_size. """ classes = data['groundtruth_classes'] boxes = data['groundtruth_boxes'] if self._include_mask: masks = data['groundtruth_instance_masks'] is_crowds = data['groundtruth_is_crowd'] # Skips annotations with `is_crowd` = True. if self._skip_crowd_during_training and self._is_training: num_groundtruths = tf.shape(classes)[0] with tf.control_dependencies([num_groundtruths, is_crowds]): indices = tf.cond( tf.greater(tf.size(is_crowds), 0), lambda: tf.where(tf.logical_not(is_crowds))[:, 0], lambda: tf.cast(tf.range(num_groundtruths), tf.int64)) classes = tf.gather(classes, indices) boxes = tf.gather(boxes, indices) if self._include_mask: masks = tf.gather(masks, indices) # Gets original image and its size. image = data['image'] image_shape = tf.shape(image)[0:2] # Normalizes image with mean and std pixel values. image = input_utils.normalize_image(image) # Flips image randomly during training. if self._aug_rand_hflip: if self._include_mask: image, boxes, masks = input_utils.random_horizontal_flip( image, boxes, masks) else: image, boxes = input_utils.random_horizontal_flip( image, boxes) # Converts boxes from normalized coordinates to pixel coordinates. # Now the coordinates of boxes are w.r.t. the original image. boxes = box_utils.denormalize_boxes(boxes, image_shape) # Resizes and crops image. image, image_info = input_utils.resize_and_crop_image( image, self._output_size, padded_size=input_utils.compute_padded_size( self._output_size, 2 ** self._max_level), aug_scale_min=self._aug_scale_min, aug_scale_max=self._aug_scale_max) image_height, image_width, _ = image.get_shape().as_list() # Resizes and crops boxes. # Now the coordinates of boxes are w.r.t the scaled image. image_scale = image_info[2, :] offset = image_info[3, :] boxes = input_utils.resize_and_crop_boxes( boxes, image_scale, image_info[1, :], offset) # Filters out ground truth boxes that are all zeros. indices = box_utils.get_non_empty_box_indices(boxes) boxes = tf.gather(boxes, indices) classes = tf.gather(classes, indices) if self._include_mask: masks = tf.gather(masks, indices) # Transfer boxes to the original image space and do normalization. cropped_boxes = boxes + tf.tile(tf.expand_dims(offset, axis=0), [1, 2]) cropped_boxes /= tf.tile(tf.expand_dims(image_scale, axis=0), [1, 2]) cropped_boxes = box_utils.normalize_boxes(cropped_boxes, image_shape) num_masks = tf.shape(masks)[0] masks = tf.image.crop_and_resize( tf.expand_dims(masks, axis=-1), cropped_boxes, box_indices=tf.range(num_masks, dtype=tf.int32), crop_size=[self._mask_crop_size, self._mask_crop_size], method='bilinear') masks = tf.squeeze(masks, axis=-1) # Class manipulation. # Filter out novel split classes from training. if self._train_class != 'all': valid_classes = tf.cast( class_utils.coco_split_class_ids(self._train_class), dtype=classes.dtype) match = tf.reduce_any(tf.equal( tf.expand_dims(valid_classes, 1), tf.expand_dims(classes, 0)), 0) # kill novel split classes and boxes. boxes = tf.gather(boxes, tf.where(match)[:, 0]) classes = tf.gather(classes, tf.where(match)[:, 0]) if self._include_mask: masks = tf.gather(masks, tf.where(match)[:, 0]) # Assigns anchor targets. # Note that after the target assignment, box targets are absolute pixel # offsets w.r.t. the scaled image. input_anchor = anchor.Anchor( self._min_level, self._max_level, self._num_scales, self._aspect_ratios, self._anchor_size, (image_height, image_width)) anchor_labeler = anchor.OlnAnchorLabeler( input_anchor, self._rpn_match_threshold, self._rpn_unmatched_threshold, self._rpn_batch_size_per_im, self._rpn_fg_fraction, # for centerness target. self._has_centerness, self._rpn_center_match_iou_threshold, self._rpn_center_unmatched_iou_threshold, self._rpn_num_center_samples_per_im,) if self._has_centerness: rpn_score_targets, _, rpn_lrtb_targets, rpn_center_targets = ( anchor_labeler.label_anchors_lrtb( gt_boxes=boxes, gt_labels=tf.cast( tf.expand_dims(classes, axis=-1), dtype=tf.float32))) else: rpn_score_targets, rpn_box_targets = anchor_labeler.label_anchors( boxes, tf.cast(tf.expand_dims(classes, axis=-1), dtype=tf.float32)) # For base rpn, dummy placeholder for centerness target. rpn_center_targets = rpn_score_targets.copy() # If bfloat16 is used, casts input image to tf.bfloat16. if self._use_bfloat16: image = tf.cast(image, dtype=tf.bfloat16) inputs = { 'image': image, 'image_info': image_info, } # Packs labels for model_fn outputs. labels = { 'anchor_boxes': input_anchor.multilevel_boxes, 'image_info': image_info, 'rpn_score_targets': rpn_score_targets, 'rpn_box_targets': (rpn_lrtb_targets if self._has_centerness else rpn_box_targets), 'rpn_center_targets': rpn_center_targets, } # If class_agnostic, convert to binary classes. if self._class_agnostic: classes = tf.where(tf.greater(classes, 0), tf.ones_like(classes), tf.zeros_like(classes)) inputs['gt_boxes'] = input_utils.pad_to_fixed_size(boxes, self._max_num_instances, -1) inputs['gt_classes'] = input_utils.pad_to_fixed_size( classes, self._max_num_instances, -1) if self._include_mask: inputs['gt_masks'] = input_utils.pad_to_fixed_size( masks, self._max_num_instances, -1) return inputs, labels
14,218
42.35061
89
py