configs/experiment/projects/bridging/dinosaur/_preprocessing_coco_dino_feature_recon_randcrop.yaml
# @package _global_
dataset:
eval_transforms:
03a_preprocessing:
_target_: ocl.transforms.Map
transform:
_target_: torchvision.transforms.Compose
transforms:
- _target_: ocl.preprocessing.InstanceMasksToDenseMasks
- _target_: ocl.preprocessing.AddSegmentationMaskFromInstanceMask
# Drop instance_category again as some images do not contain it
- "${lambda_fn:'lambda data: {k: v for k, v in data.items() if k != \"instance_category\"\
}'}"
- _target_: ocl.preprocessing.AddEmptyMasks
mask_keys:
- instance_mask
- segmentation_mask
fields:
- image
- instance_mask
- instance_category
batch_transform: false
03b_preprocessing:
_target_: ocl.transforms.SimpleTransform
transforms:
image:
_target_: torchvision.transforms.Compose
transforms:
- _target_: torchvision.transforms.ToTensor
- _target_: torchvision.transforms.Resize
size: 224
interpolation: ${torchvision_interpolation_mode:BICUBIC}
- "${lambda_fn:'lambda image: image.clamp(0.0, 1.0)'}" # Bicubic interpolation can get out of range
- _target_: torchvision.transforms.CenterCrop
size: 224
- _target_: torchvision.transforms.Normalize
mean: [0.485, 0.456, 0.406]
std: [0.229, 0.224, 0.225]
instance_mask:
_target_: torchvision.transforms.Compose
transforms:
- _target_: ocl.preprocessing.DenseMaskToTensor
- _target_: ocl.preprocessing.ResizeNearestExact
size: 224
- _target_: torchvision.transforms.CenterCrop
size: 224
segmentation_mask:
_target_: torchvision.transforms.Compose
transforms:
- _target_: ocl.preprocessing.DenseMaskToTensor
- _target_: ocl.preprocessing.ResizeNearestExact
size: 224
- _target_: torchvision.transforms.CenterCrop
size: 224
batch_transform: false
train_transforms:
03b_preprocessing:
_target_: ocl.transforms.SimpleTransform
transforms:
image:
_target_: torchvision.transforms.Compose
transforms:
- _target_: torchvision.transforms.ToTensor
- _target_: torchvision.transforms.Resize
size: 224
interpolation: ${torchvision_interpolation_mode:BICUBIC}
- "${lambda_fn:'lambda image: image.clamp(0.0, 1.0)'}" # Bicubic interpolation can get out of range
- _target_: torchvision.transforms.RandomCrop
size: 224
- _target_: torchvision.transforms.RandomHorizontalFlip
- _target_: torchvision.transforms.Normalize
mean: [0.485, 0.456, 0.406]
std: [0.229, 0.224, 0.225]
batch_transform: false