+datamodule_GrandPrix
This commit is contained in:
332
Grand_Prix/Training/openpifpaf_GrandPrix/GrandPrix_kp.py
Normal file
332
Grand_Prix/Training/openpifpaf_GrandPrix/GrandPrix_kp.py
Normal file
@@ -0,0 +1,332 @@
|
|||||||
|
"""
|
||||||
|
Interface for custom data.
|
||||||
|
|
||||||
|
This module handles datasets and is the class that you need to inherit from for your custom dataset.
|
||||||
|
This class gives you all the handles so that you can train with a new –dataset=mydataset.
|
||||||
|
The particular configuration of keypoints and skeleton is specified in the headmeta instances
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
import argparse
|
||||||
|
import torch
|
||||||
|
import numpy as np
|
||||||
|
try:
|
||||||
|
from pycocotools.coco import COCO
|
||||||
|
except ImportError:
|
||||||
|
COCO = None
|
||||||
|
|
||||||
|
from openpifpaf.datasets import DataModule
|
||||||
|
from openpifpaf import encoder, headmeta, metric, transforms
|
||||||
|
from openpifpaf.datasets import collate_images_anns_meta, collate_images_targets_meta
|
||||||
|
from openpifpaf.plugins.coco import CocoDataset as CocoLoader
|
||||||
|
|
||||||
|
from .constants import get_constants, training_weights_local_centrality
|
||||||
|
from .metrics import MeanPixelError
|
||||||
|
|
||||||
|
|
||||||
|
class GrandPrix_Kp(DataModule):
|
||||||
|
"""
|
||||||
|
DataModule for the GrandPrix Dataset.
|
||||||
|
"""
|
||||||
|
|
||||||
|
train_annotations = 'GrandPrix_dataset_v1_15/annotations/train.json'
|
||||||
|
val_annotations = 'GrandPrix_dataset_v1_15/annotations/val.json'
|
||||||
|
eval_annotations = val_annotations
|
||||||
|
train_image_dir = 'GrandPrix_dataset_v1_15/images/train/'
|
||||||
|
val_image_dir = 'GrandPrix_dataset_v1_15/images/val/'
|
||||||
|
eval_image_dir = val_image_dir
|
||||||
|
|
||||||
|
n_images = None
|
||||||
|
square_edge = 513
|
||||||
|
extended_scale = False
|
||||||
|
orientation_invariant = 0.0
|
||||||
|
blur = 0.0
|
||||||
|
augmentation = True
|
||||||
|
rescale_images = 1.0
|
||||||
|
upsample_stride = 1
|
||||||
|
min_kp_anns = 1
|
||||||
|
b_min = 1 # 1 pixel
|
||||||
|
|
||||||
|
eval_annotation_filter = True
|
||||||
|
eval_long_edge = 0 # set to zero to deactivate rescaling
|
||||||
|
eval_orientation_invariant = 0.0
|
||||||
|
eval_extended_scale = False
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
super().__init__()
|
||||||
|
if self.weights is not None:
|
||||||
|
caf_weights = []
|
||||||
|
for bone in self.CAR_SKELETON:
|
||||||
|
caf_weights.append(max(self.weights[bone[0] - 1],
|
||||||
|
self.weights[bone[1] - 1]))
|
||||||
|
w_np = np.array(caf_weights)
|
||||||
|
caf_weights = list(w_np / np.sum(w_np) * len(caf_weights))
|
||||||
|
else:
|
||||||
|
caf_weights = None
|
||||||
|
cif = headmeta.Cif('cif', 'apollo',
|
||||||
|
keypoints=self.CAR_KEYPOINTS,
|
||||||
|
sigmas=self.CAR_SIGMAS,
|
||||||
|
pose=self.CAR_POSE,
|
||||||
|
draw_skeleton=self.CAR_SKELETON,
|
||||||
|
score_weights=self.CAR_SCORE_WEIGHTS,
|
||||||
|
training_weights=self.weights)
|
||||||
|
caf = headmeta.Caf('caf', 'apollo',
|
||||||
|
keypoints=self.CAR_KEYPOINTS,
|
||||||
|
sigmas=self.CAR_SIGMAS,
|
||||||
|
pose=self.CAR_POSE,
|
||||||
|
skeleton=self.CAR_SKELETON,
|
||||||
|
training_weights=caf_weights)
|
||||||
|
|
||||||
|
cif.upsample_stride = self.upsample_stride
|
||||||
|
caf.upsample_stride = self.upsample_stride
|
||||||
|
self.head_metas = [cif, caf]
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def cli(cls, parser: argparse.ArgumentParser):
|
||||||
|
group = parser.add_argument_group('data module Apollo')
|
||||||
|
|
||||||
|
group.add_argument('--GrandPrix-train-annotations',
|
||||||
|
default=cls.train_annotations)
|
||||||
|
group.add_argument('--GrandPrix-val-annotations',
|
||||||
|
default=cls.val_annotations)
|
||||||
|
group.add_argument('--GrandPrix-train-image-dir',
|
||||||
|
default=cls.train_image_dir)
|
||||||
|
group.add_argument('--GrandPrix-val-image-dir',
|
||||||
|
default=cls.val_image_dir)
|
||||||
|
|
||||||
|
group.add_argument('--GrandPrix-square-edge',
|
||||||
|
default=cls.square_edge, type=int,
|
||||||
|
help='square edge of input images')
|
||||||
|
assert not cls.extended_scale
|
||||||
|
group.add_argument('--GrandPrix-extended-scale',
|
||||||
|
default=False, action='store_true',
|
||||||
|
help='augment with an extended scale range')
|
||||||
|
group.add_argument('--GrandPrix-orientation-invariant',
|
||||||
|
default=cls.orientation_invariant, type=float,
|
||||||
|
help='augment with random orientations')
|
||||||
|
group.add_argument('--GrandPrix-blur',
|
||||||
|
default=cls.blur, type=float,
|
||||||
|
help='augment with blur')
|
||||||
|
assert cls.augmentation
|
||||||
|
group.add_argument('--GrandPrix-no-augmentation',
|
||||||
|
dest='GrandPrix_augmentation',
|
||||||
|
default=True, action='store_false',
|
||||||
|
help='do not apply data augmentation')
|
||||||
|
group.add_argument('--GrandPrix-rescale-images',
|
||||||
|
default=cls.rescale_images, type=float,
|
||||||
|
help='overall rescale factor for images')
|
||||||
|
group.add_argument('--GrandPrix-upsample',
|
||||||
|
default=cls.upsample_stride, type=int,
|
||||||
|
help='head upsample stride')
|
||||||
|
group.add_argument('--GrandPrix-min-kp-anns',
|
||||||
|
default=cls.min_kp_anns, type=int,
|
||||||
|
help='filter images with fewer keypoint annotations')
|
||||||
|
group.add_argument('--GrandPrix-bmin',
|
||||||
|
default=cls.b_min, type=int,
|
||||||
|
help='b minimum in pixels')
|
||||||
|
group.add_argument('--GrandPrix-apply-local-centrality-weights',
|
||||||
|
dest='GrandPrix_apply_local_centrality',
|
||||||
|
default=False, action='store_true',
|
||||||
|
help='Weigh the CIF and CAF head during training.')
|
||||||
|
|
||||||
|
# evaluation
|
||||||
|
assert cls.eval_annotation_filter
|
||||||
|
group.add_argument('--GrandPrix-no-eval-annotation-filter',
|
||||||
|
dest='GrandPrix_eval_annotation_filter',
|
||||||
|
default=True, action='store_false')
|
||||||
|
group.add_argument('--GrandPrix-eval-long-edge', default=cls.eval_long_edge, type=int,
|
||||||
|
help='set to zero to deactivate rescaling')
|
||||||
|
assert not cls.eval_extended_scale
|
||||||
|
group.add_argument('--GrandPrix-eval-extended-scale', default=False, action='store_true')
|
||||||
|
group.add_argument('--GrandPrix-eval-orientation-invariant',
|
||||||
|
default=cls.eval_orientation_invariant, type=float)
|
||||||
|
group.add_argument('--GrandPrix-use-15-kps', default=False, action='store_true',
|
||||||
|
help=('The ApolloCar3D dataset can '
|
||||||
|
'be trained with 24 or 66 kps. If you want to train a model '
|
||||||
|
'with 24 kps activate this flag. Change the annotations '
|
||||||
|
'path to the json files with 24 kps.'))
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def configure(cls, args: argparse.Namespace):
|
||||||
|
# extract global information
|
||||||
|
cls.debug = args.debug
|
||||||
|
cls.pin_memory = args.pin_memory
|
||||||
|
|
||||||
|
# Apollo specific
|
||||||
|
cls.train_annotations = args.GrandPrix_train_annotations
|
||||||
|
cls.val_annotations = args.GrandPrix_val_annotations
|
||||||
|
cls.eval_annotations = cls.val_annotations
|
||||||
|
cls.train_image_dir = args.GrandPrix_train_image_dir
|
||||||
|
cls.val_image_dir = args.GrandPrix_val_image_dir
|
||||||
|
cls.eval_image_dir = cls.val_image_dir
|
||||||
|
|
||||||
|
cls.square_edge = args.GrandPrix_square_edge
|
||||||
|
cls.extended_scale = args.GrandPrix_extended_scale
|
||||||
|
cls.orientation_invariant = args.GrandPrix_orientation_invariant
|
||||||
|
cls.blur = args.GrandPrix_blur
|
||||||
|
cls.augmentation = args.GrandPrix_augmentation # loaded by the dest name
|
||||||
|
cls.rescale_images = args.GrandPrix_rescale_images
|
||||||
|
cls.upsample_stride = args.GrandPrix_upsample
|
||||||
|
cls.min_kp_anns = args.GrandPrix_min_kp_anns
|
||||||
|
cls.b_min = args.GrandPrix_bmin
|
||||||
|
if args.GrandPrix_use_15_kps:
|
||||||
|
(cls.CAR_KEYPOINTS, cls.CAR_SKELETON, cls.HFLIP, cls.CAR_SIGMAS, cls.CAR_POSE,
|
||||||
|
cls.CAR_CATEGORIES, cls.CAR_SCORE_WEIGHTS) = get_constants(24)
|
||||||
|
else:
|
||||||
|
(cls.CAR_KEYPOINTS, cls.CAR_SKELETON, cls.HFLIP, cls.CAR_SIGMAS, cls.CAR_POSE,
|
||||||
|
cls.CAR_CATEGORIES, cls.CAR_SCORE_WEIGHTS) = get_constants(66)
|
||||||
|
# evaluation
|
||||||
|
cls.eval_annotation_filter = args.GrandPrix_eval_annotation_filter
|
||||||
|
cls.eval_long_edge = args.GrandPrix_eval_long_edge
|
||||||
|
cls.eval_orientation_invariant = args.GrandPrix_eval_orientation_invariant
|
||||||
|
cls.eval_extended_scale = args.GrandPrix_eval_extended_scale
|
||||||
|
if args.GrandPrix_apply_local_centrality:
|
||||||
|
if args.GrandPrix_use_15_kps:
|
||||||
|
raise Exception("Applying local centrality weights only works with 66 kps.")
|
||||||
|
cls.weights = training_weights_local_centrality
|
||||||
|
else:
|
||||||
|
cls.weights = None
|
||||||
|
|
||||||
|
def _preprocess(self):
|
||||||
|
encoders = (encoder.Cif(self.head_metas[0], bmin=self.b_min),
|
||||||
|
encoder.Caf(self.head_metas[1], bmin=self.b_min))
|
||||||
|
|
||||||
|
if not self.augmentation:
|
||||||
|
return transforms.Compose([
|
||||||
|
transforms.NormalizeAnnotations(),
|
||||||
|
transforms.RescaleAbsolute(self.square_edge),
|
||||||
|
transforms.CenterPad(self.square_edge),
|
||||||
|
transforms.EVAL_TRANSFORM,
|
||||||
|
transforms.Encoders(encoders),
|
||||||
|
])
|
||||||
|
|
||||||
|
if self.extended_scale:
|
||||||
|
rescale_t = transforms.RescaleRelative(
|
||||||
|
scale_range=(0.2 * self.rescale_images,
|
||||||
|
2.0 * self.rescale_images),
|
||||||
|
power_law=True, stretch_range=(0.75, 1.33))
|
||||||
|
else:
|
||||||
|
rescale_t = transforms.RescaleRelative(
|
||||||
|
scale_range=(0.33 * self.rescale_images,
|
||||||
|
1.33 * self.rescale_images),
|
||||||
|
power_law=True, stretch_range=(0.75, 1.33))
|
||||||
|
|
||||||
|
return transforms.Compose([
|
||||||
|
transforms.NormalizeAnnotations(),
|
||||||
|
# transforms.AnnotationJitter(),
|
||||||
|
transforms.RandomApply(transforms.HFlip(self.CAR_KEYPOINTS, self.HFLIP), 0.5),
|
||||||
|
rescale_t,
|
||||||
|
transforms.RandomApply(transforms.Blur(), self.blur),
|
||||||
|
transforms.RandomChoice(
|
||||||
|
[transforms.RotateBy90(),
|
||||||
|
transforms.RotateUniform(30.0)],
|
||||||
|
[self.orientation_invariant, 0.2],
|
||||||
|
),
|
||||||
|
transforms.Crop(self.square_edge, use_area_of_interest=True),
|
||||||
|
transforms.CenterPad(self.square_edge),
|
||||||
|
transforms.MinSize(min_side=32.0),
|
||||||
|
transforms.TRAIN_TRANSFORM,
|
||||||
|
transforms.Encoders(encoders),
|
||||||
|
])
|
||||||
|
|
||||||
|
def train_loader(self):
|
||||||
|
train_data = CocoLoader(
|
||||||
|
image_dir=self.train_image_dir,
|
||||||
|
ann_file=self.train_annotations,
|
||||||
|
preprocess=self._preprocess(),
|
||||||
|
annotation_filter=True,
|
||||||
|
min_kp_anns=self.min_kp_anns,
|
||||||
|
category_ids=[1],
|
||||||
|
)
|
||||||
|
return torch.utils.data.DataLoader(
|
||||||
|
train_data, batch_size=self.batch_size, shuffle=not self.debug,
|
||||||
|
pin_memory=self.pin_memory, num_workers=self.loader_workers, drop_last=True,
|
||||||
|
collate_fn=collate_images_targets_meta)
|
||||||
|
|
||||||
|
def val_loader(self):
|
||||||
|
val_data = CocoLoader(
|
||||||
|
image_dir=self.val_image_dir,
|
||||||
|
ann_file=self.val_annotations,
|
||||||
|
preprocess=self._preprocess(),
|
||||||
|
annotation_filter=True,
|
||||||
|
min_kp_anns=self.min_kp_anns,
|
||||||
|
category_ids=[1],
|
||||||
|
)
|
||||||
|
return torch.utils.data.DataLoader(
|
||||||
|
val_data, batch_size=self.batch_size, shuffle=False,
|
||||||
|
pin_memory=self.pin_memory, num_workers=self.loader_workers, drop_last=True,
|
||||||
|
collate_fn=collate_images_targets_meta)
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def common_eval_preprocess(cls):
|
||||||
|
rescale_t = None
|
||||||
|
if cls.eval_extended_scale:
|
||||||
|
assert cls.eval_long_edge
|
||||||
|
rescale_t = [
|
||||||
|
transforms.DeterministicEqualChoice([
|
||||||
|
transforms.RescaleAbsolute(cls.eval_long_edge),
|
||||||
|
transforms.RescaleAbsolute((cls.eval_long_edge - 1) // 2 + 1),
|
||||||
|
], salt=1)
|
||||||
|
]
|
||||||
|
elif cls.eval_long_edge:
|
||||||
|
rescale_t = transforms.RescaleAbsolute(cls.eval_long_edge)
|
||||||
|
|
||||||
|
if cls.batch_size == 1:
|
||||||
|
padding_t = transforms.CenterPadTight(16)
|
||||||
|
else:
|
||||||
|
assert cls.eval_long_edge
|
||||||
|
padding_t = transforms.CenterPad(cls.eval_long_edge)
|
||||||
|
|
||||||
|
orientation_t = None
|
||||||
|
if cls.eval_orientation_invariant:
|
||||||
|
orientation_t = transforms.DeterministicEqualChoice([
|
||||||
|
None,
|
||||||
|
transforms.RotateBy90(fixed_angle=90),
|
||||||
|
transforms.RotateBy90(fixed_angle=180),
|
||||||
|
transforms.RotateBy90(fixed_angle=270),
|
||||||
|
], salt=3)
|
||||||
|
|
||||||
|
return [
|
||||||
|
transforms.NormalizeAnnotations(),
|
||||||
|
rescale_t,
|
||||||
|
padding_t,
|
||||||
|
orientation_t,
|
||||||
|
]
|
||||||
|
|
||||||
|
def _eval_preprocess(self):
|
||||||
|
return transforms.Compose([
|
||||||
|
*self.common_eval_preprocess(),
|
||||||
|
transforms.ToAnnotations([
|
||||||
|
transforms.ToKpAnnotations(
|
||||||
|
self.CAR_CATEGORIES,
|
||||||
|
keypoints_by_category={1: self.head_metas[0].keypoints},
|
||||||
|
skeleton_by_category={1: self.head_metas[1].skeleton},
|
||||||
|
),
|
||||||
|
transforms.ToCrowdAnnotations(self.CAR_CATEGORIES),
|
||||||
|
]),
|
||||||
|
transforms.EVAL_TRANSFORM,
|
||||||
|
])
|
||||||
|
|
||||||
|
def eval_loader(self):
|
||||||
|
eval_data = CocoLoader(
|
||||||
|
image_dir=self.eval_image_dir,
|
||||||
|
ann_file=self.eval_annotations,
|
||||||
|
preprocess=self._eval_preprocess(),
|
||||||
|
annotation_filter=self.eval_annotation_filter,
|
||||||
|
min_kp_anns=self.min_kp_anns if self.eval_annotation_filter else 0,
|
||||||
|
category_ids=[1] if self.eval_annotation_filter else [],
|
||||||
|
)
|
||||||
|
return torch.utils.data.DataLoader(
|
||||||
|
eval_data, batch_size=self.batch_size, shuffle=False,
|
||||||
|
pin_memory=self.pin_memory, num_workers=self.loader_workers, drop_last=False,
|
||||||
|
collate_fn=collate_images_anns_meta)
|
||||||
|
|
||||||
|
# TODO: make sure that 24kp flag is activated when evaluating a 24kp model
|
||||||
|
def metrics(self):
|
||||||
|
return [metric.Coco(
|
||||||
|
COCO(self.eval_annotations),
|
||||||
|
max_per_image=20,
|
||||||
|
category_ids=[1],
|
||||||
|
iou_type='keypoints',
|
||||||
|
keypoint_oks_sigmas=self.CAR_SIGMAS
|
||||||
|
), MeanPixelError()]
|
||||||
16
Grand_Prix/Training/openpifpaf_GrandPrix/__init__.py
Normal file
16
Grand_Prix/Training/openpifpaf_GrandPrix/__init__.py
Normal file
@@ -0,0 +1,16 @@
|
|||||||
|
import openpifpaf
|
||||||
|
|
||||||
|
from . import GrandPrix_kp
|
||||||
|
|
||||||
|
|
||||||
|
def register():
|
||||||
|
openpifpaf.DATAMODULES['GrandPrix'] = GrandPrix_kp.GrandPrix_Kp
|
||||||
|
openpifpaf.CHECKPOINT_URLS['shufflenetv2k16-apollo-24'] = \
|
||||||
|
"http://github.com/DuncanZauss/openpifpaf_assets/releases/" \
|
||||||
|
"download/v0.1.0/shufflenetv2k16-201113-135121-apollo.pkl.epoch290"
|
||||||
|
openpifpaf.CHECKPOINT_URLS['shufflenetv2k16-apollo-66'] = \
|
||||||
|
"http://github.com/DuncanZauss/openpifpaf_assets/releases/" \
|
||||||
|
"download/v0.1.0/sk16_apollo_66kp.pkl"
|
||||||
|
openpifpaf.CHECKPOINT_URLS['shufflenetv2k30-apollo-66'] = \
|
||||||
|
"http://github.com/DuncanZauss/openpifpaf_assets/releases/" \
|
||||||
|
"download/v0.1.0/sk30_apollo_66kp.pkl"
|
||||||
272
Grand_Prix/Training/openpifpaf_GrandPrix/constants.py
Normal file
272
Grand_Prix/Training/openpifpaf_GrandPrix/constants.py
Normal file
@@ -0,0 +1,272 @@
|
|||||||
|
import os
|
||||||
|
|
||||||
|
import numpy as np
|
||||||
|
try:
|
||||||
|
import matplotlib.cm as mplcm
|
||||||
|
from matplotlib.animation import FuncAnimation
|
||||||
|
from mpl_toolkits.mplot3d import Axes3D
|
||||||
|
except ImportError:
|
||||||
|
pass
|
||||||
|
|
||||||
|
import openpifpaf
|
||||||
|
|
||||||
|
CAR_KEYPOINTS_24 = list(range(1, 16))
|
||||||
|
|
||||||
|
CAR_SKELETON_24 = [[1, 2],
|
||||||
|
[2, 14],
|
||||||
|
[14, 9],
|
||||||
|
[2, 3],
|
||||||
|
[14, 5],
|
||||||
|
[3, 5],
|
||||||
|
[3, 4],
|
||||||
|
[4, 6],
|
||||||
|
[6, 7],
|
||||||
|
[7, 8],
|
||||||
|
[8, 9],
|
||||||
|
[5, 8],
|
||||||
|
[9, 10],
|
||||||
|
[10, 11],
|
||||||
|
[11, 12],
|
||||||
|
[12, 13],
|
||||||
|
[13, 11],
|
||||||
|
[13, 1],
|
||||||
|
[11, 1],
|
||||||
|
[5, 7],
|
||||||
|
[8, 15],
|
||||||
|
[7, 15]]
|
||||||
|
|
||||||
|
CAR_CATEGORIES_24 = ['GrandPrix']
|
||||||
|
|
||||||
|
CAR_SCORE_WEIGHTS_24 = [1.0]*len(CAR_KEYPOINTS_24)
|
||||||
|
|
||||||
|
CAR_SIGMAS_24 = [0.05] * len(CAR_KEYPOINTS_24)
|
||||||
|
|
||||||
|
CAR_POSE_24 = np.array([[-7.49196479e+00, 2.52538184e-01, 0.00000000e+00],
|
||||||
|
[ 7.49196479e+00, 8.41793948e-02, 0.00000000e+00],
|
||||||
|
[ 2.53379934e+01, 1.33845214e+01, 0.00000000e+00],
|
||||||
|
[ 3.16514470e+01, 2.54221725e+01, 0.00000000e+00],
|
||||||
|
[ 2.48329184e+01, 1.40579572e+01, 1.25999990e+01],
|
||||||
|
[ 3.14830882e+01, 2.55905313e+01, 2.65859990e+01],
|
||||||
|
[ 2.51696353e+01, 1.33845214e+01, 3.07859970e+01],
|
||||||
|
[ 2.45803806e+01, -4.78138895e+01, 1.39859990e+01],
|
||||||
|
[ 1.17009335e+01, -4.89082199e+01, 9.78599900e+00],
|
||||||
|
[-7.23942728e+00, -4.86556828e+01, 9.78599900e+00],
|
||||||
|
[-3.21565248e+01, -4.86556834e+01, 9.78599900e+00],
|
||||||
|
[-4.99183727e+01, 6.65016983e+00, 1.05000000e+01],
|
||||||
|
[-3.94801318e+01, 1.17009349e+01, 1.05000000e+01],
|
||||||
|
[ 9.25973199e+00, -7.57614426e+00, 5.04000000e+00],
|
||||||
|
[ 3.40084714e+01, -1.02025414e+02, 6.29999970e+00]])
|
||||||
|
|
||||||
|
HFLIP_24 = {
|
||||||
|
'1': '1',
|
||||||
|
'2': '2',
|
||||||
|
'3': '3',
|
||||||
|
'4': '4',
|
||||||
|
'5': '5',
|
||||||
|
'6': '6',
|
||||||
|
'7': '7',
|
||||||
|
'8': '8',
|
||||||
|
'9': '9',
|
||||||
|
'10': '10',
|
||||||
|
'11': '11',
|
||||||
|
'12': '12',
|
||||||
|
'13': '13',
|
||||||
|
'14': '14',
|
||||||
|
'15': '15',
|
||||||
|
'16': '16',
|
||||||
|
'17': '17',
|
||||||
|
'18': '18',
|
||||||
|
'19': '19',
|
||||||
|
'20': '20',
|
||||||
|
'21': '21',
|
||||||
|
'22': '22'
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
training_weights_local_centrality = [
|
||||||
|
0.890968488270775,
|
||||||
|
0.716506138617812,
|
||||||
|
1.05674590410869,
|
||||||
|
0.764774195768455,
|
||||||
|
0.637682585483328,
|
||||||
|
0.686680807728366,
|
||||||
|
0.955422595797394,
|
||||||
|
0.936714585642375,
|
||||||
|
1.34823795445326,
|
||||||
|
1.38308992581967,
|
||||||
|
1.32689945125819,
|
||||||
|
1.38838655605483,
|
||||||
|
1.18980184904613,
|
||||||
|
1.02584355494795,
|
||||||
|
0.90969156732068,
|
||||||
|
1.24732068576104,
|
||||||
|
1.11338768064342,
|
||||||
|
0.933815217550391,
|
||||||
|
0.852297518872114,
|
||||||
|
1.04167641424727,
|
||||||
|
1.01668968075247,
|
||||||
|
1.34625964088011,
|
||||||
|
0.911796331039028,
|
||||||
|
0.866206536337413,
|
||||||
|
1.55957820407853,
|
||||||
|
0.730844382675724,
|
||||||
|
0.651138644197359,
|
||||||
|
0.758018559633786,
|
||||||
|
1.31842501396691,
|
||||||
|
1.32186116654782,
|
||||||
|
0.744347016851606,
|
||||||
|
0.636390683664723,
|
||||||
|
0.715244950821949,
|
||||||
|
1.63122349407032,
|
||||||
|
0.849835699185461,
|
||||||
|
0.910488007220499,
|
||||||
|
1.44244151650561,
|
||||||
|
1.14150437331681,
|
||||||
|
1.19808610191343,
|
||||||
|
0.960186788642886,
|
||||||
|
1.05023623286937,
|
||||||
|
1.19761709710598,
|
||||||
|
1.3872216313401,
|
||||||
|
1.01256700741214,
|
||||||
|
1.1167909667759,
|
||||||
|
1.27893496336199,
|
||||||
|
1.54475684725655,
|
||||||
|
1.40343733870633,
|
||||||
|
1.45552060866114,
|
||||||
|
1.47264222155031,
|
||||||
|
0.970060423999993,
|
||||||
|
0.944450314768933,
|
||||||
|
0.623987071240172,
|
||||||
|
0.5745237907704,
|
||||||
|
0.66890646050993,
|
||||||
|
0.978411632994504,
|
||||||
|
0.587396395188292,
|
||||||
|
0.76307999741129,
|
||||||
|
0.609793563449648,
|
||||||
|
0.67983566494545,
|
||||||
|
0.685883538168462,
|
||||||
|
0.753587600664775,
|
||||||
|
0.770335133588157,
|
||||||
|
0.764713638033368,
|
||||||
|
0.792364155965385,
|
||||||
|
0.796435233566833
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def get_constants(num_kps):
|
||||||
|
if num_kps == 24:
|
||||||
|
CAR_POSE_24[:, 2] = 2.0
|
||||||
|
return [CAR_KEYPOINTS_24, CAR_SKELETON_24, HFLIP_24, CAR_SIGMAS_24,
|
||||||
|
CAR_POSE_24, CAR_CATEGORIES_24, CAR_SCORE_WEIGHTS_24]
|
||||||
|
if num_kps == 66:
|
||||||
|
CAR_POSE_66[:, 2] = 2.0
|
||||||
|
return [CAR_KEYPOINTS_66, CAR_SKELETON_66, HFLIP_66, CAR_SIGMAS_66,
|
||||||
|
CAR_POSE_66, CAR_CATEGORIES_66, CAR_SCORE_WEIGHTS_66]
|
||||||
|
# using no if-elif-else construction due to pylint no-else-return error
|
||||||
|
raise Exception("Only poses with 24 or 66 keypoints are available.")
|
||||||
|
|
||||||
|
|
||||||
|
def draw_ann(ann, *, keypoint_painter, filename=None, margin=0.5, aspect=None, **kwargs):
|
||||||
|
from openpifpaf import show # pylint: disable=import-outside-toplevel
|
||||||
|
|
||||||
|
bbox = ann.bbox()
|
||||||
|
xlim = bbox[0] - margin, bbox[0] + bbox[2] + margin
|
||||||
|
ylim = bbox[1] - margin, bbox[1] + bbox[3] + margin
|
||||||
|
if aspect == 'equal':
|
||||||
|
fig_w = 5.0
|
||||||
|
else:
|
||||||
|
fig_w = 5.0 / (ylim[1] - ylim[0]) * (xlim[1] - xlim[0])
|
||||||
|
|
||||||
|
with show.canvas(filename, figsize=(fig_w, 5), nomargin=True, **kwargs) as ax:
|
||||||
|
ax.set_axis_off()
|
||||||
|
ax.set_xlim(*xlim)
|
||||||
|
ax.set_ylim(*ylim)
|
||||||
|
|
||||||
|
if aspect is not None:
|
||||||
|
ax.set_aspect(aspect)
|
||||||
|
|
||||||
|
keypoint_painter.annotation(ax, ann)
|
||||||
|
|
||||||
|
|
||||||
|
def draw_skeletons(pose, sigmas, skel, kps, scr_weights):
|
||||||
|
from openpifpaf.annotation import Annotation # pylint: disable=import-outside-toplevel
|
||||||
|
from openpifpaf import show # pylint: disable=import-outside-toplevel
|
||||||
|
|
||||||
|
scale = np.sqrt(
|
||||||
|
(np.max(pose[:, 0]) - np.min(pose[:, 0]))
|
||||||
|
* (np.max(pose[:, 1]) - np.min(pose[:, 1]))
|
||||||
|
)
|
||||||
|
|
||||||
|
show.KeypointPainter.show_joint_scales = True
|
||||||
|
keypoint_painter = show.KeypointPainter()
|
||||||
|
ann = Annotation(keypoints=kps, skeleton=skel, score_weights=scr_weights)
|
||||||
|
ann.set(pose, np.array(sigmas) * scale)
|
||||||
|
os.makedirs('docs', exist_ok=True)
|
||||||
|
draw_ann(ann, filename='docs/skeleton_car.png', keypoint_painter=keypoint_painter)
|
||||||
|
|
||||||
|
|
||||||
|
def plot3d_red(ax_2D, p3d, skeleton):
|
||||||
|
skeleton = [(bone[0] - 1, bone[1] - 1) for bone in skeleton]
|
||||||
|
|
||||||
|
rot_p90_x = np.array([[1, 0, 0], [0, 0, 1], [0, 1, 0]])
|
||||||
|
p3d = p3d @ rot_p90_x
|
||||||
|
|
||||||
|
fig = ax_2D.get_figure()
|
||||||
|
ax = Axes3D(fig, auto_add_to_figure=False)
|
||||||
|
fig.add_axes(ax)
|
||||||
|
ax.set_axis_off()
|
||||||
|
ax_2D.set_axis_off()
|
||||||
|
|
||||||
|
ax.view_init(azim=-90, elev=20)
|
||||||
|
ax.set_xlabel('X')
|
||||||
|
ax.set_ylabel('Y')
|
||||||
|
ax.set_zlabel('Z')
|
||||||
|
max_range = np.array([p3d[:, 0].max() - p3d[:, 0].min(),
|
||||||
|
p3d[:, 1].max() - p3d[:, 1].min(),
|
||||||
|
p3d[:, 2].max() - p3d[:, 2].min()]).max() / 2.0
|
||||||
|
mid_x = (p3d[:, 0].max() + p3d[:, 0].min()) * 0.5
|
||||||
|
mid_y = (p3d[:, 1].max() + p3d[:, 1].min()) * 0.5
|
||||||
|
mid_z = (p3d[:, 2].max() + p3d[:, 2].min()) * 0.5
|
||||||
|
|
||||||
|
ax.set_xlim(mid_x - max_range, mid_x + max_range)
|
||||||
|
ax.set_ylim(mid_y - max_range, mid_y + max_range)
|
||||||
|
ax.set_zlim(mid_z - max_range, mid_z + max_range) # pylint: disable=no-member
|
||||||
|
|
||||||
|
for ci, bone in enumerate(skeleton):
|
||||||
|
c = mplcm.get_cmap('tab20')((ci % 20 + 0.05) / 20) # Same coloring as Pifpaf preds
|
||||||
|
ax.plot(p3d[bone, 0], p3d[bone, 1], p3d[bone, 2], color=c)
|
||||||
|
|
||||||
|
def animate(i):
|
||||||
|
ax.view_init(elev=10., azim=i)
|
||||||
|
return fig
|
||||||
|
|
||||||
|
return FuncAnimation(fig, animate, frames=360, interval=100)
|
||||||
|
|
||||||
|
|
||||||
|
def print_associations():
|
||||||
|
print("\nAssociations of the car skeleton with 24 keypoints")
|
||||||
|
for j1, j2 in CAR_SKELETON_24:
|
||||||
|
print(CAR_KEYPOINTS_24[j1 - 1], '-', CAR_KEYPOINTS_24[j2 - 1])
|
||||||
|
print("\nAssociations of the car skeleton with 66 keypoints")
|
||||||
|
for j1, j2 in CAR_SKELETON_66:
|
||||||
|
print(CAR_KEYPOINTS_66[j1 - 1], '-', CAR_KEYPOINTS_66[j2 - 1])
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
# print_associations()
|
||||||
|
# =============================================================================
|
||||||
|
# draw_skeletons(CAR_POSE_24, sigmas = CAR_SIGMAS_24, skel = CAR_SKELETON_24,
|
||||||
|
# kps = CAR_KEYPOINTS_24, scr_weights = CAR_SCORE_WEIGHTS_24)
|
||||||
|
# draw_skeletons(CAR_POSE_66, sigmas = CAR_SIGMAS_66, skel = CAR_SKELETON_66,
|
||||||
|
# kps = CAR_KEYPOINTS_66, scr_weights = CAR_SCORE_WEIGHTS_66)
|
||||||
|
# =============================================================================
|
||||||
|
# with openpifpaf.show.Canvas.blank(nomargin=True) as ax_2D:
|
||||||
|
# anim_66 = plot3d_red(ax_2D, CAR_POSE_66, CAR_SKELETON_66)
|
||||||
|
# anim_66.save('./CAR_66_Pose.gif', fps=30)
|
||||||
|
with openpifpaf.show.Canvas.blank(nomargin=True) as ax_2D:
|
||||||
|
anim_24 = plot3d_red(ax_2D, CAR_POSE_24, CAR_SKELETON_24)
|
||||||
|
anim_24.save('./CAR_24_Pose.gif', fps=30)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
main()
|
||||||
131
Grand_Prix/Training/openpifpaf_GrandPrix/metrics.py
Normal file
131
Grand_Prix/Training/openpifpaf_GrandPrix/metrics.py
Normal file
@@ -0,0 +1,131 @@
|
|||||||
|
import logging
|
||||||
|
|
||||||
|
import numpy as np
|
||||||
|
|
||||||
|
from openpifpaf.metric.base import Base
|
||||||
|
from openpifpaf.annotation import Annotation
|
||||||
|
|
||||||
|
try:
|
||||||
|
import scipy
|
||||||
|
except ImportError:
|
||||||
|
scipy = None
|
||||||
|
|
||||||
|
LOG = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
class MeanPixelError(Base):
|
||||||
|
"""
|
||||||
|
Calculate mean pixel error and detection rate for a given image
|
||||||
|
and category in an "all-vs-all setting"
|
||||||
|
"""
|
||||||
|
predictions = []
|
||||||
|
image_ids = []
|
||||||
|
errors = [] # mean pixel errors
|
||||||
|
detections = [] # detection rate
|
||||||
|
errors_scaled = [] # mean pixel errors
|
||||||
|
detections_scaled = [] # detection rate
|
||||||
|
px_ref = 368 # CPM crop size in pixels
|
||||||
|
|
||||||
|
def accumulate(self, predictions, image_meta, *, ground_truth=None):
|
||||||
|
errors = []
|
||||||
|
detections = []
|
||||||
|
errors_scaled = []
|
||||||
|
detections_scaled = []
|
||||||
|
|
||||||
|
# Filter ground-truth
|
||||||
|
for annotation in ground_truth:
|
||||||
|
if not isinstance(annotation, Annotation):
|
||||||
|
continue
|
||||||
|
indices_gt = np.nonzero(annotation.data[:, 2] > 1.0)
|
||||||
|
if indices_gt[0].size <= 3:
|
||||||
|
continue
|
||||||
|
gts = annotation.data[indices_gt, 0:2].squeeze()
|
||||||
|
width = float(annotation.fixed_bbox[2])
|
||||||
|
height = float(annotation.fixed_bbox[3])
|
||||||
|
scale = np.array([self.px_ref / width, self.px_ref / height]).reshape(1, 2)
|
||||||
|
|
||||||
|
# Evaluate each keypoint
|
||||||
|
for idx, gt in zip(indices_gt[0], gts):
|
||||||
|
preds = np.array([p.data[idx] for p in predictions]).reshape(-1, 3)[:, 0:2]
|
||||||
|
if preds.size <= 0:
|
||||||
|
continue
|
||||||
|
i = np.argmin(np.linalg.norm(preds - gt, axis=1))
|
||||||
|
dist = preds[i:i + 1] - gt
|
||||||
|
dist_scaled = dist * scale
|
||||||
|
d = float(np.linalg.norm(dist, axis=1))
|
||||||
|
d_scaled = float(np.linalg.norm(dist_scaled, axis=1))
|
||||||
|
|
||||||
|
# Prediction correct if error less than 10 pixels
|
||||||
|
if d < 10:
|
||||||
|
errors.append(d)
|
||||||
|
detections.append(1)
|
||||||
|
else:
|
||||||
|
detections.append(0)
|
||||||
|
if d_scaled < 10:
|
||||||
|
errors_scaled.append(d)
|
||||||
|
detections_scaled.append(1)
|
||||||
|
else:
|
||||||
|
detections_scaled.append(0)
|
||||||
|
|
||||||
|
# Stats for a single image
|
||||||
|
mpe = average(errors)
|
||||||
|
mpe_scaled = average(errors_scaled)
|
||||||
|
det_rate = 100 * average(detections)
|
||||||
|
det_rate_scaled = 100 * average(detections_scaled)
|
||||||
|
LOG.info('Mean Pixel Error (scaled): %s (%s) Det. Rate (scaled): %s (%s)',
|
||||||
|
str(mpe)[:4], str(mpe_scaled)[:4], str(det_rate)[:4], str(det_rate_scaled)[:4])
|
||||||
|
|
||||||
|
# Accumulate stats
|
||||||
|
self.errors.extend(errors)
|
||||||
|
self.detections.extend(detections)
|
||||||
|
self.errors_scaled.extend(errors_scaled)
|
||||||
|
self.detections_scaled.extend(detections_scaled)
|
||||||
|
|
||||||
|
def write_predictions(self, filename, *, additional_data=None):
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
|
def stats(self):
|
||||||
|
mpe = average(self.errors)
|
||||||
|
mpe_scaled = average(self.errors_scaled)
|
||||||
|
det_rate = 100 * average(self.detections)
|
||||||
|
det_rate_scaled = 100 * average(self.detections_scaled)
|
||||||
|
LOG.info('Final Results: \nMean Pixel Error [scaled] : %f [%f] '
|
||||||
|
'\nDetection Rate [scaled]: %f [%f]',
|
||||||
|
mpe, mpe_scaled, det_rate, det_rate_scaled)
|
||||||
|
data = {
|
||||||
|
'stats': [mpe, mpe_scaled, det_rate, det_rate_scaled],
|
||||||
|
'text_labels': ['Mean Pixel Error',
|
||||||
|
'Mean Pixel Error Scaled',
|
||||||
|
'Detection Rate [%]',
|
||||||
|
'Detection Rate Scaled[%]'],
|
||||||
|
}
|
||||||
|
return data
|
||||||
|
|
||||||
|
|
||||||
|
def hungarian_matching(gts, predictions, thresh=0.5):
|
||||||
|
cost = np.zeros((len(gts), len(predictions)))
|
||||||
|
|
||||||
|
for i, (dg, vg) in enumerate(gts):
|
||||||
|
for j, pred in enumerate(predictions):
|
||||||
|
p = np.array(pred.data)
|
||||||
|
dp = p[:, 0:2][vg > 1.0]
|
||||||
|
vp = p[:, 2][vg > 1.0]
|
||||||
|
|
||||||
|
dp[vp < thresh] = -100
|
||||||
|
dp[vp < thresh] = -100
|
||||||
|
|
||||||
|
# measure the per-keypoint distance
|
||||||
|
distances = np.clip(np.linalg.norm(dp - dg, axis=1), 0, 10)
|
||||||
|
cost[i, j] = float(np.mean(distances))
|
||||||
|
|
||||||
|
assert np.max(cost) < 11
|
||||||
|
row, cols = scipy.optimize.linear_sum_assignment(cost)
|
||||||
|
return row, cols, cost
|
||||||
|
|
||||||
|
|
||||||
|
def average(my_list, *, empty_value=0.0):
|
||||||
|
"""calculate mean of a list"""
|
||||||
|
if not my_list:
|
||||||
|
return empty_value
|
||||||
|
|
||||||
|
return sum(my_list) / float(len(my_list))
|
||||||
Reference in New Issue
Block a user