| /* |
| * Copyright (c) 2016, Alliance for Open Media. All rights reserved |
| * |
| * This source code is subject to the terms of the BSD 2 Clause License and |
| * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License |
| * was not distributed with this source code in the LICENSE file, you can |
| * obtain it at www.aomedia.org/license/software. If the Alliance for Open |
| * Media Patent License 1.0 was not distributed with this source code in the |
| * PATENTS file, you can obtain it at www.aomedia.org/license/patent. |
| */ |
| |
| #ifndef AOM_AOM_DSP_FLOW_ESTIMATION_H_ |
| #define AOM_AOM_DSP_FLOW_ESTIMATION_H_ |
| |
| #include "aom_dsp/pyramid.h" |
| #include "aom_dsp/flow_estimation/corner_detect.h" |
| #include "aom_ports/mem.h" |
| #include "aom_scale/yv12config.h" |
| |
| #ifdef __cplusplus |
| extern "C" { |
| #endif |
| |
| #define MAX_PARAMDIM 6 |
| #define MIN_INLIER_PROB 0.1 |
| |
| /* clang-format off */ |
| enum { |
| IDENTITY = 0, // identity transformation, 0-parameter |
| TRANSLATION = 1, // translational motion 2-parameter |
| ROTZOOM = 2, // simplified affine with rotation + zoom only, 4-parameter |
| AFFINE = 3, // affine, 6-parameter |
| TRANS_TYPES, |
| } UENUM1BYTE(TransformationType); |
| /* clang-format on */ |
| |
| // number of parameters used by each transformation in TransformationTypes |
| static const int trans_model_params[TRANS_TYPES] = { 0, 2, 4, 6 }; |
| |
| // Available methods which can be used for global motion estimation |
| typedef enum { |
| GLOBAL_MOTION_METHOD_FEATURE_MATCH, |
| GLOBAL_MOTION_METHOD_DISFLOW, |
| GLOBAL_MOTION_METHOD_LAST = GLOBAL_MOTION_METHOD_DISFLOW, |
| GLOBAL_MOTION_METHODS |
| } GlobalMotionMethod; |
| |
| typedef struct { |
| double params[MAX_PARAMDIM]; |
| int *inliers; |
| int num_inliers; |
| } MotionModel; |
| |
| // Data structure to store a single correspondence point during global |
| // motion search. |
| // |
| // A correspondence (x, y) -> (rx, ry) means that point (x, y) in the |
| // source frame corresponds to point (rx, ry) in the ref frame. |
| typedef struct { |
| double x, y; |
| double rx, ry; |
| } Correspondence; |
| |
| // For each global motion method, how many pyramid levels should we allocate? |
| // Note that this is a maximum, and fewer levels will be allocated if the frame |
| // is not large enough to need all of the specified levels |
| extern const int global_motion_pyr_levels[GLOBAL_MOTION_METHODS]; |
| |
| // Which global motion method should we use in practice? |
| // Disflow is both faster and gives better results than feature matching in |
| // practically all cases, so we use disflow by default |
| static const GlobalMotionMethod default_global_motion_method = |
| GLOBAL_MOTION_METHOD_DISFLOW; |
| |
| extern const double kIdentityParams[MAX_PARAMDIM]; |
| |
| // Compute a global motion model between the given source and ref frames. |
| // |
| // As is standard for video codecs, the resulting model maps from (x, y) |
| // coordinates in `src` to the corresponding points in `ref`, regardless |
| // of the temporal order of the two frames. |
| // |
| // Returns true if global motion estimation succeeded, false if not. |
| // The output models should only be used if this function succeeds. |
| bool aom_compute_global_motion(TransformationType type, YV12_BUFFER_CONFIG *src, |
| YV12_BUFFER_CONFIG *ref, int bit_depth, |
| GlobalMotionMethod gm_method, |
| MotionModel *motion_models, |
| int num_motion_models, bool *mem_alloc_failed); |
| |
| #ifdef __cplusplus |
| } |
| #endif |
| |
| #endif // AOM_AOM_DSP_FLOW_ESTIMATION_H_ |