2018-05-06 08:19:09 +01:00
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
|
|
// Copyright (C) 2016, Carnegie Mellon University and University of Cambridge,
|
|
|
|
|
|
// all rights reserved.
|
|
|
|
|
|
//
|
|
|
|
|
|
// THIS SOFTWARE IS PROVIDED <20>AS IS<49> FOR ACADEMIC USE ONLY AND ANY EXPRESS
|
|
|
|
|
|
// OR IMPLIED WARRANTIES WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
|
|
|
|
|
|
// THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
|
|
|
|
// PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS
|
|
|
|
|
|
// BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY.
|
|
|
|
|
|
// OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
|
|
|
|
// SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
|
|
// HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
|
|
|
|
|
|
// STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
|
|
|
|
|
|
// ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
|
|
|
|
// POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
|
|
//
|
|
|
|
|
|
// Notwithstanding the license granted herein, Licensee acknowledges that certain components
|
|
|
|
|
|
// of the Software may be covered by so-called <20>open source<63> software licenses (<28>Open Source
|
|
|
|
|
|
// Components<74>), which means any software licenses approved as open source licenses by the
|
|
|
|
|
|
// Open Source Initiative or any substantially similar licenses, including without limitation any
|
|
|
|
|
|
// license that, as a condition of distribution of the software licensed under such license,
|
|
|
|
|
|
// requires that the distributor make the software available in source code format. Licensor shall
|
|
|
|
|
|
// provide a list of Open Source Components for a particular version of the Software upon
|
|
|
|
|
|
// Licensee<65>s request. Licensee will comply with the applicable terms of such licenses and to
|
|
|
|
|
|
// the extent required by the licenses covering Open Source Components, the terms of such
|
|
|
|
|
|
// licenses will apply in lieu of the terms of this Agreement. To the extent the terms of the
|
|
|
|
|
|
// licenses applicable to Open Source Components prohibit any of the restrictions in this
|
|
|
|
|
|
// License Agreement with respect to such Open Source Component, such restrictions will not
|
|
|
|
|
|
// apply to such Open Source Component. To the extent the terms of the licenses applicable to
|
|
|
|
|
|
// Open Source Components require Licensor to make an offer to provide source code or
|
|
|
|
|
|
// related information in connection with the Software, such offer is hereby made. Any request
|
|
|
|
|
|
// for source code or related information should be directed to cl-face-tracker-distribution@lists.cam.ac.uk
|
|
|
|
|
|
// Licensee acknowledges receipt of notices for the Open Source Components for the initial
|
|
|
|
|
|
// delivery of the Software.
|
|
|
|
|
|
|
|
|
|
|
|
// * Any publications arising from the use of this software, including but
|
|
|
|
|
|
// not limited to academic journal and conference publications, technical
|
|
|
|
|
|
// reports and manuals, must cite at least one of the following works:
|
|
|
|
|
|
//
|
|
|
|
|
|
// OpenFace: an open source facial behavior analysis toolkit
|
|
|
|
|
|
// Tadas Baltru<72>aitis, Peter Robinson, and Louis-Philippe Morency
|
|
|
|
|
|
// in IEEE Winter Conference on Applications of Computer Vision, 2016
|
|
|
|
|
|
//
|
|
|
|
|
|
// Rendering of Eyes for Eye-Shape Registration and Gaze Estimation
|
|
|
|
|
|
// Erroll Wood, Tadas Baltru<72>aitis, Xucong Zhang, Yusuke Sugano, Peter Robinson, and Andreas Bulling
|
|
|
|
|
|
// in IEEE International. Conference on Computer Vision (ICCV), 2015
|
|
|
|
|
|
//
|
|
|
|
|
|
// Cross-dataset learning and person-speci?c normalisation for automatic Action Unit detection
|
|
|
|
|
|
// Tadas Baltru<72>aitis, Marwa Mahmoud, and Peter Robinson
|
|
|
|
|
|
// in Facial Expression Recognition and Analysis Challenge,
|
|
|
|
|
|
// IEEE International Conference on Automatic Face and Gesture Recognition, 2015
|
|
|
|
|
|
//
|
|
|
|
|
|
// Constrained Local Neural Fields for robust facial landmark detection in the wild.
|
|
|
|
|
|
// Tadas Baltru<72>aitis, Peter Robinson, and Louis-Philippe Morency.
|
|
|
|
|
|
// in IEEE Int. Conference on Computer Vision Workshops, 300 Faces in-the-Wild Challenge, 2013.
|
|
|
|
|
|
//
|
|
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
|
|
|
|
|
|
|
|
#include "stdafx.h"
|
|
|
|
|
|
|
|
|
|
|
|
#include "CEN_patch_expert.h"
|
|
|
|
|
|
|
|
|
|
|
|
// OpenCV includes
|
|
|
|
|
|
#include <opencv2/core/core.hpp>
|
|
|
|
|
|
#include <opencv2/imgproc.hpp>
|
|
|
|
|
|
|
|
|
|
|
|
// Local includes
|
|
|
|
|
|
#include "LandmarkDetectorUtils.h"
|
|
|
|
|
|
|
|
|
|
|
|
// For exponential
|
|
|
|
|
|
#include <math.h>
|
|
|
|
|
|
|
|
|
|
|
|
using namespace LandmarkDetector;
|
|
|
|
|
|
|
|
|
|
|
|
// Copy constructor
|
2018-05-07 21:54:52 +01:00
|
|
|
|
CEN_patch_expert::CEN_patch_expert(const CEN_patch_expert& other) : confidence(other.confidence), width_support(other.width_support), height_support(other.height_support)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
|
|
|
|
|
|
// Copy the layer weights in a deep way
|
|
|
|
|
|
for (size_t i = 0; i < other.weights.size(); ++i)
|
|
|
|
|
|
{
|
|
|
|
|
|
this->weights.push_back(other.weights[i].clone());
|
|
|
|
|
|
this->biases.push_back(other.biases[i].clone());
|
|
|
|
|
|
this->activation_function.push_back(other.activation_function[i]);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//===========================================================================
|
|
|
|
|
|
void CEN_patch_expert::Read(ifstream &stream)
|
|
|
|
|
|
{
|
|
|
|
|
|
|
|
|
|
|
|
// Setting up OpenBLAS
|
|
|
|
|
|
goto_set_num_threads(1);
|
|
|
|
|
|
openblas_set_num_threads(1);
|
|
|
|
|
|
|
|
|
|
|
|
// Sanity check
|
|
|
|
|
|
int read_type;
|
|
|
|
|
|
|
|
|
|
|
|
stream.read((char*)&read_type, 4);
|
|
|
|
|
|
assert(read_type == 6);
|
|
|
|
|
|
|
|
|
|
|
|
// the number of neurons for this patch
|
|
|
|
|
|
int num_layers;
|
2018-05-07 21:54:52 +01:00
|
|
|
|
stream.read((char*)&width_support, 4);
|
|
|
|
|
|
stream.read((char*)&height_support, 4);
|
2018-05-06 08:19:09 +01:00
|
|
|
|
stream.read((char*)&num_layers, 4);
|
|
|
|
|
|
|
|
|
|
|
|
if (num_layers == 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
// empty patch due to landmark being invisible at that orientation (or visible through mirroring)
|
|
|
|
|
|
stream.read((char*)&confidence, 8);
|
|
|
|
|
|
return;
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
activation_function.resize(num_layers);
|
|
|
|
|
|
weights.resize(num_layers);
|
|
|
|
|
|
biases.resize(num_layers);
|
|
|
|
|
|
|
|
|
|
|
|
for (int i = 0; i < num_layers; i++)
|
|
|
|
|
|
{
|
|
|
|
|
|
int neuron_type;
|
|
|
|
|
|
stream.read((char*)&neuron_type, 4);
|
|
|
|
|
|
activation_function[i] = neuron_type;
|
|
|
|
|
|
|
|
|
|
|
|
cv::Mat_<double> bias;
|
|
|
|
|
|
LandmarkDetector::ReadMatBin(stream, bias);
|
|
|
|
|
|
|
|
|
|
|
|
cv::Mat_<double> weight;
|
|
|
|
|
|
LandmarkDetector::ReadMatBin(stream, weight);
|
|
|
|
|
|
|
|
|
|
|
|
weights[i] = weight;
|
|
|
|
|
|
biases[i] = bias;
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Read the patch confidence
|
|
|
|
|
|
stream.read((char*)&confidence, 8);
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Contrast normalize the input for response map computation
|
|
|
|
|
|
void contrastNorm(const cv::Mat_<float>& input, cv::Mat_<float>& output)
|
|
|
|
|
|
{
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int num_cols = input.cols;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int num_rows = input.rows;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
output = input.clone();
|
|
|
|
|
|
|
|
|
|
|
|
cv::MatConstIterator_<float> p = input.begin();
|
|
|
|
|
|
|
|
|
|
|
|
// Compute row wise
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int y = 0; y < num_rows; ++y)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
|
|
|
|
|
|
cv::Scalar mean_s = cv::mean(input(cv::Rect(1,y,num_cols-1, 1)));
|
|
|
|
|
|
float mean = (float)mean_s[0];
|
|
|
|
|
|
|
|
|
|
|
|
p++;
|
|
|
|
|
|
|
|
|
|
|
|
float sum_sq = 0;
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int x = 1; x < num_cols; ++x)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
float curr = *p++;
|
|
|
|
|
|
sum_sq += (curr - mean) * (curr - mean);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
float norm = sqrt(sum_sq);
|
|
|
|
|
|
|
|
|
|
|
|
if (norm == 0)
|
|
|
|
|
|
norm = 1;
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int x = 1; x < num_cols; ++x)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
output.at<float>(y, x) = (output.at<float>(y, x) - mean) / norm;
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
void im2colBias(const cv::Mat_<float>& input, const unsigned int width, const unsigned int height, cv::Mat_<float>& output)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int m = input.rows;
|
|
|
|
|
|
const unsigned int n = input.cols;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// determine how many blocks there will be with a sliding window of width x height in the input
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int yB = m - height + 1;
|
|
|
|
|
|
const unsigned int xB = n - width + 1;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// Allocate the output size
|
|
|
|
|
|
if(output.rows != xB*yB && output.cols != width * height + 1)
|
|
|
|
|
|
{
|
|
|
|
|
|
output = cv::Mat::ones(xB*yB, width * height + 1, CV_32F);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Iterate over the blocks
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int j = 0; j< xB; j++)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int i = 0; i< yB; i++)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
2018-05-07 22:15:40 +01:00
|
|
|
|
unsigned int rowIdx = i + j*yB;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
for (unsigned int yy = 0; yy < height; ++yy)
|
|
|
|
|
|
for (unsigned int xx = 0; xx < width; ++xx)
|
|
|
|
|
|
{
|
2018-05-07 22:15:40 +01:00
|
|
|
|
unsigned int colIdx = xx*height + yy;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
output.at<float>(rowIdx, colIdx + 1) = input.at<float>(i + yy, j + xx);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//===========================================================================
|
|
|
|
|
|
void CEN_patch_expert::Response(const cv::Mat_<float> &area_of_interest, cv::Mat_<float> &response)
|
|
|
|
|
|
{
|
|
|
|
|
|
|
2018-05-07 21:54:52 +01:00
|
|
|
|
int response_height = area_of_interest.rows - height_support + 1;
|
|
|
|
|
|
int response_width = area_of_interest.cols - width_support + 1;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
cv::Mat_<float> input_col;
|
2018-05-07 21:54:52 +01:00
|
|
|
|
im2colBias(area_of_interest, width_support, height_support, input_col);
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// Mean and standard deviation normalization
|
|
|
|
|
|
contrastNorm(input_col, response);
|
|
|
|
|
|
response = response.t();
|
|
|
|
|
|
|
|
|
|
|
|
for (size_t layer = 0; layer < activation_function.size(); ++layer)
|
|
|
|
|
|
{
|
|
|
|
|
|
|
|
|
|
|
|
// We are performing response = weights[layers] * response(t), but in OpenBLAS as that is significantly quicker than OpenCV
|
|
|
|
|
|
cv::Mat_<float> resp = response;
|
|
|
|
|
|
float* m1 = (float*)resp.data;
|
|
|
|
|
|
cv::Mat_<float> weight = weights[layer];
|
|
|
|
|
|
float* m2 = (float*)weight.data;
|
|
|
|
|
|
|
|
|
|
|
|
cv::Mat_<float> resp_blas(weight.rows, resp.cols);
|
|
|
|
|
|
float* m3 = (float*)resp_blas.data;
|
|
|
|
|
|
|
|
|
|
|
|
// Perform matrix multiplication in OpenBLAS (fortran call)
|
|
|
|
|
|
float alpha1 = 1.0;
|
|
|
|
|
|
float beta1 = 0.0;
|
2018-05-07 21:54:52 +01:00
|
|
|
|
sgemm_("N", "N", &resp.cols, &weight.rows, &weight.cols, &alpha1, m1, &resp.cols, m2, &weight.cols, &beta1, m3, &resp.cols);
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// The above is a faster version of this, by calling the fortran version directly
|
|
|
|
|
|
//cblas_sgemm(CblasColMajor, CblasNoTrans, CblasNoTrans, resp.cols, weight.rows, weight.cols, 1, m1, resp.cols, m2, weight.cols, 0.0, m3, resp.cols);
|
|
|
|
|
|
|
|
|
|
|
|
// Adding the bias (bit ugly, but the fastest way to do this)
|
|
|
|
|
|
response = resp_blas;
|
|
|
|
|
|
|
|
|
|
|
|
float* data = (float*)response.data;
|
|
|
|
|
|
size_t height = response.rows;
|
|
|
|
|
|
size_t width = response.cols;
|
|
|
|
|
|
float* data_b = (float*)biases[layer].data;
|
|
|
|
|
|
for (size_t y = 0; y < height; ++y)
|
|
|
|
|
|
{
|
|
|
|
|
|
float bias = data_b[y];
|
|
|
|
|
|
for (size_t x = 0; x < width; ++x)
|
|
|
|
|
|
{
|
|
|
|
|
|
float in = *data + bias;
|
|
|
|
|
|
*data++ = in;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Perform activation and add bias at the same time
|
|
|
|
|
|
if (activation_function[layer] == 0) // Sigmoid
|
|
|
|
|
|
{
|
|
|
|
|
|
|
|
|
|
|
|
size_t resp_size = response.rows * response.cols;
|
|
|
|
|
|
|
|
|
|
|
|
// Iterate over the data directly
|
|
|
|
|
|
float* data = (float*)response.data;
|
|
|
|
|
|
|
|
|
|
|
|
for (size_t counter = 0; counter < resp_size; ++counter)
|
|
|
|
|
|
{
|
|
|
|
|
|
float in = *data;
|
|
|
|
|
|
*data++ = 1.0 / (1.0 + exp(-(in)));
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
else if (activation_function[layer] == 2)// ReLU
|
|
|
|
|
|
{
|
|
|
|
|
|
cv::threshold(response, response, 0, 0, cv::THRESH_TOZERO);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
response = response.t();
|
|
|
|
|
|
response = response.reshape(1, response_height);
|
|
|
|
|
|
response = response.t();
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Perform im2col, while at the same time doing contrast normalization and adding a bias term (also skip every other region)
|
|
|
|
|
|
void im2colBiasSparseContrastNorm(const cv::Mat_<float>& input, const unsigned int width, const unsigned int height, cv::Mat_<float>& output)
|
|
|
|
|
|
{
|
|
|
|
|
|
const unsigned int m = input.rows;
|
|
|
|
|
|
const unsigned int n = input.cols;
|
|
|
|
|
|
|
|
|
|
|
|
// determine how many blocks there will be with a sliding window of width x height in the input
|
|
|
|
|
|
const unsigned int yB = m - height + 1;
|
|
|
|
|
|
const unsigned int xB = n - width + 1;
|
|
|
|
|
|
|
|
|
|
|
|
// As we will be skipping half of the outputs
|
|
|
|
|
|
const unsigned int out_size = (yB*xB - 1) / 2;
|
|
|
|
|
|
|
|
|
|
|
|
// Allocate the output size
|
|
|
|
|
|
if (output.rows != out_size && output.cols != width * height + 1)
|
|
|
|
|
|
{
|
|
|
|
|
|
output = cv::Mat::ones(out_size, width * height + 1, CV_32F);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Iterate over the blocks, skipping every second block
|
|
|
|
|
|
unsigned int rowIdx = 0;
|
|
|
|
|
|
unsigned int skipCounter = 0;
|
|
|
|
|
|
for (unsigned int j = 0; j< xB; j++)
|
|
|
|
|
|
{
|
|
|
|
|
|
for (unsigned int i = 0; i< yB; i++)
|
|
|
|
|
|
{
|
|
|
|
|
|
// Skip every second row
|
|
|
|
|
|
skipCounter++;
|
|
|
|
|
|
if ((skipCounter + 1) % 2 == 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
continue;
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
float* Mo = output.ptr<float>(rowIdx);
|
|
|
|
|
|
|
|
|
|
|
|
float sum = 0;
|
|
|
|
|
|
|
|
|
|
|
|
for (unsigned int yy = 0; yy < height; ++yy)
|
|
|
|
|
|
{
|
|
|
|
|
|
const float* Mi = input.ptr<float>(i + yy);
|
|
|
|
|
|
for (unsigned int xx = 0; xx < width; ++xx)
|
|
|
|
|
|
{
|
|
|
|
|
|
int colIdx = xx*height + yy;
|
|
|
|
|
|
float in = Mi[j + xx];
|
|
|
|
|
|
sum += in;
|
|
|
|
|
|
|
|
|
|
|
|
Mo[colIdx+1] = in;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Working out the mean
|
|
|
|
|
|
float mean = sum / (float)(width * height);
|
|
|
|
|
|
|
|
|
|
|
|
float sum_sq = 0;
|
|
|
|
|
|
const unsigned int num_items = width*height + 1;
|
|
|
|
|
|
// Working out the sum squared and subtracting the mean
|
|
|
|
|
|
for (unsigned int x = 1; x < num_items; ++x)
|
|
|
|
|
|
{
|
|
|
|
|
|
float in = Mo[x] - mean;
|
|
|
|
|
|
Mo[x] = in;
|
|
|
|
|
|
sum_sq += in * in;
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
float norm = sqrt(sum_sq);
|
|
|
|
|
|
|
|
|
|
|
|
// Avoiding division by 0
|
|
|
|
|
|
if (norm == 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
norm = 1;
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Flip multiplication to division for speed
|
|
|
|
|
|
norm = 1.0 / norm;
|
|
|
|
|
|
|
|
|
|
|
|
for (unsigned int x = 1; x < num_items; ++x)
|
|
|
|
|
|
{
|
|
|
|
|
|
Mo[x] *= norm;
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
rowIdx++;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
void im2colBiasSparse(const cv::Mat_<float>& input, const unsigned int width, const unsigned int height, cv::Mat_<float>& output)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int m = input.rows;
|
|
|
|
|
|
const unsigned int n = input.cols;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// determine how many blocks there will be with a sliding window of width x height in the input
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int yB = m - height + 1;
|
|
|
|
|
|
const unsigned int xB = n - width + 1;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// As we will be skipping half of the outputs
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int out_size = (yB*xB - 1) / 2;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// Allocate the output size
|
|
|
|
|
|
if (output.rows != out_size && output.cols != width * height + 1)
|
|
|
|
|
|
{
|
|
|
|
|
|
output = cv::Mat::ones(out_size, width * height + 1, CV_32F);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Iterate over the blocks, skipping every second block
|
2018-05-07 22:15:40 +01:00
|
|
|
|
unsigned int rowIdx = 0;
|
|
|
|
|
|
unsigned int skipCounter = 0;
|
|
|
|
|
|
for (unsigned int j = 0; j< xB; j++)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int i = 0; i< yB; i++)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
// Skip every second row
|
|
|
|
|
|
skipCounter++;
|
|
|
|
|
|
if ((skipCounter + 1) % 2 == 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
continue;
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
for (unsigned int yy = 0; yy < height; ++yy)
|
|
|
|
|
|
{
|
|
|
|
|
|
for (unsigned int xx = 0; xx < width; ++xx)
|
|
|
|
|
|
{
|
2018-05-07 22:15:40 +01:00
|
|
|
|
unsigned int colIdx = xx*height + yy;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
output.at<float>(rowIdx, colIdx + 1) = input.at<float>(i + yy, j + xx);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
rowIdx++;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// As the sparse patch expert output with interpolation, this function creates an interpolation matrix
|
|
|
|
|
|
void LandmarkDetector::interpolationMatrix(cv::Mat_<float>& mapMatrix, int response_height, int response_width, int input_width, int input_height)
|
|
|
|
|
|
{
|
|
|
|
|
|
int m = input_height;
|
|
|
|
|
|
int n = input_width;
|
|
|
|
|
|
|
|
|
|
|
|
// determine how many blocks there will be with a sliding window of width x height in the input
|
|
|
|
|
|
int yB = m - 11 + 1;
|
|
|
|
|
|
int xB = n - 11 + 1;
|
|
|
|
|
|
|
|
|
|
|
|
// As we will be skipping half of the outputs
|
|
|
|
|
|
int out_size = (yB*xB - 1) / 2;
|
|
|
|
|
|
|
|
|
|
|
|
mapMatrix.create(out_size, response_height * response_width);
|
|
|
|
|
|
mapMatrix.setTo(0.0f);
|
|
|
|
|
|
|
|
|
|
|
|
// Find a mapping from indices in the computed sparse response and the original full response
|
|
|
|
|
|
cv::Mat_<int> value_id_matrix(response_width, response_height, 0);
|
|
|
|
|
|
|
|
|
|
|
|
int ind = 0;
|
|
|
|
|
|
for (int k = 0; k < value_id_matrix.rows * value_id_matrix.cols; ++k)
|
|
|
|
|
|
{
|
|
|
|
|
|
if (k % 2 != 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
value_id_matrix.at<int>(k) = ind;
|
|
|
|
|
|
ind++;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
value_id_matrix = value_id_matrix.t();
|
|
|
|
|
|
|
|
|
|
|
|
int skip_counter = 0;
|
|
|
|
|
|
for (int x = 0; x < response_width; ++x)
|
|
|
|
|
|
{
|
|
|
|
|
|
for (int y = 0; y < response_height; ++y)
|
|
|
|
|
|
{
|
|
|
|
|
|
int mapping_col = x * response_height + y;
|
|
|
|
|
|
skip_counter++;
|
|
|
|
|
|
if (skip_counter % 2 == 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
int val_id = value_id_matrix.at<int>(y, x);
|
|
|
|
|
|
mapMatrix.at<float>(val_id, mapping_col) = 1;
|
|
|
|
|
|
continue;
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
float num_neigh = 0.0;
|
|
|
|
|
|
vector<int> val_ids;
|
|
|
|
|
|
if (x - 1 >= 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
num_neigh++;
|
|
|
|
|
|
val_ids.push_back(value_id_matrix.at<int>(y, x - 1));
|
|
|
|
|
|
}
|
|
|
|
|
|
if (y - 1 >= 0)
|
|
|
|
|
|
{
|
|
|
|
|
|
num_neigh++;
|
|
|
|
|
|
val_ids.push_back(value_id_matrix.at<int>(y - 1, x));
|
|
|
|
|
|
}
|
|
|
|
|
|
if (x + 1 < response_width)
|
|
|
|
|
|
{
|
|
|
|
|
|
num_neigh++;
|
|
|
|
|
|
val_ids.push_back(value_id_matrix.at<int>(y, x + 1));
|
|
|
|
|
|
}
|
|
|
|
|
|
if (y + 1 < response_height)
|
|
|
|
|
|
{
|
|
|
|
|
|
num_neigh++;
|
|
|
|
|
|
val_ids.push_back(value_id_matrix.at<int>(y + 1, x));
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
for (size_t k = 0; k < val_ids.size(); ++k)
|
|
|
|
|
|
{
|
|
|
|
|
|
mapMatrix.at<float>(val_ids[k], mapping_col) = 1.0 / num_neigh;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//===========================================================================
|
|
|
|
|
|
void CEN_patch_expert::ResponseSparse(const cv::Mat_<float> &area_of_interest, cv::Mat_<float> &response, cv::Mat_<float>& mapMatrix, cv::Mat_<float>& im2col_prealloc)
|
|
|
|
|
|
{
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int response_height = area_of_interest.rows - height_support + 1;
|
|
|
|
|
|
const unsigned int response_width = area_of_interest.cols - width_support + 1;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// Extract im2col but in a sparse way and contrast normalize
|
2018-05-07 21:54:52 +01:00
|
|
|
|
im2colBiasSparseContrastNorm(area_of_interest, width_support, height_support, im2col_prealloc);
|
2018-05-06 08:19:09 +01:00
|
|
|
|
response = im2col_prealloc.t();
|
|
|
|
|
|
|
|
|
|
|
|
for (size_t layer = 0; layer < activation_function.size(); ++layer)
|
|
|
|
|
|
{
|
|
|
|
|
|
|
|
|
|
|
|
// We are performing response = weights[layers] * response(t), but in OpenBLAS as that is significantly quicker than OpenCV
|
|
|
|
|
|
cv::Mat_<float> resp = response;
|
|
|
|
|
|
float* m1 = (float*)resp.data;
|
|
|
|
|
|
cv::Mat_<float> weight = weights[layer];
|
|
|
|
|
|
float* m2 = (float*)weight.data;
|
|
|
|
|
|
|
|
|
|
|
|
cv::Mat_<float> resp_blas(weight.rows, resp.cols);
|
|
|
|
|
|
float* m3 = (float*)resp_blas.data;
|
|
|
|
|
|
|
|
|
|
|
|
// Perform matrix multiplication in OpenBLAS (fortran call)
|
|
|
|
|
|
float alpha1 = 1.0;
|
|
|
|
|
|
float beta1 = 0.0;
|
2018-05-07 21:54:52 +01:00
|
|
|
|
sgemm_("N", "N", &resp.cols, &weight.rows, &weight.cols, &alpha1, m1, &resp.cols, m2, &weight.cols, &beta1, m3, &resp.cols);
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// The above is a faster version of this, by calling the fortran version directly
|
|
|
|
|
|
//cblas_sgemm(CblasColMajor, CblasNoTrans, CblasNoTrans, resp.cols, weight.rows, weight.cols, 1, m1, resp.cols, m2, weight.cols, 0.0, m3, resp.cols);
|
|
|
|
|
|
|
|
|
|
|
|
// Adding the bias (bit ugly, but the fastest way to do this)
|
|
|
|
|
|
response = resp_blas;
|
|
|
|
|
|
|
|
|
|
|
|
float* data = (float*)response.data;
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned height = response.rows;
|
|
|
|
|
|
const unsigned width = response.cols;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
float* data_b = (float*)biases[layer].data;
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int y = 0; y < height; ++y)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
float bias = data_b[y];
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int x = 0; x < width; ++x)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
float in = *data + bias;
|
|
|
|
|
|
*data++ = in;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Perform activation and add bias at the same time
|
|
|
|
|
|
if (activation_function[layer] == 0) // Sigmoid
|
|
|
|
|
|
{
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int resp_size = response.rows * response.cols;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// Iterate over the data directly
|
|
|
|
|
|
float* data = (float*)response.data;
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int counter = 0; counter < resp_size; ++counter)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
float in = *data;
|
|
|
|
|
|
*data++ = 1.0 / (1.0 + exp(-(in)));
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
else if (activation_function[layer] == 2)// ReLU
|
|
|
|
|
|
{
|
|
|
|
|
|
cv::threshold(response, response, 0, 0, cv::THRESH_TOZERO);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
response = response * mapMatrix;
|
|
|
|
|
|
response = response.t();
|
|
|
|
|
|
response = response.reshape(1, response_height);
|
|
|
|
|
|
response = response.t();
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//===========================================================================
|
|
|
|
|
|
void CEN_patch_expert::ResponseSparse_mirror(const cv::Mat_<float> &area_of_interest, cv::Mat_<float> &response, cv::Mat_<float>& mapMatrix, cv::Mat_<float>& im2col_prealloc)
|
|
|
|
|
|
{
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int response_height = area_of_interest.rows - height_support + 1;
|
|
|
|
|
|
const unsigned int response_width = area_of_interest.cols - width_support + 1;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
cv::flip(area_of_interest, area_of_interest, 1);
|
|
|
|
|
|
|
|
|
|
|
|
// Extract im2col but in a sparse way and contrast normalize
|
2018-05-07 21:54:52 +01:00
|
|
|
|
im2colBiasSparseContrastNorm(area_of_interest, width_support, height_support, im2col_prealloc);
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
response = im2col_prealloc.t();
|
|
|
|
|
|
|
|
|
|
|
|
for (size_t layer = 0; layer < activation_function.size(); ++layer)
|
|
|
|
|
|
{
|
|
|
|
|
|
|
|
|
|
|
|
// We are performing response = weights[layers] * response, but in OpenBLAS as that is significantly quicker than OpenCV
|
|
|
|
|
|
cv::Mat_<float> resp = response;
|
|
|
|
|
|
float* m1 = (float*)resp.data;
|
|
|
|
|
|
cv::Mat_<float> weight = weights[layer];
|
|
|
|
|
|
float* m2 = (float*)weight.data;
|
|
|
|
|
|
|
|
|
|
|
|
cv::Mat_<float> resp_blas(weight.rows, resp.cols);
|
|
|
|
|
|
float* m3 = (float*)resp_blas.data;
|
|
|
|
|
|
|
|
|
|
|
|
// Perform matrix multiplication in OpenBLAS (fortran call)
|
|
|
|
|
|
float alpha1 = 1.0;
|
|
|
|
|
|
float beta1 = 0.0;
|
2018-05-07 21:54:52 +01:00
|
|
|
|
sgemm_("N", "N", &resp.cols, &weight.rows, &weight.cols, &alpha1, m1, &resp.cols, m2, &weight.cols, &beta1, m3, &resp.cols);
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// The above is a faster version of this, by calling the fortran version directly
|
|
|
|
|
|
//cblas_sgemm(CblasColMajor, CblasNoTrans, CblasNoTrans, resp.cols, weight.rows, weight.cols, 1, m1, resp.cols, m2, weight.cols, 0.0, m3, resp.cols);
|
|
|
|
|
|
|
|
|
|
|
|
// Adding the bias (bit ugly, but the fastest way to do this)
|
|
|
|
|
|
response = resp_blas;
|
|
|
|
|
|
|
|
|
|
|
|
float* data = (float*)response.data;
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int height = response.rows;
|
|
|
|
|
|
const unsigned int width = response.cols;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
float* data_b = (float*)biases[layer].data;
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int y = 0; y < height; ++y)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
float bias = data_b[y];
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int x = 0; x < width; ++x)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
float in = *data + bias;
|
|
|
|
|
|
*data++ = in;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Perform activation and add bias at the same time
|
|
|
|
|
|
if (activation_function[layer] == 0) // Sigmoid
|
|
|
|
|
|
{
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int resp_size = response.rows * response.cols;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// Iterate over the data directly
|
|
|
|
|
|
float* data = (float*)response.data;
|
|
|
|
|
|
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int counter = 0; counter < resp_size; ++counter)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
float in = *data;
|
|
|
|
|
|
*data++ = 1.0 / (1.0 + exp(-(in)));
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
else if (activation_function[layer] == 2)// ReLU
|
|
|
|
|
|
{
|
|
|
|
|
|
cv::threshold(response, response, 0, 0, cv::THRESH_TOZERO);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
response = response * mapMatrix;
|
|
|
|
|
|
response = response.t();
|
|
|
|
|
|
response = response.reshape(1, response_height);
|
|
|
|
|
|
response = response.t();
|
|
|
|
|
|
cv::flip(response, response, 1);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void CEN_patch_expert::ResponseSparse_mirror_joint(const cv::Mat_<float> &area_of_interest_left, const cv::Mat_<float> &area_of_interest_right, cv::Mat_<float> &response_left, cv::Mat_<float> &response_right, cv::Mat_<float>& mapMatrix, cv::Mat_<float>& im2col_prealloc_left, cv::Mat_<float>& im2col_prealloc_right)
|
|
|
|
|
|
{
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int response_height = area_of_interest_left.rows - height_support + 1;
|
|
|
|
|
|
const unsigned int response_width = area_of_interest_left.cols - width_support + 1;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
cv::flip(area_of_interest_right, area_of_interest_right, 1);
|
|
|
|
|
|
|
|
|
|
|
|
// Extract im2col but in a sparse way and contrast normalize
|
2018-05-07 21:54:52 +01:00
|
|
|
|
im2colBiasSparseContrastNorm(area_of_interest_left, width_support, height_support, im2col_prealloc_left);
|
|
|
|
|
|
im2colBiasSparseContrastNorm(area_of_interest_right, width_support, height_support, im2col_prealloc_right);
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
cv::Mat_<float> response;
|
|
|
|
|
|
cv::vconcat(im2col_prealloc_left, im2col_prealloc_right, response);
|
|
|
|
|
|
|
|
|
|
|
|
response = response.t();
|
|
|
|
|
|
|
|
|
|
|
|
for (size_t layer = 0; layer < activation_function.size(); ++layer)
|
|
|
|
|
|
{
|
|
|
|
|
|
|
|
|
|
|
|
// We are performing response = weights[layers] * response(t), but in OpenBLAS as that is significantly quicker than OpenCV
|
|
|
|
|
|
cv::Mat_<float> resp = response;
|
|
|
|
|
|
float* m1 = (float*)resp.data;
|
|
|
|
|
|
cv::Mat_<float> weight = weights[layer];
|
|
|
|
|
|
float* m2 = (float*)weight.data;
|
|
|
|
|
|
|
|
|
|
|
|
// TODO, this could poss be pre-allocated based on biggest one
|
|
|
|
|
|
cv::Mat_<float> resp_blas(weight.rows, resp.cols);
|
|
|
|
|
|
float* m3 = (float*)resp_blas.data;
|
|
|
|
|
|
|
|
|
|
|
|
// Perform matrix multiplication in OpenBLAS (fortran call)
|
|
|
|
|
|
float alpha1 = 1.0;
|
|
|
|
|
|
float beta1 = 0.0;
|
2018-05-07 21:54:52 +01:00
|
|
|
|
sgemm_("N", "N", &resp.cols, &weight.rows, &weight.cols, &alpha1, m1, &resp.cols, m2, &weight.cols, &beta1, m3, &resp.cols);
|
2018-05-06 08:19:09 +01:00
|
|
|
|
|
|
|
|
|
|
// The above is a faster version of this, by calling the fortran version directly
|
|
|
|
|
|
//cblas_sgemm(CblasColMajor, CblasNoTrans, CblasNoTrans, resp.cols, weight.rows, weight.cols, 1, m1, resp.cols, m2, weight.cols, 0.0, m3, resp.cols);
|
|
|
|
|
|
|
|
|
|
|
|
// Adding the bias (bit ugly, but the fastest way to do this)
|
|
|
|
|
|
response = resp_blas;
|
|
|
|
|
|
|
|
|
|
|
|
float* data = (float*)response.data;
|
2018-05-07 22:15:40 +01:00
|
|
|
|
const unsigned int height = response.rows;
|
|
|
|
|
|
const unsigned int width = response.cols;
|
2018-05-06 08:19:09 +01:00
|
|
|
|
float* data_b = (float*)biases[layer].data;
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int y = 0; y < height; ++y)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
float bias = data_b[y];
|
2018-05-07 22:15:40 +01:00
|
|
|
|
for (unsigned int x = 0; x < width; ++x)
|
2018-05-06 08:19:09 +01:00
|
|
|
|
{
|
|
|
|
|
|
float in = *data + bias;
|
|
|
|
|
|
*data++ = in;
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Perform activation and add bias at the same time
|
|
|
|
|
|
if (activation_function[layer] == 0) // Sigmoid
|
|
|
|
|
|
{
|
|
|
|
|
|
|
|
|
|
|
|
cv::exp(-response, response);
|
|
|
|
|
|
response = 1.0 / (1.0 + response);
|
|
|
|
|
|
|
|
|
|
|
|
// Iterate over the data directly
|
|
|
|
|
|
/*size_t resp_size = response.rows * response.cols;
|
|
|
|
|
|
|
|
|
|
|
|
float* data = (float*)response.data;
|
|
|
|
|
|
|
|
|
|
|
|
for (size_t counter = 0; counter < resp_size; ++counter)
|
|
|
|
|
|
{
|
|
|
|
|
|
float in = *data;
|
|
|
|
|
|
*data++ = 1.0 / (1.0 + std::exp(-(in)));
|
|
|
|
|
|
}*/
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
else if (activation_function[layer] == 2)// ReLU
|
|
|
|
|
|
{
|
|
|
|
|
|
cv::threshold(response, response, 0, 0, cv::THRESH_TOZERO);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
response_left = response(cv::Rect(0, 0, response.cols / 2, 1));
|
|
|
|
|
|
response_right = response(cv::Rect(response.cols / 2, 0, response.cols / 2, 1));
|
|
|
|
|
|
|
|
|
|
|
|
response_left = response_left * mapMatrix;
|
|
|
|
|
|
response_right = response_right * mapMatrix;
|
|
|
|
|
|
|
|
|
|
|
|
response_left = response_left.t();
|
|
|
|
|
|
response_right = response_right.t();
|
|
|
|
|
|
|
|
|
|
|
|
response_left = response_left.reshape(1, response_height);
|
|
|
|
|
|
response_left = response_left.t();
|
|
|
|
|
|
|
|
|
|
|
|
response_right = response_right.reshape(1, response_height);
|
|
|
|
|
|
response_right = response_right.t();
|
|
|
|
|
|
|
|
|
|
|
|
cv::flip(response_right, response_right, 1);
|
|
|
|
|
|
}
|