/////////////////////////////////////////////////////////////////////////////// // Copyright (C) 2016, Carnegie Mellon University and University of Cambridge, // all rights reserved. // // THIS SOFTWARE IS PROVIDED “AS IS” FOR ACADEMIC USE ONLY AND ANY EXPRESS // OR IMPLIED WARRANTIES WARRANTIES, INCLUDING, BUT NOT LIMITED TO, // THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR // PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS // BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY. // OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF // SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) // HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, // STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN // ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE // POSSIBILITY OF SUCH DAMAGE. // // Notwithstanding the license granted herein, Licensee acknowledges that certain components // of the Software may be covered by so-called “open source” software licenses (“Open Source // Components”), which means any software licenses approved as open source licenses by the // Open Source Initiative or any substantially similar licenses, including without limitation any // license that, as a condition of distribution of the software licensed under such license, // requires that the distributor make the software available in source code format. Licensor shall // provide a list of Open Source Components for a particular version of the Software upon // Licensee’s request. Licensee will comply with the applicable terms of such licenses and to // the extent required by the licenses covering Open Source Components, the terms of such // licenses will apply in lieu of the terms of this Agreement. To the extent the terms of the // licenses applicable to Open Source Components prohibit any of the restrictions in this // License Agreement with respect to such Open Source Component, such restrictions will not // apply to such Open Source Component. To the extent the terms of the licenses applicable to // Open Source Components require Licensor to make an offer to provide source code or // related information in connection with the Software, such offer is hereby made. Any request // for source code or related information should be directed to cl-face-tracker-distribution@lists.cam.ac.uk // Licensee acknowledges receipt of notices for the Open Source Components for the initial // delivery of the Software. // * Any publications arising from the use of this software, including but // not limited to academic journal and conference publications, technical // reports and manuals, must cite at least one of the following works: // // OpenFace: an open source facial behavior analysis toolkit // Tadas Baltrušaitis, Peter Robinson, and Louis-Philippe Morency // in IEEE Winter Conference on Applications of Computer Vision, 2016 // // Rendering of Eyes for Eye-Shape Registration and Gaze Estimation // Erroll Wood, Tadas Baltrušaitis, Xucong Zhang, Yusuke Sugano, Peter Robinson, and Andreas Bulling // in IEEE International. Conference on Computer Vision (ICCV), 2015 // // Cross-dataset learning and person-speci?c normalisation for automatic Action Unit detection // Tadas Baltrušaitis, Marwa Mahmoud, and Peter Robinson // in Facial Expression Recognition and Analysis Challenge, // IEEE International Conference on Automatic Face and Gesture Recognition, 2015 // // Constrained Local Neural Fields for robust facial landmark detection in the wild. // Tadas Baltrušaitis, Peter Robinson, and Louis-Philippe Morency. // in IEEE Int. Conference on Computer Vision Workshops, 300 Faces in-the-Wild Challenge, 2013. // /////////////////////////////////////////////////////////////////////////////// #include "stdafx.h" #include "CEN_patch_expert.h" // OpenCV includes #include #include // Local includes #include "LandmarkDetectorUtils.h" // For exponential #include using namespace LandmarkDetector; // Copy constructor CEN_patch_expert::CEN_patch_expert(const CEN_patch_expert& other) : confidence(other.confidence), width_support(other.width_support), height_support(other.height_support) { // Copy the layer weights in a deep way for (size_t i = 0; i < other.weights.size(); ++i) { this->weights.push_back(other.weights[i].clone()); this->biases.push_back(other.biases[i].clone()); this->activation_function.push_back(other.activation_function[i]); } } //=========================================================================== void CEN_patch_expert::Read(ifstream &stream) { // Setting up OpenBLAS goto_set_num_threads(1); openblas_set_num_threads(1); // Sanity check int read_type; stream.read((char*)&read_type, 4); assert(read_type == 6); // the number of neurons for this patch int num_layers; stream.read((char*)&width_support, 4); stream.read((char*)&height_support, 4); stream.read((char*)&num_layers, 4); if (num_layers == 0) { // empty patch due to landmark being invisible at that orientation (or visible through mirroring) stream.read((char*)&confidence, 8); return; } activation_function.resize(num_layers); weights.resize(num_layers); biases.resize(num_layers); for (int i = 0; i < num_layers; i++) { int neuron_type; stream.read((char*)&neuron_type, 4); activation_function[i] = neuron_type; cv::Mat_ bias; LandmarkDetector::ReadMatBin(stream, bias); cv::Mat_ weight; LandmarkDetector::ReadMatBin(stream, weight); weights[i] = weight; biases[i] = bias; } // Read the patch confidence stream.read((char*)&confidence, 8); } // Contrast normalize the input for response map computation void contrastNorm(const cv::Mat_& input, cv::Mat_& output) { const unsigned int num_cols = input.cols; const unsigned int num_rows = input.rows; output = input.clone(); cv::MatConstIterator_ p = input.begin(); // Compute row wise for (unsigned int y = 0; y < num_rows; ++y) { cv::Scalar mean_s = cv::mean(input(cv::Rect(1,y,num_cols-1, 1))); float mean = (float)mean_s[0]; p++; float sum_sq = 0; for (unsigned int x = 1; x < num_cols; ++x) { float curr = *p++; sum_sq += (curr - mean) * (curr - mean); } float norm = sqrt(sum_sq); if (norm == 0) norm = 1; for (unsigned int x = 1; x < num_cols; ++x) { output.at(y, x) = (output.at(y, x) - mean) / norm; } } } void im2colBias(const cv::Mat_& input, const unsigned int width, const unsigned int height, cv::Mat_& output) { const unsigned int m = input.rows; const unsigned int n = input.cols; // determine how many blocks there will be with a sliding window of width x height in the input const unsigned int yB = m - height + 1; const unsigned int xB = n - width + 1; // Allocate the output size if(output.rows != xB*yB && output.cols != width * height + 1) { output = cv::Mat::ones(xB*yB, width * height + 1, CV_32F); } // Iterate over the blocks for (unsigned int j = 0; j< xB; j++) { for (unsigned int i = 0; i< yB; i++) { unsigned int rowIdx = i + j*yB; for (unsigned int yy = 0; yy < height; ++yy) for (unsigned int xx = 0; xx < width; ++xx) { unsigned int colIdx = xx*height + yy; output.at(rowIdx, colIdx + 1) = input.at(i + yy, j + xx); } } } } //=========================================================================== void CEN_patch_expert::Response(const cv::Mat_ &area_of_interest, cv::Mat_ &response) { int response_height = area_of_interest.rows - height_support + 1; int response_width = area_of_interest.cols - width_support + 1; cv::Mat_ input_col; im2colBias(area_of_interest, width_support, height_support, input_col); // Mean and standard deviation normalization contrastNorm(input_col, response); response = response.t(); for (size_t layer = 0; layer < activation_function.size(); ++layer) { // We are performing response = weights[layers] * response(t), but in OpenBLAS as that is significantly quicker than OpenCV cv::Mat_ resp = response; float* m1 = (float*)resp.data; cv::Mat_ weight = weights[layer]; float* m2 = (float*)weight.data; cv::Mat_ resp_blas(weight.rows, resp.cols); float* m3 = (float*)resp_blas.data; // Perform matrix multiplication in OpenBLAS (fortran call) float alpha1 = 1.0; float beta1 = 0.0; sgemm_("N", "N", &resp.cols, &weight.rows, &weight.cols, &alpha1, m1, &resp.cols, m2, &weight.cols, &beta1, m3, &resp.cols); // The above is a faster version of this, by calling the fortran version directly //cblas_sgemm(CblasColMajor, CblasNoTrans, CblasNoTrans, resp.cols, weight.rows, weight.cols, 1, m1, resp.cols, m2, weight.cols, 0.0, m3, resp.cols); // Adding the bias (bit ugly, but the fastest way to do this) response = resp_blas; float* data = (float*)response.data; size_t height = response.rows; size_t width = response.cols; float* data_b = (float*)biases[layer].data; for (size_t y = 0; y < height; ++y) { float bias = data_b[y]; for (size_t x = 0; x < width; ++x) { float in = *data + bias; *data++ = in; } } // Perform activation and add bias at the same time if (activation_function[layer] == 0) // Sigmoid { size_t resp_size = response.rows * response.cols; // Iterate over the data directly float* data = (float*)response.data; for (size_t counter = 0; counter < resp_size; ++counter) { float in = *data; *data++ = 1.0 / (1.0 + exp(-(in))); } } else if (activation_function[layer] == 2)// ReLU { cv::threshold(response, response, 0, 0, cv::THRESH_TOZERO); } } response = response.t(); response = response.reshape(1, response_height); response = response.t(); } // Perform im2col, while at the same time doing contrast normalization and adding a bias term (also skip every other region) void im2colBiasSparseContrastNorm(const cv::Mat_& input, const unsigned int width, const unsigned int height, cv::Mat_& output) { const unsigned int m = input.rows; const unsigned int n = input.cols; // determine how many blocks there will be with a sliding window of width x height in the input const unsigned int yB = m - height + 1; const unsigned int xB = n - width + 1; // As we will be skipping half of the outputs const unsigned int out_size = (yB*xB - 1) / 2; // Allocate the output size if (output.rows != out_size && output.cols != width * height + 1) { output = cv::Mat::ones(out_size, width * height + 1, CV_32F); } // Iterate over the blocks, skipping every second block unsigned int rowIdx = 0; unsigned int skipCounter = 0; for (unsigned int j = 0; j< xB; j++) { for (unsigned int i = 0; i< yB; i++) { // Skip every second row skipCounter++; if ((skipCounter + 1) % 2 == 0) { continue; } float* Mo = output.ptr(rowIdx); float sum = 0; for (unsigned int yy = 0; yy < height; ++yy) { const float* Mi = input.ptr(i + yy); for (unsigned int xx = 0; xx < width; ++xx) { int colIdx = xx*height + yy; float in = Mi[j + xx]; sum += in; Mo[colIdx+1] = in; } } // Working out the mean float mean = sum / (float)(width * height); float sum_sq = 0; const unsigned int num_items = width*height + 1; // Working out the sum squared and subtracting the mean for (unsigned int x = 1; x < num_items; ++x) { float in = Mo[x] - mean; Mo[x] = in; sum_sq += in * in; } float norm = sqrt(sum_sq); // Avoiding division by 0 if (norm == 0) { norm = 1; } // Flip multiplication to division for speed norm = 1.0 / norm; for (unsigned int x = 1; x < num_items; ++x) { Mo[x] *= norm; } rowIdx++; } } } void im2colBiasSparse(const cv::Mat_& input, const unsigned int width, const unsigned int height, cv::Mat_& output) { const unsigned int m = input.rows; const unsigned int n = input.cols; // determine how many blocks there will be with a sliding window of width x height in the input const unsigned int yB = m - height + 1; const unsigned int xB = n - width + 1; // As we will be skipping half of the outputs const unsigned int out_size = (yB*xB - 1) / 2; // Allocate the output size if (output.rows != out_size && output.cols != width * height + 1) { output = cv::Mat::ones(out_size, width * height + 1, CV_32F); } // Iterate over the blocks, skipping every second block unsigned int rowIdx = 0; unsigned int skipCounter = 0; for (unsigned int j = 0; j< xB; j++) { for (unsigned int i = 0; i< yB; i++) { // Skip every second row skipCounter++; if ((skipCounter + 1) % 2 == 0) { continue; } for (unsigned int yy = 0; yy < height; ++yy) { for (unsigned int xx = 0; xx < width; ++xx) { unsigned int colIdx = xx*height + yy; output.at(rowIdx, colIdx + 1) = input.at(i + yy, j + xx); } } rowIdx++; } } } // As the sparse patch expert output with interpolation, this function creates an interpolation matrix void LandmarkDetector::interpolationMatrix(cv::Mat_& mapMatrix, int response_height, int response_width, int input_width, int input_height) { int m = input_height; int n = input_width; // determine how many blocks there will be with a sliding window of width x height in the input int yB = m - 11 + 1; int xB = n - 11 + 1; // As we will be skipping half of the outputs int out_size = (yB*xB - 1) / 2; mapMatrix.create(out_size, response_height * response_width); mapMatrix.setTo(0.0f); // Find a mapping from indices in the computed sparse response and the original full response cv::Mat_ value_id_matrix(response_width, response_height, 0); int ind = 0; for (int k = 0; k < value_id_matrix.rows * value_id_matrix.cols; ++k) { if (k % 2 != 0) { value_id_matrix.at(k) = ind; ind++; } } value_id_matrix = value_id_matrix.t(); int skip_counter = 0; for (int x = 0; x < response_width; ++x) { for (int y = 0; y < response_height; ++y) { int mapping_col = x * response_height + y; skip_counter++; if (skip_counter % 2 == 0) { int val_id = value_id_matrix.at(y, x); mapMatrix.at(val_id, mapping_col) = 1; continue; } float num_neigh = 0.0; vector val_ids; if (x - 1 >= 0) { num_neigh++; val_ids.push_back(value_id_matrix.at(y, x - 1)); } if (y - 1 >= 0) { num_neigh++; val_ids.push_back(value_id_matrix.at(y - 1, x)); } if (x + 1 < response_width) { num_neigh++; val_ids.push_back(value_id_matrix.at(y, x + 1)); } if (y + 1 < response_height) { num_neigh++; val_ids.push_back(value_id_matrix.at(y + 1, x)); } for (size_t k = 0; k < val_ids.size(); ++k) { mapMatrix.at(val_ids[k], mapping_col) = 1.0 / num_neigh; } } } } //=========================================================================== void CEN_patch_expert::ResponseSparse(const cv::Mat_ &area_of_interest, cv::Mat_ &response, cv::Mat_& mapMatrix, cv::Mat_& im2col_prealloc) { const unsigned int response_height = area_of_interest.rows - height_support + 1; const unsigned int response_width = area_of_interest.cols - width_support + 1; // Extract im2col but in a sparse way and contrast normalize im2colBiasSparseContrastNorm(area_of_interest, width_support, height_support, im2col_prealloc); response = im2col_prealloc.t(); for (size_t layer = 0; layer < activation_function.size(); ++layer) { // We are performing response = weights[layers] * response(t), but in OpenBLAS as that is significantly quicker than OpenCV cv::Mat_ resp = response; float* m1 = (float*)resp.data; cv::Mat_ weight = weights[layer]; float* m2 = (float*)weight.data; cv::Mat_ resp_blas(weight.rows, resp.cols); float* m3 = (float*)resp_blas.data; // Perform matrix multiplication in OpenBLAS (fortran call) float alpha1 = 1.0; float beta1 = 0.0; sgemm_("N", "N", &resp.cols, &weight.rows, &weight.cols, &alpha1, m1, &resp.cols, m2, &weight.cols, &beta1, m3, &resp.cols); // The above is a faster version of this, by calling the fortran version directly //cblas_sgemm(CblasColMajor, CblasNoTrans, CblasNoTrans, resp.cols, weight.rows, weight.cols, 1, m1, resp.cols, m2, weight.cols, 0.0, m3, resp.cols); // Adding the bias (bit ugly, but the fastest way to do this) response = resp_blas; float* data = (float*)response.data; const unsigned height = response.rows; const unsigned width = response.cols; float* data_b = (float*)biases[layer].data; for (unsigned int y = 0; y < height; ++y) { float bias = data_b[y]; for (unsigned int x = 0; x < width; ++x) { float in = *data + bias; *data++ = in; } } // Perform activation and add bias at the same time if (activation_function[layer] == 0) // Sigmoid { const unsigned int resp_size = response.rows * response.cols; // Iterate over the data directly float* data = (float*)response.data; for (unsigned int counter = 0; counter < resp_size; ++counter) { float in = *data; *data++ = 1.0 / (1.0 + exp(-(in))); } } else if (activation_function[layer] == 2)// ReLU { cv::threshold(response, response, 0, 0, cv::THRESH_TOZERO); } } response = response * mapMatrix; response = response.t(); response = response.reshape(1, response_height); response = response.t(); } //=========================================================================== void CEN_patch_expert::ResponseSparse_mirror(const cv::Mat_ &area_of_interest, cv::Mat_ &response, cv::Mat_& mapMatrix, cv::Mat_& im2col_prealloc) { const unsigned int response_height = area_of_interest.rows - height_support + 1; const unsigned int response_width = area_of_interest.cols - width_support + 1; cv::flip(area_of_interest, area_of_interest, 1); // Extract im2col but in a sparse way and contrast normalize im2colBiasSparseContrastNorm(area_of_interest, width_support, height_support, im2col_prealloc); response = im2col_prealloc.t(); for (size_t layer = 0; layer < activation_function.size(); ++layer) { // We are performing response = weights[layers] * response, but in OpenBLAS as that is significantly quicker than OpenCV cv::Mat_ resp = response; float* m1 = (float*)resp.data; cv::Mat_ weight = weights[layer]; float* m2 = (float*)weight.data; cv::Mat_ resp_blas(weight.rows, resp.cols); float* m3 = (float*)resp_blas.data; // Perform matrix multiplication in OpenBLAS (fortran call) float alpha1 = 1.0; float beta1 = 0.0; sgemm_("N", "N", &resp.cols, &weight.rows, &weight.cols, &alpha1, m1, &resp.cols, m2, &weight.cols, &beta1, m3, &resp.cols); // The above is a faster version of this, by calling the fortran version directly //cblas_sgemm(CblasColMajor, CblasNoTrans, CblasNoTrans, resp.cols, weight.rows, weight.cols, 1, m1, resp.cols, m2, weight.cols, 0.0, m3, resp.cols); // Adding the bias (bit ugly, but the fastest way to do this) response = resp_blas; float* data = (float*)response.data; const unsigned int height = response.rows; const unsigned int width = response.cols; float* data_b = (float*)biases[layer].data; for (unsigned int y = 0; y < height; ++y) { float bias = data_b[y]; for (unsigned int x = 0; x < width; ++x) { float in = *data + bias; *data++ = in; } } // Perform activation and add bias at the same time if (activation_function[layer] == 0) // Sigmoid { const unsigned int resp_size = response.rows * response.cols; // Iterate over the data directly float* data = (float*)response.data; for (unsigned int counter = 0; counter < resp_size; ++counter) { float in = *data; *data++ = 1.0 / (1.0 + exp(-(in))); } } else if (activation_function[layer] == 2)// ReLU { cv::threshold(response, response, 0, 0, cv::THRESH_TOZERO); } } response = response * mapMatrix; response = response.t(); response = response.reshape(1, response_height); response = response.t(); cv::flip(response, response, 1); } void CEN_patch_expert::ResponseSparse_mirror_joint(const cv::Mat_ &area_of_interest_left, const cv::Mat_ &area_of_interest_right, cv::Mat_ &response_left, cv::Mat_ &response_right, cv::Mat_& mapMatrix, cv::Mat_& im2col_prealloc_left, cv::Mat_& im2col_prealloc_right) { const unsigned int response_height = area_of_interest_left.rows - height_support + 1; const unsigned int response_width = area_of_interest_left.cols - width_support + 1; cv::flip(area_of_interest_right, area_of_interest_right, 1); // Extract im2col but in a sparse way and contrast normalize im2colBiasSparseContrastNorm(area_of_interest_left, width_support, height_support, im2col_prealloc_left); im2colBiasSparseContrastNorm(area_of_interest_right, width_support, height_support, im2col_prealloc_right); cv::Mat_ response; cv::vconcat(im2col_prealloc_left, im2col_prealloc_right, response); response = response.t(); for (size_t layer = 0; layer < activation_function.size(); ++layer) { // We are performing response = weights[layers] * response(t), but in OpenBLAS as that is significantly quicker than OpenCV cv::Mat_ resp = response; float* m1 = (float*)resp.data; cv::Mat_ weight = weights[layer]; float* m2 = (float*)weight.data; // TODO, this could poss be pre-allocated based on biggest one cv::Mat_ resp_blas(weight.rows, resp.cols); float* m3 = (float*)resp_blas.data; // Perform matrix multiplication in OpenBLAS (fortran call) float alpha1 = 1.0; float beta1 = 0.0; sgemm_("N", "N", &resp.cols, &weight.rows, &weight.cols, &alpha1, m1, &resp.cols, m2, &weight.cols, &beta1, m3, &resp.cols); // The above is a faster version of this, by calling the fortran version directly //cblas_sgemm(CblasColMajor, CblasNoTrans, CblasNoTrans, resp.cols, weight.rows, weight.cols, 1, m1, resp.cols, m2, weight.cols, 0.0, m3, resp.cols); // Adding the bias (bit ugly, but the fastest way to do this) response = resp_blas; float* data = (float*)response.data; const unsigned int height = response.rows; const unsigned int width = response.cols; float* data_b = (float*)biases[layer].data; for (unsigned int y = 0; y < height; ++y) { float bias = data_b[y]; for (unsigned int x = 0; x < width; ++x) { float in = *data + bias; *data++ = in; } } // Perform activation and add bias at the same time if (activation_function[layer] == 0) // Sigmoid { cv::exp(-response, response); response = 1.0 / (1.0 + response); // Iterate over the data directly /*size_t resp_size = response.rows * response.cols; float* data = (float*)response.data; for (size_t counter = 0; counter < resp_size; ++counter) { float in = *data; *data++ = 1.0 / (1.0 + std::exp(-(in))); }*/ } else if (activation_function[layer] == 2)// ReLU { cv::threshold(response, response, 0, 0, cv::THRESH_TOZERO); } } response_left = response(cv::Rect(0, 0, response.cols / 2, 1)); response_right = response(cv::Rect(response.cols / 2, 0, response.cols / 2, 1)); response_left = response_left * mapMatrix; response_right = response_right * mapMatrix; response_left = response_left.t(); response_right = response_right.t(); response_left = response_left.reshape(1, response_height); response_left = response_left.t(); response_right = response_right.reshape(1, response_height); response_right = response_right.t(); cv::flip(response_right, response_right, 1); }