问题
I am working with normxcorr2
function in Matlab for template matching. However, what I want to do is different from what normxcorr2
does. The built-in normxcorr2
computes cross-correlation taking into account all the pixels in a rectangular template. But I only want certain pixels to participate in the normalized cross correlation process.
For example, I want only the ring-like white region in the following image to be used as a template while correlating. (the black region inside should not be used for computation)
And thus if I correlate the above template with the following image, I am likely to get a normalized value of around 1.0 (considering the outer circles are of same diameter in both images)
I have checked out this solution:- matlab template matching only for 0 (or 1) in matrix but it is not generic.
Can anybody help me with a more general solution in matlab? This can be used to track objects by their outer shapes
EDIT:- This is the entire image for those who want to have a look at it. I want to detect the ball only by the outer circular rim, not the inner details.
回答1:
Ok, let's give it a try... This solution tries to use existing normxcorr2
implementation and modify it to solve yoru problem.
The formula for normalized cross correlation is:
In this case you want to change the integration boundaries for every window. This is turn affects both standard deviations and the correlation itself. Lets tackle it in several steps:
Step #1: Get the correlation right
We can do this by modifying the template image:
template_fix = template;
mean_template_mask = mean(template(mask == 1));
template_fix(mask == 0) = mean_template_mask;
result = normxcorr2(template_fix, query)
Notice that by making this change we make the mean value of the template to be equal to the mean value of the template in side the mask. this way all template pixels outside the mask don't contribute to the integration as they are equal to the mean value.
Step #2: Fix template std
size_mask = sum(mask(:));
size_template = numel(template);
std_template = std2(template);
std_template_masked = sqrt(sum((template_fix(:) - mean_template_mask).^2)/size_mask);
result = result * (std_template/std_template_masked);
Step #3: Fix query std
sum_filt = ones(size(template));
std_query = filter2(query.^2, sum_filt) - filter2(query, sum_filt).^2/size_template;
std_query = sqrt(std_query/size_template);
std_query_mask = filter2(query.^2, mask) - filter2(query, mask).^2/size_mask;
std_query_mask = sqrt(std_query_mask/size_mask);
result = result .* std_query ./ std_query_mask;
My Matlab is not responding so I didn't have the chance to test it in practice. Unless I missed some errors it should be mathematically equivalent.
This solution does some extra convolutions but it doesn't process overlapping pixels more than once.
If you use the same template image multiple times then you could refactor steps 1 & 2 to run only once for preprocessing. Although both shouldn't be computationally expensive.
Different approach: Straight forward
Here is a different, straightforward approach that doesn't use the original normxcorr2
function. This code can be easily optimized for memory usage in expense of readability.
% q for query, t for template and mask for mask
% shape = 'full' or 'same' or 'valid'
t_mask = t .* mask;
n = numel(mask);
tq_sum = filter2(t_mask,q, shape);
q_sum = filter2(mask, q, shape);
q_mean = q_sum/n;
t_sum = sum(t_mask(:));
t_mean = t_sum/n;
res1 = tq_sum - t_mean*q_sum - t_sum*q_mean + t_mean*q_mean*n;
t_std = sqrt((sum(t_mask(:).^2) - sum(t_mask(:)).^2/n)/n);
q_std = sqrt((filter2(mask, q.^2, shape) - q_sum.^2/n)/n);
res = res1 ./ (n * t_std * q_std)
回答2:
This is a derivative of a previous post where I provided an answer to here: matlab template matching only for 0 (or 1) in matrix
However, this solution used for
loops which is quite inefficient. As such, we can use im2col
, bsxfun
and col2im
to help us perform this more quickly. im2col
takes overlapping regions in your image and places them each into individual columns. Essentially, this takes sliding windows of your image like you would with any kind of spatial image filtering and collects all pixels within a sliding window and places each window as individual columns.
Supposing the size of your template is M x N
, and the size of your image you want to search in is R x C
, and supposing that your image template is called imTemplate
while the image you want to search in is imSearch
, we can do the following setup. Let's also assume that both images are binary.
[M, N] = size(imTemplate);
[R, C] = size(imSearch);
%// Cast to double for precision
imTemplate = im2double(imTemplate);
imSearch = im2double(imSearch);
neigh = im2col(imSearch, [M, N]);
templateCol = imTemplate(:); %// Ensures we place template into single column
Now, you wish to exclude all pixels that are inside the circular boundary. As such, what we can do is invert the image so that black pixels become white, then remove all of the pixels around the border. This should then give us the interior of the circle.
imInvert = ~imTemplate;
imInvertNoBorder = imclearborder(imInvert, 8); %// Search 8-pixel neighbourhood
We will use this to figure out what pixels we are going to remove from searching. This can be done by:
rowsToRemove = imInvertNoBorder(:) == 1;
Now, what we can do is finally remove those pixels that are within the interior of the circle to not be searched in our correlation scheme.
neigh(rowsToRemove,:) = [];
What we can do now is compute the NCC over all of these columns. If you recall, the NCC between two signals is the following:
(source: www.jot.fm)
As such, we need to subtract the mean from each neighbourhood, and we also need to subtract the mean from each of the columns. We then compute the formula as shown above. We can easily achieve this vectorized in MATLAB like so:
neighMeanSubtract = bsxfun(@minus, neigh, mean(neigh));
templateMeanSubtract = templateCol - mean(templateCol);
We can compute the numerator of the NCC for each neighbourhood (before we sum) as follows:
numerator = bsxfun(@times, neighMeanSubtract, templateMeanSubtract);
Now, all we have to do is sum all of the columns and that will give us our final numerator:
sumNumerator = sum(numerator);
The denominator can be computed like so:
denominator1 = sqrt(sum(neighMeanSubtract.*neighMeanSubtract));
denominator2 = sqrt(sum(templateMeanSubtract.*templateMeanSubtract));
sumDenominator = denominator1 .* denominator2;
Finally, our NCC can be computed as so:
NCC = sumNumerator ./ sumDenominator;
You'll notice that this is a single row of values. Each row corresponds to the output defined at a neighbourhood. As such, we also need to reshape this back into a matrix, and so you can use col2im
:
finalOutput = col2im(NCC, [M, N], [R, C]);
The above statement will take overlapping neighbourhoods of M x N
defined in NCC, and reshapes it so that it becomes a R x C
matrix. Sometimes, you will get a divide by zero error, especially if the neighbourhood search window is all uniform. As such, you will get NaN
numbers. Areas of no variation are assumed to have no correlation in image processing, and so let's zero these locations:
finalOutput(isnan(finalOutput)) = 0;
If you want to find the location of where the highest correlation is, simply do:
[rowNCC, colNCC] = find(finalOutput == max(finalOutput(:)));
If you want to interpret negative correlation, that completely depends on your application. If you want to ensure that your template matching algorithm is rotationally invariant, then you should actually check for the max of the absolute values. Negative correlation simply means that the match between the template and a neighbourhood is simply rotated. As such, a better way to find the best neighbourhood is:
maxCoeff = max(abs(finalOutput(:)));
[rowNCC, colNCC] = find(abs(finalOutput) == maxCoeff);
For your copying and pasting pleasure, here is the code in its entirety:
function [rowNCC, colNCC] = testCorr(imTemplate, imSearch)
[M, N] = size(imTemplate);
[R, C] = size(imSearch);
%// Cast to double for precision
imTemplate = im2double(imTemplate);
imSearch = im2double(imSearch);
neigh = im2col(imSearch, [M, N]);
templateCol = imTemplate(:); %// Ensures we place template into single column
imInvert = ~imTemplate;
imInvertNoBorder = imclearborder(imInvert, 8); %// Search 8-pixel neighbourhood
rowsToRemove = imInvertNoBorder(:) == 1;
neigh(rowsToRemove,:) = [];
neighMeanSubtract = bsxfun(@minus, neigh, mean(neigh));
templateMeanSubtract = templateCol - mean(templateCol);
numerator = bsxfun(@times, neighMeanSubtract, templateMeanSubtract);
sumNumerator = sum(numerator);
denominator1 = sqrt(sum(neighMeanSubtract.*neighMeanSubtract));
denominator2 = sqrt(sum(templateMeanSubtract.*templateMeanSubtract));
sumDenominator = denominator1 .* denominator2;
NCC = sumNumerator ./ sumDenominator;
finalOutput = col2im(NCC, [M, N], [R, C]);
finalOutput(isnan(finalOutput)) = 0;
maxCoeff = max(abs(finalOutput(:)));
[rowNCC, colNCC] = find(abs(finalOutput) == maxCoeff);
end
Good luck!
来源:https://stackoverflow.com/questions/24185744/matching-object-outer-shape-using-normalized-cross-correlation