# Optimization code. One line takses 86% of working time.

2 views (last 30 days)
Jan Kowal on 13 May 2016
Commented: Ahmet Cecen on 14 May 2016
Hello Everyone! I am developing a Clasificator for hand written signs and digits using Logical Regression and Softmax function.
I have a problem with one line in one function because it takes too much time.
function [grad] = logistic_cost_function(xTrain, yTrain, w)
%xTrain = [30134 x 11760]
%xTrain = [30134 x 1]
%w = [36 x 11760]
N = size(xTrain,1);
K = 36;
ARG = w * xTrain';
for n = 1 : N
xRow = xTrain(n,:);
Probability = softmax(ARG(:,n));
label_n = yTrain(n);
for k = 1 : K
IND = label_n == k;
prob_k = Probability(k);
end
end
end It is the time counted on only 2 epochs. My point is to count a few thousands of epchos.
Does anyone have any idea how to improve an algorithm?
Bet Regards
Ahmet Cecen on 14 May 2016
I believe that might be the point. Don't think of it as a negative probability, think of IND as an indicator function. You either subtract expectation by the probability of hitting the case, or add expectation by the probability of not hitting the case. I may be wrong though as I have no idea what he is actually trying to do.

Ahmet Cecen on 14 May 2016
Edited: Ahmet Cecen on 14 May 2016
for k = 1 : K
IND = label_n == k;
prob_k = Probability(k);
end
Here is a way out of this loop (well hopefully, as I can't actually run the code to verify):
IND = zeros(K,1); IND(label_n)=1;
prob_k = Probability;
% (IND - prob_k) IS A COLUMN VECTOR, SO OUTER PRODUCT.
Hmm... turns out no need for bsxfun.
Ahmet Cecen on 14 May 2016
Yeah, in a previous version of my answer I didn't see that operation was an outer product for some reason and used a weird combination of bsxfun and repmat. Sometimes when I focus on trying to understand other people's code the simple stuff manages to elude me.