Skip to content

Commit

Permalink
Tranform classic loops to range-based for loops in module ml
Browse files Browse the repository at this point in the history
Changes are based on the result of run-clang-tidy -header-filter='.*' -checks='-*,modernize-loop-convert' -fix
  • Loading branch information
Heiko Thiel committed Feb 9, 2019
1 parent 1cb184d commit ab5284a
Show file tree
Hide file tree
Showing 2 changed files with 24 additions and 24 deletions.
12 changes: 6 additions & 6 deletions ml/src/densecrf.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -52,8 +52,8 @@ pcl::DenseCrf::DenseCrf (int N, int m) :
//////////////////////////////////////////////////////////////////////////////////////////////////////////////////////
pcl::DenseCrf::~DenseCrf ()
{
for(size_t i = 0; i < pairwise_potential_.size (); i++ )
delete pairwise_potential_[i];
for(auto &p : pairwise_potential_)
delete p;
}

//////////////////////////////////////////////////////////////////////////////////////////////////////////////////////
Expand Down Expand Up @@ -294,11 +294,11 @@ pcl::DenseCrf::runInference (float relax)
next_[i] = -unary_[i];

// Add up all pairwise potentials
for( size_t i = 0; i < pairwise_potential_.size(); i++ )
pairwise_potential_[i]->compute( next_, current_, tmp_, M_ );
for(auto &p : pairwise_potential_)
p->compute( next_, current_, tmp_, M_ );

// Exponentiate and normalize
expAndNormalize( current_, next_, 1.0, relax );
// Exponentiate and normalize
expAndNormalize( current_, next_, 1.0, relax );
}

void
Expand Down
36 changes: 18 additions & 18 deletions ml/src/svm_wrapper.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -134,10 +134,10 @@ pcl::SVMTrain::scaleFactors (std::vector<SVMData> training_set, svm_scaling &sca
{
int max = 0;

for (size_t i = 0; i < training_set.size() ; i++)
for (size_t j = 0; j < training_set[i].SV.size() ; j++)
if (training_set[i].SV[j].idx > max)
max = training_set[i].SV[j].idx; // max number of features
for (const auto &svm_data : training_set)
for (const auto &sample : svm_data.SV)
if (sample.idx > max)
max = sample.idx; // max number of features

max += 1;

Expand All @@ -151,13 +151,13 @@ pcl::SVMTrain::scaleFactors (std::vector<SVMData> training_set, svm_scaling &sca
scaling.obj[i].value = 0;
}

for (size_t i = 0; i < training_set.size(); i++)
for (size_t j = 0; j < training_set[i].SV.size(); j++)
for (auto &svm_data : training_set)
for (const auto &sample : svm_data.SV)
// save scaling factor finding the maximum value
if (module (training_set[i].SV[j].value) > scaling.obj[ training_set[i].SV[j].idx ].value)
if (module (sample.value) > scaling.obj[sample.idx].value)
{
scaling.obj[ training_set[i].SV[j].idx ].index = 1;
scaling.obj[ training_set[i].SV[j].idx ].value = module (training_set[i].SV[j].value);
scaling.obj[sample.idx].index = 1;
scaling.obj[sample.idx].value = module (sample.value);
}
};

Expand Down Expand Up @@ -462,18 +462,18 @@ pcl::SVM::saveProblem (const char *filename, bool labelled = false)
}


for (size_t j = 0; j < training_set_.size() ; j++)
for (const auto &svm_data : training_set_)
{

if (labelled)
{
assert (std::isfinite (training_set_[j].label));
myfile << training_set_[j].label << " ";
assert (std::isfinite (svm_data.label));
myfile << svm_data.label << " ";
}

for (size_t i = 0; i < training_set_[j].SV.size(); i++)
if (std::isfinite (training_set_[j].SV[i].value))
myfile << training_set_[j].SV[i].idx << ":" << training_set_[j].SV[i].value << " ";
for (const auto &sample : svm_data.SV)
if (std::isfinite (sample.value))
myfile << sample.idx << ":" << sample.value << " ";

myfile << "\n";
}
Expand Down Expand Up @@ -899,10 +899,10 @@ pcl::SVMClassify::saveClassificationResult (const char *filename)
output << "\n";
}

for (size_t i = 0; i < prediction_.size(); i++)
for (const auto &prediction : prediction_)
{
for (size_t j = 0; j < prediction_[i].size(); j++)
output << prediction_[i][j] << " ";
for (const double value : prediction)
output << value << " ";

output << "\n";
}
Expand Down

0 comments on commit ab5284a

Please sign in to comment.