[opengm] 268/386: seems to be working
Ghislain Vaillant
ghisvail-guest at moszumanska.debian.org
Wed Aug 31 08:38:08 UTC 2016
This is an automated email from the git hooks/post-receive script.
ghisvail-guest pushed a commit to branch debian/master
in repository opengm.
commit 1c9a6e17bab42aea1bd16fd5c43dbb68635e4a5f
Author: DerThorsten <thorsten.beier at iwr.uni-heidelberg.de>
Date: Tue Mar 10 16:46:56 2015 +0100
seems to be working
---
fubar/noisy_squares_3_labels.py | 4 ++--
include/opengm/learning/rws.hxx | 12 ++++++------
2 files changed, 8 insertions(+), 8 deletions(-)
diff --git a/fubar/noisy_squares_3_labels.py b/fubar/noisy_squares_3_labels.py
index 75e3482..2d9bf7a 100644
--- a/fubar/noisy_squares_3_labels.py
+++ b/fubar/noisy_squares_3_labels.py
@@ -98,10 +98,10 @@ infCls = opengm.inference.TrwsExternal
param = opengm.InfParam()
-learner = learning.subgradientSSVM(dataset, learningRate=1.0, C=0.9, learningMode='batch',maxIterations=200, averaging=-1)
+learner = learning.subgradientSSVM(dataset, learningRate=1.0, C=0.9, learningMode='batch',maxIterations=5, averaging=-1)
learner.learn(infCls=infCls,parameter=param,connectedComponents=False,infMode='n')
-learner = learning.rws(dataset, learningRate=1.0, C=0.05,maxIterations=5000, p=20, sigma=2.0)
+learner = learning.rws(dataset, learningRate=1.0, C=1.0,maxIterations=5000, p=100, sigma=1.3)
learner.learn(infCls=infCls,parameter=param,connectedComponents=False,infMode='n')
diff --git a/include/opengm/learning/rws.hxx b/include/opengm/learning/rws.hxx
index dadc60a..42c7cd0 100644
--- a/include/opengm/learning/rws.hxx
+++ b/include/opengm/learning/rws.hxx
@@ -210,14 +210,14 @@ namespace opengm {
opengm::infer<INF>(gm, para, arg);
lossVec[p] = dataset_.getLoss(arg, gmi);
- featureAcc_.accumulateModelFeatures(gm, dataset_.getGT(gmi).begin(), arg.begin());
+ //featureAcc_.accumulateModelFeatures(gm, dataset_.getGT(gmi).begin(), arg.begin());
// update weights
- const double wChange =updateWeights();
+ //const double wChange =updateWeights();
}
- for(size_t wi=0; wi<nWegihts; ++wi){
- gradient[wi] = featureAcc_.getWeight(wi);
- }
+ //for(size_t wi=0; wi<nWegihts; ++wi){
+ // gradient[wi] = featureAcc_.getWeight(wi);
+ //}
std::fill(gradient.begin(), gradient.end(),0.0);
for(size_t p=0; p<para_.p_; ++p){
for(size_t wi=0; wi<nWegihts; ++wi){
@@ -230,7 +230,7 @@ namespace opengm {
// do update
for(size_t wi=0; wi<nWegihts; ++wi){
const ValueType oldWeight = currentWeights[wi];
- const ValueType newWeights = (1.0 + oldWeight - actualLearningRate*gradient[wi])*para_.C_;
+ const ValueType newWeights = (oldWeight - actualLearningRate*gradient[wi])*para_.C_;
//std::cout<<"wi "<<newWeights<<"\n";
dataset_.getWeights().setWeight(wi, newWeights);
}
--
Alioth's /usr/local/bin/git-commit-notice on /srv/git.debian.org/git/debian-science/packages/opengm.git
More information about the debian-science-commits
mailing list