Stephen:

Your calls to best.svm() do not tune anything unless you specify the
parameter ranges (see the examples on the help page). Your calls are
just using the defaults which are very unlikely to yield models with
good performance.

[I think some day, I will have to remove the defaults in svm()...]

Another point: why aren't you using classification machines (which is
done automatically by providing a factor as dependent variable)?

There is classAgreement() in e1071, too, you might want to look at.

Cheers,
David






Hi  
 
I am doing this  sort of thing:
 
POLY:
 
> > obj = best.tune(svm, similarity ~., data = training, kernel =
"polynomial")
> summary(obj)
 
Call:
 best.tune(svm, similarity ~ ., data = training, kernel = "polynomial") 
 
Parameters:
   SVM-Type:  eps-regression 
 SVM-Kernel:  polynomial 
       cost:  1 
     degree:  3 
      gamma:  0.04545455 
     coef.0:  0 
    epsilon:  0.1 
 
 
Number of Support Vectors:  754
 
> svm.model <- svm(similarity ~., data = training, kernel  =
"polynomial", cost = 1, degree = 3, gamma = 0.04545455, coef.0 = 0,
epsilon = 0.1)
> pred=predict(svm.model, testing)
> pred[pred > .5] = 1
> pred[pred <= .5] = 0  
> table(testing$similarity, pred)
   pred
    0  1 
  0 30  8
  1 70 63
> obj = best.tune(svm, similarity ~., data = training, kernel =
"linear")
> summary(obj)
 
LINEAR:
 
Call:
 best.tune(svm, similarity ~ ., data = training, kernel = "linear") 
 
Parameters:
   SVM-Type:  eps-regression 
 SVM-Kernel:  linear 
       cost:  1 
      gamma:  0.04545455 
    epsilon:  0.1 
 
 
Number of Support Vectors:  697
 
> svm.model <- svm(similarity ~., data = training, kernel  = "linear",
cost = 1, gamma = 0.04545455, epsilon = 0.1)
> pred=predict(svm.model, testing)
> pred[pred > .5] = 1
> pred[pred <= .5] = 0  
> table(testing$similarity, pred)
   pred
    0   1  
  0   6  32
  1   4 129
 
 
RADIAL:
 
> obj = best.tune(svm, similarity ~., data = training, kernel =
"radial")
> summary(obj)
 
Call:
 best.tune(svm, similarity ~ ., data = training, kernel = "linear") 
 
Parameters:
   SVM-Type:  eps-regression 
 SVM-Kernel:  linear 
       cost:  1 
      gamma:  0.04545455 
    epsilon:  0.1 
 
 
Number of Support Vectors:  697
 
> svm.model <- svm(similarity ~., data = training, kernel  = "radial",
cost = 1, gamma = 0.04545455, epsilon = 0.1)
> pred=predict(svm.model, testing)
> pred[pred > .5] = 1
> pred[pred <= .5] = 0  
> table(testing$similarity, pred)
   pred
    0  1 
  0 27 11
  1 64 69
 
 
SIGMOID:
 
> obj = best.tune(svm, similarity ~., data = training, kernel =
"sigmoid")
> summary(obj)
 
Call:
 best.tune(svm, similarity ~ ., data = training, kernel = "sigmoid") 
 
Parameters:
   SVM-Type:  eps-regression 
 SVM-Kernel:  sigmoid 
       cost:  1 
      gamma:  0.04545455 
     coef.0:  0 
    epsilon:  0.1 
 
 
Number of Support Vectors:  986
 
> svm.model <- svm(similarity ~., data = training, kernel  = "sigmoid",
cost = 1, gamma = 0.04545455, coef.0 = 0, epsilon = 0.1)
> pred=predict(svm.model, testing)
> pred[pred > .5] = 1
> pred[pred <= .5] = 0  
> table(testing$similarity, pred)
   pred
    0   1  
  0   8  30
  1  26 107
>
 
and then taking out the kappa statistic to see if I am getting anything
significant.
 
I get kappas of 15 - 17% - I don't think that is very good.  I know
kappa is really for comparing the outcomes of two taggers but it seems a
good way to measure if your results might be by chance.
 
Two questions:
 
Any comments on Kappa and what it might be telling me?
 
What can I do to tune my kernels further?
 
Stephen
-- 
Dr. David Meyer
Department of Information Systems

Vienna University of Economics and Business Administration
Augasse 2-6, A-1090 Wien, Austria, Europe
Fax: +43-1-313 36x746 
Tel: +43-1-313 36x4393
HP:  http://wi.wu-wien.ac.at/~meyer/

______________________________________________
[EMAIL PROTECTED] mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html

Reply via email to