Eval for 0.9.1
diff --git a/build.sbt b/build.sbt
index a7e7858..0115703 100644
--- a/build.sbt
+++ b/build.sbt
@@ -7,6 +7,7 @@
 organization := "io.prediction"
 
 libraryDependencies ++= Seq(
-  "io.prediction"    %% "core"          % pioVersion.value % "provided",
+  //"io.prediction"    %% "core"          % pioVersion.value % "provided",
+  "io.prediction"    %% "core"          % "0.9.1-SNAPSHOT" % "provided",
   "org.apache.spark" %% "spark-core"    % "1.2.0" % "provided",
   "org.apache.spark" %% "spark-mllib"   % "1.2.0" % "provided")
diff --git a/src/main/scala/Evaluation.scala b/src/main/scala/Evaluation.scala
index 8a903ee..a635b93 100644
--- a/src/main/scala/Evaluation.scala
+++ b/src/main/scala/Evaluation.scala
@@ -10,16 +10,15 @@
 import org.apache.spark.SparkContext._
 import org.apache.spark.rdd.RDD
 
-case class Precision 
-  extends AverageMetric[EmptyEvaluationInfo, 
-      Query, PredictedResult, ActualResult] {
+case class Accuracy
+  extends AverageMetric[EmptyEvaluationInfo, Query, PredictedResult, ActualResult] {
   def calculate(query: Query, predicted: PredictedResult, actual: ActualResult)
   : Double = (if (predicted.label == actual.label) 1.0 else 0.0)
 }
 
-object PrecisionEvaluation extends Evaluation {
+object AccuracyEvaluation extends Evaluation {
   // Define Engine and Metric used in Evaluation
-  engineMetric = (ClassificationEngine(), new Precision())
+  engineMetric = (ClassificationEngine(), new Accuracy())
 }
 
 object EngineParamsList extends EngineParamsGenerator {
@@ -29,7 +28,7 @@
   // the data is read, and a evalK parameter is used to define the
   // cross-validation.
   private[this] val baseEP = EngineParams(
-    dataSourceParams = DataSourceParams(appId = 18, evalK = Some(5)))
+    dataSourceParams = DataSourceParams(appId = 19, evalK = Some(5)))
 
   // Second, we specify the engine params list by explicitly listing all
   // algorithm parameters. In this case, we evaluate 3 engine params, each with