From ef2c4bea3f70f3e013d4eb7d1e413c47b8c262f7 Mon Sep 17 00:00:00 2001 From: Zheng RuiFeng Date: Tue, 6 Sep 2016 18:28:05 +0800 Subject: [PATCH] update --- .../main/scala/org/apache/spark/ml/feature/MaxAbsScaler.scala | 2 +- .../main/scala/org/apache/spark/ml/feature/MinMaxScaler.scala | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/mllib/src/main/scala/org/apache/spark/ml/feature/MaxAbsScaler.scala b/mllib/src/main/scala/org/apache/spark/ml/feature/MaxAbsScaler.scala index acabf0b892660..2fcfaa8dbda04 100644 --- a/mllib/src/main/scala/org/apache/spark/ml/feature/MaxAbsScaler.scala +++ b/mllib/src/main/scala/org/apache/spark/ml/feature/MaxAbsScaler.scala @@ -159,7 +159,7 @@ object MaxAbsScalerModel extends MLReadable[MaxAbsScalerModel] { override protected def saveImpl(path: String): Unit = { DefaultParamsWriter.saveMetadata(instance, path, sc) - val data = new Data(instance.maxAbs) + val data = Data(instance.maxAbs) val dataPath = new Path(path, "data").toString sparkSession.createDataFrame(Seq(data)).repartition(1).write.parquet(dataPath) } diff --git a/mllib/src/main/scala/org/apache/spark/ml/feature/MinMaxScaler.scala b/mllib/src/main/scala/org/apache/spark/ml/feature/MinMaxScaler.scala index 28cbe1cb01e9a..4dcc5dd3c8452 100644 --- a/mllib/src/main/scala/org/apache/spark/ml/feature/MinMaxScaler.scala +++ b/mllib/src/main/scala/org/apache/spark/ml/feature/MinMaxScaler.scala @@ -223,7 +223,7 @@ object MinMaxScalerModel extends MLReadable[MinMaxScalerModel] { override protected def saveImpl(path: String): Unit = { DefaultParamsWriter.saveMetadata(instance, path, sc) - val data = new Data(instance.originalMin, instance.originalMax) + val data = Data(instance.originalMin, instance.originalMax) val dataPath = new Path(path, "data").toString sparkSession.createDataFrame(Seq(data)).repartition(1).write.parquet(dataPath) }