Skip to content

Commit

Permalink
Chore: rename mmlspark to synapseml
Browse files Browse the repository at this point in the history
  • Loading branch information
mhamilton723 committed Oct 14, 2021
1 parent ecc6868 commit 4232043
Show file tree
Hide file tree
Showing 562 changed files with 2,304 additions and 2,328 deletions.
2 changes: 1 addition & 1 deletion .chglog/CHANGELOG.tpl.md
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@
{{ end -}}

## Acknowledgements
We would like to acknowledge the developers and contributors, both internal and external who helped create this version of MMLSpark.\n
We would like to acknowledge the developers and contributors, both internal and external who helped create this version of SynapseML.\n

{{ end -}}

2 changes: 1 addition & 1 deletion .chglog/config.yml
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@ style: github
template: CHANGELOG.tpl.md
info:
title: CHANGELOG
repository_url: https://github.com/Azure/mmlspark
repository_url: https://github.com/Microsoft/SynapseML
options:
commit_groups:
title_maps:
Expand Down
4 changes: 2 additions & 2 deletions .github/ISSUE_TEMPLATE/bug_report.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ Steps to reproduce the behavior, code snippets encouraged
A clear and concise description of what you expected to happen.

**Info (please complete the following information):**
- MMLSpark Version: [e.g. v0.17]
- SynapseML Version: [e.g. v0.17]
- Spark Version [e.g. 2.4.3]
- Spark Platform [e.g. Databricks]

Expand All @@ -26,7 +26,7 @@ A clear and concise description of what you expected to happen.
Please post the stacktrace here if applicable
```

If the bug pertains to a specific feature please tag the appropriate [CODEOWNER](https://github.com/Azure/mmlspark/blob/master/CODEOWNERS) for better visibility
If the bug pertains to a specific feature please tag the appropriate [CODEOWNER](https://github.com/Microsoft/SynapseML/blob/master/CODEOWNERS) for better visibility

**Additional context**
Add any other context about the problem here.
2 changes: 1 addition & 1 deletion .github/config.yml
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ newPRWelcomeComment: >
- `style: Remove nulls from CNTKModel`
- `test: Add test coverage for CNTKModel`
Make sure to check out the [developer guide](https://github.com/Azure/mmlspark/blob/master/CONTRIBUTING.md) for guidance on testing your change.
Make sure to check out the [developer guide](https://github.com/Microsoft/SynapseML/blob/master/CONTRIBUTING.md) for guidance on testing your change.
# Configuration for first-pr-merge - https://github.com/behaviorbot/first-pr-merge

Expand Down
8 changes: 4 additions & 4 deletions CONTRIBUTING.md
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
## Interested in contributing to MMLSpark? We're excited to work with you.
## Interested in contributing to SynapseML? We're excited to work with you.

### You can contribute in many ways:

Expand Down Expand Up @@ -32,7 +32,7 @@ this process:

#### Implement your contribution

- Fork the MMLSpark repository.
- Fork the SynapseML repository.
- Implement your algorithm in Scala, using our wrapper generation mechanism to
produce PySpark bindings.
- Use SparkML `PipelineStage`s so your algorithm can be used as a part of
Expand All @@ -41,7 +41,7 @@ this process:
- Implement model saving and loading by extending SparkML `MLReadable`.
- Use good Scala style.
- Binary dependencies should be on Maven Central.
- See this [pull request](https://github.com/Azure/mmlspark/pull/22) for an
- See this [pull request](https://github.com/Microsoft/SynapseML/pull/22) for an
example contribution.

#### Implement tests
Expand All @@ -65,7 +65,7 @@ this process:

- In most cases, you should squash your commits into one.
- Open a pull request, and link it to the discussion issue you created earlier.
- An MMLSpark core team member will trigger a build to test your changes.
- An SynapseML core team member will trigger a build to test your changes.
- Fix any build failures. (The pull request will have comments from the build
with useful links.)
- Wait for code reviews from core team members and others.
Expand Down
94 changes: 47 additions & 47 deletions README.md

Large diffs are not rendered by default.

34 changes: 17 additions & 17 deletions build.sbt
Original file line number Diff line number Diff line change
Expand Up @@ -8,10 +8,10 @@ import scala.xml.transform.{RewriteRule, RuleTransformer}
import BuildUtils._
import xerial.sbt.Sonatype._

val condaEnvName = "mmlspark"
val condaEnvName = "synapseml"
val sparkVersion = "3.1.2"
name := "mmlspark"
ThisBuild / organization := "com.microsoft.ml.spark"
name := "synapseml"
ThisBuild / organization := "com.microsoft.azure"
ThisBuild / scalaVersion := "2.12.10"

val scalaMajorVersion = 2.12
Expand Down Expand Up @@ -118,7 +118,7 @@ generatePythonDoc := {
).value
val targetDir = artifactPath.in(packageBin).in(Compile).in(root).value.getParentFile
val codegenDir = join(targetDir, "generated")
val dir = join(codegenDir, "src", "python", "mmlspark")
val dir = join(codegenDir, "src", "python", "synapse","ml")
join(dir, "__init__.py").createNewFile()
runCmd(activateCondaEnv.value ++ Seq("sphinx-apidoc", "-f", "-o", "doc", "."), dir)
runCmd(activateCondaEnv.value ++ Seq("sphinx-build", "-b", "html", "doc", "../../../doc/pyspark"), dir)
Expand All @@ -145,7 +145,7 @@ publishDocs := {
uploadToBlob(unifiedDocDir.toString, version.value, "docs")
}

val release = TaskKey[Unit]("release", "publish the library to mmlspark blob")
val release = TaskKey[Unit]("release", "publish the library to synapseml blob")
release := Def.taskDyn {
val v = isSnapshot.value
if (!v) {
Expand All @@ -159,7 +159,7 @@ release := Def.taskDyn {
}
}

val publishBadges = TaskKey[Unit]("publishBadges", "publish badges to mmlspark blob")
val publishBadges = TaskKey[Unit]("publishBadges", "publish badges to synapseml blob")
publishBadges := {
def enc(s: String): String = {
s.replaceAllLiterally("_", "__").replaceAllLiterally(" ", "_").replaceAllLiterally("-", "--")
Expand Down Expand Up @@ -205,8 +205,8 @@ lazy val core = (project in file("core"))
sbtVersion,
baseDirectory
),
name := "mmlspark-core",
buildInfoPackage := "com.microsoft.ml.spark.build",
name := "synapseml-core",
buildInfoPackage := "com.microsoft.azure.synapse.ml.build",
): _*)

lazy val deepLearning = (project in file("deep-learning"))
Expand All @@ -217,23 +217,23 @@ lazy val deepLearning = (project in file("deep-learning"))
"com.microsoft.cntk" % "cntk" % "2.4",
"com.microsoft.onnxruntime" % "onnxruntime_gpu" % "1.8.1"
),
name := "mmlspark-deep-learning",
name := "synapseml-deep-learning",
): _*)

lazy val lightgbm = (project in file("lightgbm"))
.enablePlugins(SbtPlugin)
.dependsOn(core % "test->test;compile->compile")
.settings(settings ++ Seq(
libraryDependencies += ("com.microsoft.ml.lightgbm" % "lightgbmlib" % "3.2.110"),
name := "mmlspark-lightgbm"
name := "synapseml-lightgbm"
): _*)

lazy val vw = (project in file("vw"))
.enablePlugins(SbtPlugin)
.dependsOn(core % "test->test;compile->compile")
.settings(settings ++ Seq(
libraryDependencies += ("com.github.vowpalwabbit" % "vw-jni" % "8.9.1"),
name := "mmlspark-vw"
name := "synapseml-vw"
): _*)

lazy val cognitive = (project in file("cognitive"))
Expand All @@ -242,15 +242,15 @@ lazy val cognitive = (project in file("cognitive"))
.settings(settings ++ Seq(
libraryDependencies += ("com.microsoft.cognitiveservices.speech" % "client-sdk" % "1.14.0"),
resolvers += speechResolver,
name := "mmlspark-cognitive"
name := "synapseml-cognitive"
): _*)

lazy val opencv = (project in file("opencv"))
.enablePlugins(SbtPlugin)
.dependsOn(core % "test->test;compile->compile")
.settings(settings ++ Seq(
libraryDependencies += ("org.openpnp" % "opencv" % "3.2.0-1"),
name := "mmlspark-opencv"
name := "synapseml-opencv"
): _*)

lazy val root = (project in file("."))
Expand All @@ -265,7 +265,7 @@ lazy val root = (project in file("."))
.enablePlugins(ScalaUnidocPlugin && SbtPlugin)
.disablePlugins(CodegenPlugin)
.settings(settings ++ Seq(
name := "mmlspark",
name := "synapseml",
))

val setupTask = TaskKey[Unit]("setup", "set up library for intellij")
Expand All @@ -278,8 +278,8 @@ setupTask := {
}

sonatypeProjectHosting := Some(
GitHubHosting("Azure", "MMLSpark", "mmlspark-support@microsot.com"))
homepage := Some(url("https://github.com/Azure/mmlspark"))
GitHubHosting("Azure", "SynapseML", "mmlspark-support@microsot.com"))
homepage := Some(url("https://github.com/Microsoft/SynapseML"))
developers := List(
Developer("mhamilton723", "Mark Hamilton",
"mmlspark-support@microsoft.com", url("https://github.com/mhamilton723")),
Expand All @@ -289,7 +289,7 @@ developers := List(
"mmlspark-support@microsoft.com", url("https://github.com/drdarshan"))
)

licenses += ("MIT", url("https://github.com/Azure/mmlspark/blob/master/LICENSE"))
licenses += ("MIT", url("https://github.com/Microsoft/SynapseML/blob/master/LICENSE"))

credentials += Credentials("Sonatype Nexus Repository Manager",
"oss.sonatype.org",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,14 +14,14 @@

def streamToAzureSearch(df, **options):
jvm = SparkContext.getOrCreate()._jvm
writer = jvm.com.microsoft.ml.spark.cognitive.AzureSearchWriter
writer = jvm.com.microsoft.azure.synapse.ml.cognitive.AzureSearchWriter
return writer.stream(df._jdf, options)

setattr(pyspark.sql.DataFrame, 'streamToAzureSearch', streamToAzureSearch)

def writeToAzureSearch(df, **options):
jvm = SparkContext.getOrCreate()._jvm
writer = jvm.com.microsoft.ml.spark.cognitive.AzureSearchWriter
writer = jvm.com.microsoft.azure.synapse.ml.cognitive.AzureSearchWriter
writer.write(df._jdf, options)

setattr(pyspark.sql.DataFrame, 'writeToAzureSearch', writeToAzureSearch)
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,8 @@
if sys.version >= '3':
basestring = str

from mmlspark.cognitive._BingImageSearch import _BingImageSearch
from mmlspark.stages import Lambda
from synapse.ml.cognitive._BingImageSearch import _BingImageSearch
from synapse.ml.stages import Lambda
from pyspark.ml.common import inherit_doc
from pyspark.sql import SparkSession

Expand All @@ -32,10 +32,10 @@ def setMarketCol(self, value):

@staticmethod
def getUrlTransformer(imageCol, urlCol):
bis = SparkSession.builder.getOrCreate()._jvm.com.microsoft.ml.spark.cognitive.BingImageSearch
bis = SparkSession.builder.getOrCreate()._jvm.com.microsoft.azure.synapse.ml.cognitive.BingImageSearch
return Lambda._from_java(bis.getUrlTransformer(imageCol,urlCol))

@staticmethod
def downloadFromUrls(pathCol, bytesCol, concurrency, timeout):
bis = SparkSession.builder.getOrCreate()._jvm.com.microsoft.ml.spark.cognitive.BingImageSearch
bis = SparkSession.builder.getOrCreate()._jvm.com.microsoft.azure.synapse.ml.cognitive.BingImageSearch
return Lambda._from_java(bis.downloadFromUrls(pathCol, bytesCol, concurrency, timeout))
Original file line number Diff line number Diff line change
@@ -1,13 +1,13 @@
// Copyright (C) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License. See LICENSE in project root for information.

package com.microsoft.ml.spark.cognitive
package com.microsoft.azure.synapse.ml.cognitive

import com.microsoft.ml.spark.cognitive.AnomalyDetectorProtocol._
import com.microsoft.ml.spark.core.contracts.HasOutputCol
import com.microsoft.ml.spark.core.schema.DatasetExtensions
import com.microsoft.ml.spark.io.http.ErrorUtils
import com.microsoft.ml.spark.logging.BasicLogging
import com.microsoft.azure.synapse.ml.core.contracts.HasOutputCol
import com.microsoft.azure.synapse.ml.core.schema.DatasetExtensions
import com.microsoft.azure.synapse.ml.io.http.ErrorUtils
import com.microsoft.azure.synapse.ml.logging.BasicLogging
import AnomalyDetectorProtocol._
import org.apache.http.entity.{AbstractHttpEntity, StringEntity}
import org.apache.spark.injections.UDFUtils
import org.apache.spark.ml.ComplexParamsReadable
Expand Down
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
// Copyright (C) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License. See LICENSE in project root for information.

package com.microsoft.ml.spark.cognitive
package com.microsoft.azure.synapse.ml.cognitive

import com.microsoft.ml.spark.core.schema.SparkBindings
import com.microsoft.azure.synapse.ml.core.schema.SparkBindings
import spray.json.DefaultJsonProtocol._
import spray.json.RootJsonFormat

Expand Down
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
// Copyright (C) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License. See LICENSE in project root for information.

package com.microsoft.ml.spark
package com.microsoft.azure.synapse.ml.cognitive

import java.io.InputStream

Expand Down
Original file line number Diff line number Diff line change
@@ -1,27 +1,26 @@
// Copyright (C) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License. See LICENSE in project root for information.

package com.microsoft.ml.spark.cognitive
package com.microsoft.azure.synapse.ml.cognitive

import com.microsoft.ml.spark.io.http.{ErrorUtils, SimpleHTTPTransformer}
import com.microsoft.ml.spark.io.powerbi.StreamMaterializer
import com.microsoft.ml.spark.stages.{FixedMiniBatchTransformer, HasBatchSize, Lambda}
import com.microsoft.azure.synapse.ml.cognitive.AzureSearchProtocol._
import com.microsoft.azure.synapse.ml.io.http.{ErrorUtils, SimpleHTTPTransformer}
import com.microsoft.azure.synapse.ml.io.powerbi.StreamMaterializer
import com.microsoft.azure.synapse.ml.logging.BasicLogging
import com.microsoft.azure.synapse.ml.stages.{FixedMiniBatchTransformer, HasBatchSize, Lambda}
import org.apache.http.Consts
import org.apache.http.entity.{AbstractHttpEntity, ContentType, StringEntity}
import org.apache.log4j.{LogManager, Logger}
import org.apache.spark.injections.UDFUtils
import org.apache.spark.internal.{Logging => SLogging}
import org.apache.spark.ml.param._
import org.apache.spark.ml.util._
import org.apache.spark.ml.{ComplexParamsReadable, NamespaceInjections, PipelineModel}
import org.apache.spark.sql.functions.{col, expr, struct, to_json, udf}
import org.apache.spark.sql.functions.{col, expr, struct, to_json}
import org.apache.spark.sql.streaming.DataStreamWriter
import org.apache.spark.sql.types._
import org.apache.spark.sql.{DataFrame, Dataset, Row}
import com.microsoft.ml.spark.cognitive.AzureSearchProtocol._
import spray.json._
import DefaultJsonProtocol._
import com.microsoft.ml.spark.logging.BasicLogging
import org.apache.spark.injections.UDFUtils

import scala.collection.JavaConverters._

Expand Down
Original file line number Diff line number Diff line change
@@ -1,10 +1,10 @@
// Copyright (C) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License. See LICENSE in project root for information.

package com.microsoft.ml.spark.cognitive
package com.microsoft.azure.synapse.ml.cognitive

import com.microsoft.ml.spark.cognitive.AzureSearchProtocol._
import com.microsoft.ml.spark.cognitive.RESTHelpers._
import AzureSearchProtocol._
import RESTHelpers._
import org.apache.commons.io.IOUtils
import org.apache.http.client.methods.{HttpGet, HttpPost}
import org.apache.http.entity.StringEntity
Expand Down
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
// Copyright (C) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License. See LICENSE in project root for information.

package com.microsoft.ml.spark.cognitive
package com.microsoft.azure.synapse.ml.cognitive

import com.microsoft.ml.spark.core.schema.SparkBindings
import com.microsoft.azure.synapse.ml.core.schema.SparkBindings
import spray.json.DefaultJsonProtocol._
import spray.json.{JsonFormat, RootJsonFormat}

Expand Down
Original file line number Diff line number Diff line change
@@ -1,11 +1,11 @@
// Copyright (C) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License. See LICENSE in project root for information.

package com.microsoft.ml.spark.cognitive
package com.microsoft.azure.synapse.ml.cognitive

import com.microsoft.ml.spark.core.utils.AsyncUtils
import com.microsoft.ml.spark.logging.BasicLogging
import com.microsoft.ml.spark.stages.Lambda
import com.microsoft.azure.synapse.ml.core.utils.AsyncUtils
import com.microsoft.azure.synapse.ml.logging.BasicLogging
import com.microsoft.azure.synapse.ml.stages.Lambda
import org.apache.commons.io.IOUtils
import org.apache.http.client.methods.{HttpGet, HttpRequestBase}
import org.apache.http.entity.AbstractHttpEntity
Expand Down
Original file line number Diff line number Diff line change
@@ -1,14 +1,14 @@
// Copyright (C) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License. See LICENSE in project root for information.

package com.microsoft.ml.spark.cognitive

import com.microsoft.ml.spark.codegen.Wrappable
import com.microsoft.ml.spark.core.contracts.HasOutputCol
import com.microsoft.ml.spark.core.schema.DatasetExtensions
import com.microsoft.ml.spark.io.http._
import com.microsoft.ml.spark.logging.BasicLogging
import com.microsoft.ml.spark.stages.{DropColumns, Lambda}
package com.microsoft.azure.synapse.ml.cognitive

import com.microsoft.azure.synapse.ml.codegen.Wrappable
import com.microsoft.azure.synapse.ml.core.contracts.HasOutputCol
import com.microsoft.azure.synapse.ml.core.schema.DatasetExtensions
import com.microsoft.azure.synapse.ml.io.http._
import com.microsoft.azure.synapse.ml.logging.BasicLogging
import com.microsoft.azure.synapse.ml.stages.{DropColumns, Lambda}
import org.apache.http.NameValuePair
import org.apache.http.client.methods.{HttpEntityEnclosingRequestBase, HttpPost, HttpRequestBase}
import org.apache.http.client.utils.URLEncodedUtils
Expand Down
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
// Copyright (C) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License. See LICENSE in project root for information.

package com.microsoft.ml.spark.cognitive
package com.microsoft.azure.synapse.ml.cognitive

case class Rectangle(left: Int, top: Int, width: Int, height: Int)

Expand Down
Loading

0 comments on commit 4232043

Please sign in to comment.