/
build.gradle
133 lines (107 loc) · 4.83 KB
/
build.gradle
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
description = "Elasticsearch Spark (for Spark 2.X)"
evaluationDependsOn(':elasticsearch-hadoop-mr')
apply plugin: 'scala'
apply plugin: 'es.hadoop.build.integration'
apply plugin: 'scala.variants'
variants {
defaultVersion '2.11.12'
targetVersions '2.10.7', '2.11.12'
}
println "Compiled using Scala ${project.ext.scalaMajorVersion} [${project.ext.scalaVersion}]"
String sparkVersion = spark20Version
// Revert to spark 2.2.0 for scala 2.10 as 2.3+ does not support scala 2.10
if (project.ext.scalaMajorVersion == '2.10') {
sparkVersion = '2.2.0'
}
compileScala {
configure(scalaCompileOptions.forkOptions) {
memoryMaximumSize = '1g'
jvmArgs = ['-XX:MaxPermSize=512m']
}
scalaCompileOptions.additionalParameters = [
"-feature",
"-unchecked",
"-deprecation",
"-Xfuture",
"-Yno-adapted-args",
"-Ywarn-dead-code",
"-Ywarn-numeric-widen",
"-Xfatal-warnings"
]
sourceCompatibility = 1.6
targetCompatibility = 1.6
}
String coreSrc = file("$projectDir/../core").absolutePath.replace('\\','/')
sourceSets {
main.scala.srcDirs += "$coreSrc/main/scala"
test.scala.srcDirs += "$coreSrc/test/scala"
itest.java.srcDirs += "$coreSrc/itest/java"
itest.scala.srcDirs += "$coreSrc/itest/scala"
itest.resources.srcDirs += "$coreSrc/itest/resources"
}
// currently the outside project folders are transformed into linked resources however
// Gradle only supports one so the project will be invalid as not all sources will be in there
// as such, they are setup here manually for Eclipse. IntelliJ probably needs a similar approach
eclipse {
project.file.whenMerged { pj ->
// eliminated resources created by gradle
linkedResources.clear()
linkedResources.add(new org.gradle.plugins.ide.eclipse.model.Link("core/main/scala", "2", "$coreSrc/main/scala", null))
linkedResources.add(new org.gradle.plugins.ide.eclipse.model.Link("core/test/scala", "2", "$coreSrc/test/scala", null))
linkedResources.add(new org.gradle.plugins.ide.eclipse.model.Link("core/itest/java", "2", "$coreSrc/itest/java", null))
linkedResources.add(new org.gradle.plugins.ide.eclipse.model.Link("core/itest/scala", "2", "$coreSrc/itest/scala", null))
linkedResources.add(new org.gradle.plugins.ide.eclipse.model.Link("core/itest/resources","2", "$coreSrc/itest/resources", null))
}
classpath.file {
whenMerged { cp ->
entries.removeAll { entry ->
entry.kind == 'src' && (entry.path in ["scala", "java", "resources"] || entry.path.startsWith("itest-") || entry.path.endsWith("-scala"))
}
entries.add(new org.gradle.plugins.ide.eclipse.model.SourceFolder("core/main/scala", null))
entries.add(new org.gradle.plugins.ide.eclipse.model.SourceFolder("core/test/scala", null))
entries.add(new org.gradle.plugins.ide.eclipse.model.SourceFolder("core/itest/java", null))
entries.add(new org.gradle.plugins.ide.eclipse.model.SourceFolder("core/itest/scala", null))
entries.add(new org.gradle.plugins.ide.eclipse.model.SourceFolder("core/itest/resources", null))
}
}
}
dependencies {
provided(project(":elasticsearch-hadoop-mr"))
provided(project(path: ":elasticsearch-hadoop-mr", configuration:"compile"))
compile("org.scala-lang:scala-library:$scalaVersion")
provided("org.apache.spark:spark-core_${project.ext.scalaMajorVersion}:$sparkVersion") {
exclude group: 'javax.servlet'
exclude group: 'org.apache.hadoop'
}
optional("org.apache.spark:spark-sql_${project.ext.scalaMajorVersion}:$sparkVersion") {
exclude group: 'org.apache.hadoop'
}
optional("org.apache.spark:spark-streaming_${project.ext.scalaMajorVersion}:$sparkVersion") {
exclude group: 'org.apache.hadoop'
}
testCompile "org.elasticsearch:securemock:1.2"
testCompile(project(":elasticsearch-hadoop-mr").sourceSets.test.output)
itestCompile(project(":elasticsearch-hadoop-mr").sourceSets.itest.output)
itestCompile("org.apache.spark:spark-sql_${project.ext.scalaMajorVersion}:$sparkVersion") {
exclude group: 'org.apache.hadoop'
}
itestCompile("org.apache.spark:spark-streaming_${project.ext.scalaMajorVersion}:$sparkVersion") {
exclude group: 'org.apache.hadoop'
}
}
jar {
from(zipTree(project(":elasticsearch-hadoop-mr").jar.archivePath)) {
include "org/elasticsearch/hadoop/**"
include "esh-build.properties"
}
}
javadoc {
source += project(":elasticsearch-hadoop-mr").sourceSets.main.allJava
classpath += files(project(":elasticsearch-hadoop-mr").sourceSets.main.compileClasspath)
}
sourcesJar {
from project(":elasticsearch-hadoop-mr").sourceSets.main.allJava.srcDirs
}
scaladoc {
title = "${rootProject.description} ${version} API"
}