Skip to content

Commit

Permalink
HIVE-2935 : Implement HiveServer2 Core code changes (4th patch of 4) …
Browse files Browse the repository at this point in the history
…(Carl Steinbach and others via Ashutosh Chauhan)

git-svn-id: https://svn.apache.org/repos/asf/hive/trunk@1455659 13f79535-47bb-0310-9956-ffa450edef68
  • Loading branch information
ashutoshc committed Mar 12, 2013
1 parent fa8dec3 commit 6b27df5
Show file tree
Hide file tree
Showing 139 changed files with 22,019 additions and 624 deletions.
41 changes: 32 additions & 9 deletions ant/src/org/apache/hadoop/hive/ant/QTestGenTask.java
Expand Up @@ -108,6 +108,8 @@ public boolean accept(File filePath) {
} }


private List<String> templatePaths = new ArrayList<String>(); private List<String> templatePaths = new ArrayList<String>();

private String hiveRootDirectory;


private String outputDirectory; private String outputDirectory;


Expand Down Expand Up @@ -185,6 +187,18 @@ public String getTemplate() {
return template; return template;
} }


public void setHiveRootDirectory(File hiveRootDirectory) {
try {
this.hiveRootDirectory = hiveRootDirectory.getCanonicalPath();
} catch (IOException ioe) {
throw new BuildException(ioe);
}
}

public String getHiveRootDirectory() {
return hiveRootDirectory;
}

public void setTemplatePath(String templatePath) throws Exception { public void setTemplatePath(String templatePath) throws Exception {
templatePaths.clear(); templatePaths.clear();
for (String relativePath : templatePath.split(",")) { for (String relativePath : templatePath.split(",")) {
Expand Down Expand Up @@ -302,14 +316,15 @@ public void execute() throws BuildException {


List<File> qFiles = new ArrayList<File>(); List<File> qFiles = new ArrayList<File>();
HashMap<String, String> qFilesMap = new HashMap<String, String>(); HashMap<String, String> qFilesMap = new HashMap<String, String>();
File hiveRootDir = null;
File queryDir = null;
File outDir = null; File outDir = null;
File resultsDir = null; File resultsDir = null;
File logDir = null; File logDir = null;


try { try {
File inpDir = null;
if (queryDirectory != null) { if (queryDirectory != null) {
inpDir = new File(queryDirectory); queryDir = new File(queryDirectory);
} }


if (queryFile != null && !queryFile.equals("")) { if (queryFile != null && !queryFile.equals("")) {
Expand All @@ -318,31 +333,37 @@ public void execute() throws BuildException {
if (includeOnly != null && !includeOnly.contains(qFile)) { if (includeOnly != null && !includeOnly.contains(qFile)) {
continue; continue;
} }
if (null != inpDir) { if (null != queryDir) {
qFiles.add(new File(inpDir, qFile)); qFiles.add(new File(queryDir, qFile));
} else { } else {
qFiles.add(new File(qFile)); qFiles.add(new File(qFile));
} }
} }
} else if (queryFileRegex != null && !queryFileRegex.equals("")) { } else if (queryFileRegex != null && !queryFileRegex.equals("")) {
qFiles.addAll(Arrays.asList(inpDir.listFiles( qFiles.addAll(Arrays.asList(queryDir.listFiles(
new QFileRegexFilter(queryFileRegex, includeOnly)))); new QFileRegexFilter(queryFileRegex, includeOnly))));
} else if (runDisabled != null && runDisabled.equals("true")) { } else if (runDisabled != null && runDisabled.equals("true")) {
qFiles.addAll(Arrays.asList(inpDir.listFiles(new DisabledQFileFilter(includeOnly)))); qFiles.addAll(Arrays.asList(queryDir.listFiles(new DisabledQFileFilter(includeOnly))));
} else { } else {
qFiles.addAll(Arrays.asList(inpDir.listFiles(new QFileFilter(includeOnly)))); qFiles.addAll(Arrays.asList(queryDir.listFiles(new QFileFilter(includeOnly))));
} }


if (excludeQueryFile != null && !excludeQueryFile.equals("")) { if (excludeQueryFile != null && !excludeQueryFile.equals("")) {
// Exclude specified query files, comma separated // Exclude specified query files, comma separated
for (String qFile : excludeQueryFile.split(",")) { for (String qFile : excludeQueryFile.split(",")) {
if (null != inpDir) { if (null != queryDir) {
qFiles.remove(new File(inpDir, qFile)); qFiles.remove(new File(queryDir, qFile));
} else { } else {
qFiles.remove(new File(qFile)); qFiles.remove(new File(qFile));
} }
} }
} }

hiveRootDir = new File(hiveRootDirectory);
if (!hiveRootDir.exists()) {
throw new BuildException("Hive Root Directory "
+ hiveRootDir.getCanonicalPath() + " does not exist");
}


Collections.sort(qFiles); Collections.sort(qFiles);
for (File qFile : qFiles) { for (File qFile : qFiles) {
Expand Down Expand Up @@ -397,6 +418,8 @@ public void execute() throws BuildException {
// For each of the qFiles generate the test // For each of the qFiles generate the test
VelocityContext ctx = new VelocityContext(); VelocityContext ctx = new VelocityContext();
ctx.put("className", className); ctx.put("className", className);
ctx.put("hiveRootDir", getEscapedCanonicalPath(hiveRootDir));
ctx.put("queryDir", getEscapedCanonicalPath(queryDir));
ctx.put("qfiles", qFiles); ctx.put("qfiles", qFiles);
ctx.put("qfilesMap", qFilesMap); ctx.put("qfilesMap", qFilesMap);
ctx.put("resultsDir", getEscapedCanonicalPath(resultsDir)); ctx.put("resultsDir", getEscapedCanonicalPath(resultsDir));
Expand Down
5 changes: 2 additions & 3 deletions bin/ext/beeline.sh
Expand Up @@ -18,13 +18,12 @@ THISSERVICE=beeline
export SERVICE_LIST="${SERVICE_LIST}${THISSERVICE} " export SERVICE_LIST="${SERVICE_LIST}${THISSERVICE} "


beeline () { beeline () {

CLASS=org.apache.hive.beeline.BeeLine;
CLASS=org.apache.hive.jdbc.beeline.HiveBeeline;
execHiveCmd $CLASS "$@" execHiveCmd $CLASS "$@"
} }


beeline_help () { beeline_help () {
CLASS=org.apache.hive.jdbc.beeline.HiveBeeline; CLASS=org.apache.hive.beeline.BeeLine;
execHiveCmd $CLASS "--help" execHiveCmd $CLASS "--help"
} }


33 changes: 33 additions & 0 deletions bin/ext/hiveserver2.sh
@@ -0,0 +1,33 @@
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

THISSERVICE=hiveserver2
export SERVICE_LIST="${SERVICE_LIST}${THISSERVICE} "

hiveserver2() {
echo "Starting HiveServer2"
CLASS=org.apache.hive.service.server.HiveServer2
if $cygwin; then
HIVE_LIB=`cygpath -w "$HIVE_LIB"`
fi
JAR=${HIVE_LIB}/hive-service-*.jar

exec $HADOOP jar $JAR $CLASS "$@"
}

hiveserver2_help() {
hiveserver2 -h
}

21 changes: 21 additions & 0 deletions bin/hiveserver2
@@ -0,0 +1,21 @@
#!/usr/bin/env bash

# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

bin=`dirname "$0"`
bin=`cd "$bin"; pwd`

. "$bin"/hive --service hiveserver2 "$@"
41 changes: 38 additions & 3 deletions build-common.xml
Expand Up @@ -38,7 +38,9 @@
<property name="build.classes" location="${build.dir}/classes"/> <property name="build.classes" location="${build.dir}/classes"/>
<property name="build.encoding" value="ISO-8859-1"/> <property name="build.encoding" value="ISO-8859-1"/>


<!-- Thrift codegen properties -->
<property name="thrift.args" value="-I ${thrift.home} --gen java:beans,hashcode --gen cpp --gen php --gen py --gen rb"/> <property name="thrift.args" value="-I ${thrift.home} --gen java:beans,hashcode --gen cpp --gen php --gen py --gen rb"/>
<property name="thrift.gen.dir" value="${basedir}/src/gen/thrift"/>


<property name="hadoop.conf.dir" location="${hadoop.root}/conf"/> <property name="hadoop.conf.dir" location="${hadoop.root}/conf"/>


Expand Down Expand Up @@ -77,6 +79,7 @@
<pathelement location="" /> <pathelement location="" />
<pathelement location="${test.src.data.dir}/conf"/> <pathelement location="${test.src.data.dir}/conf"/>
<pathelement location="${hive.conf.dir}"/> <pathelement location="${hive.conf.dir}"/>
<pathelement location="${build.dir.hive}/beeline/test/classes"/>
<pathelement location="${build.dir.hive}/cli/test/classes"/> <pathelement location="${build.dir.hive}/cli/test/classes"/>
<pathelement location="${build.dir.hive}/common/test/classes"/> <pathelement location="${build.dir.hive}/common/test/classes"/>
<pathelement location="${build.dir.hive}/hbase-handler/test/classes"/> <pathelement location="${build.dir.hive}/hbase-handler/test/classes"/>
Expand All @@ -97,6 +100,7 @@
<!-- test directory may contain hadoop jars used by tests only (e.g. mini cluster) --> <!-- test directory may contain hadoop jars used by tests only (e.g. mini cluster) -->
<fileset dir="${hive.root}/build/ivy/lib/test" includes="*.jar" erroronmissingdir="false" <fileset dir="${hive.root}/build/ivy/lib/test" includes="*.jar" erroronmissingdir="false"
excludes="**/hive_*.jar,**/hive-*.jar"/> excludes="**/hive_*.jar,**/hive-*.jar"/>
<fileset dir="${hive.root}/build/ivy/lib/test" includes="hive-testutils*.jar" Erroronmissingdir="false"/>


<!-- we strip out hadoop jars present in places other than the hadoop shimmed dir--> <!-- we strip out hadoop jars present in places other than the hadoop shimmed dir-->
<fileset dir="${hive.root}/build/ivy/lib/default" includes="*.jar" erroronmissingdir="false" <fileset dir="${hive.root}/build/ivy/lib/default" includes="*.jar" erroronmissingdir="false"
Expand Down Expand Up @@ -181,6 +185,7 @@
<pathelement location="${build.dir.hive}/classes"/> <pathelement location="${build.dir.hive}/classes"/>
<fileset dir="${build.dir.hive}" includes="*/*.jar"/> <fileset dir="${build.dir.hive}" includes="*/*.jar"/>
<fileset dir="${hive.root}/lib" includes="*.jar"/> <fileset dir="${hive.root}/lib" includes="*.jar"/>
<fileset dir="${build.ivy.lib.dir}/default" includes="junit*.jar" />
<fileset dir="${build.ivy.lib.dir}/hadoop0.${hadoop.mr.rev}.shim" includes="*.jar" erroronmissingdir="false" /> <fileset dir="${build.ivy.lib.dir}/hadoop0.${hadoop.mr.rev}.shim" includes="*.jar" erroronmissingdir="false" />
<fileset dir="${build.ivy.lib.dir}/default" includes="*.jar" <fileset dir="${build.ivy.lib.dir}/default" includes="*.jar"
excludes="**/hadoop-*.jar" excludes="**/hadoop-*.jar"
Expand All @@ -193,6 +198,7 @@
<pathelement location="${build.dir.hive}/serde/classes"/> <pathelement location="${build.dir.hive}/serde/classes"/>
<pathelement location="${build.dir.hive}/metastore/classes"/> <pathelement location="${build.dir.hive}/metastore/classes"/>
<pathelement location="${build.dir.hive}/ql/classes"/> <pathelement location="${build.dir.hive}/ql/classes"/>
<pathelement location="${build.dir.hive}/beeline/classes"/>
<pathelement location="${build.dir.hive}/cli/classes"/> <pathelement location="${build.dir.hive}/cli/classes"/>
<pathelement location="${build.dir.hive}/shims/classes"/> <pathelement location="${build.dir.hive}/shims/classes"/>
<pathelement location="${build.dir.hive}/hwi/classes"/> <pathelement location="${build.dir.hive}/hwi/classes"/>
Expand Down Expand Up @@ -226,6 +232,7 @@
<echo message="Project: ${ant.project.name}"/> <echo message="Project: ${ant.project.name}"/>
<mkdir dir="${test.data.dir}"/> <mkdir dir="${test.data.dir}"/>
<mkdir dir="${test.log.dir}/clientpositive"/> <mkdir dir="${test.log.dir}/clientpositive"/>
<mkdir dir="${test.log.dir}/beelinepositive"/>
<mkdir dir="${test.log.dir}/clientnegative"/> <mkdir dir="${test.log.dir}/clientnegative"/>
<mkdir dir="${test.log.dir}/positive"/> <mkdir dir="${test.log.dir}/positive"/>
<mkdir dir="${test.log.dir}/negative"/> <mkdir dir="${test.log.dir}/negative"/>
Expand Down Expand Up @@ -280,7 +287,7 @@
<javac <javac
encoding="${build.encoding}" encoding="${build.encoding}"
srcdir="${test.src.dir}" srcdir="${test.src.dir}"
includes="org/apache/hadoop/**/*.java" includes="org/apache/**/hive/**/*.java"
excludes="**/TestSerDe.java" excludes="**/TestSerDe.java"
destdir="${test.build.classes}" destdir="${test.build.classes}"
debug="${javac.debug}" debug="${javac.debug}"
Expand All @@ -295,7 +302,7 @@
<javac <javac
encoding="${build.encoding}" encoding="${build.encoding}"
srcdir="${test.build.src}" srcdir="${test.build.src}"
includes="org/apache/hadoop/**/*.java" includes="org/apache/**/hive/**/*.java"
destdir="${test.build.classes}" destdir="${test.build.classes}"
debug="${javac.debug}" debug="${javac.debug}"
optimize="${javac.optimize}" optimize="${javac.optimize}"
Expand Down Expand Up @@ -352,6 +359,12 @@
</not> </not>
</condition> </condition>


<condition property="disableserver" value="false">
<not>
<isset property="disableserver"/>
</not>
</condition>

<condition property="clustermode" value=""> <condition property="clustermode" value="">
<not> <not>
<isset property="clustermode"/> <isset property="clustermode"/>
Expand Down Expand Up @@ -413,15 +426,18 @@
</then> </then>
</if> </if>
<junit showoutput="${test.output}" printsummary="yes" haltonfailure="no" <junit showoutput="${test.output}" printsummary="yes" haltonfailure="no"
fork="yes" maxmemory="512m" dir="${basedir}" timeout="${test.junit.timeout}" fork="yes" maxmemory="${test.junit.maxmemory}" dir="${basedir}" timeout="${test.junit.timeout}"
errorProperty="tests.failed" failureProperty="tests.failed" filtertrace="off"> errorProperty="tests.failed" failureProperty="tests.failed" filtertrace="off">
<jvmarg value="-XX:+HeapDumpOnOutOfMemoryError"/>
<jvmarg value="-XX:HeapDumpPath=${hive.root}"/>
<env key="LANG" value="${test.lang}"/> <env key="LANG" value="${test.lang}"/>
<env key="HIVE_HADOOP_TEST_CLASSPATH" value="${hadoop.testcp}"/> <env key="HIVE_HADOOP_TEST_CLASSPATH" value="${hadoop.testcp}"/>
<env key="HADOOP_HOME" value="${hadoop.root}"/> <env key="HADOOP_HOME" value="${hadoop.root}"/>
<env key="HADOOP_CLASSPATH" path="${test.src.data.dir}/conf:${build.dir.hive}/dist/lib/derby-${derby.version}.jar:${build.dir.hive}/dist/lib/JavaEWAH-${javaewah.version}.jar:${hadoop.root}/modules/*"/> <!-- Modules needed for Hadoop 0.23 --> <env key="HADOOP_CLASSPATH" path="${test.src.data.dir}/conf:${build.dir.hive}/dist/lib/derby-${derby.version}.jar:${build.dir.hive}/dist/lib/JavaEWAH-${javaewah.version}.jar:${hadoop.root}/modules/*"/> <!-- Modules needed for Hadoop 0.23 -->
<env key="TZ" value="US/Pacific"/> <env key="TZ" value="US/Pacific"/>
<sysproperty key="test.output.overwrite" value="${overwrite}"/> <sysproperty key="test.output.overwrite" value="${overwrite}"/>
<sysproperty key="test.service.standalone.server" value="${standalone}"/> <sysproperty key="test.service.standalone.server" value="${standalone}"/>
<sysproperty key="test.service.disable.server" value="${disableserver}"/>
<sysproperty key="log4j.configuration" value="file:///${test.src.data.dir}/conf/hive-log4j.properties"/> <sysproperty key="log4j.configuration" value="file:///${test.src.data.dir}/conf/hive-log4j.properties"/>
<sysproperty key="derby.stream.error.file" value="${test.build.dir}/derby.log"/> <sysproperty key="derby.stream.error.file" value="${test.build.dir}/derby.log"/>
<sysproperty key="hive.aux.jars.path" value="file:///${test.build.dir}/test-udfs.jar"/> <sysproperty key="hive.aux.jars.path" value="file:///${test.build.dir}/test-udfs.jar"/>
Expand All @@ -440,9 +456,11 @@
<sysproperty key="build.dir.hive" value="${build.dir.hive}"/> <sysproperty key="build.dir.hive" value="${build.dir.hive}"/>
<sysproperty key="build.ivy.lib.dir" value="${build.ivy.lib.dir}"/> <sysproperty key="build.ivy.lib.dir" value="${build.ivy.lib.dir}"/>
<sysproperty key="derby.version" value="${derby.version}"/> <sysproperty key="derby.version" value="${derby.version}"/>
<sysproperty key="hive.root" value="${hive.root}"/>
<sysproperty key="hive.version" value="${version}"/> <sysproperty key="hive.version" value="${version}"/>
<sysproperty key="java.net.preferIPv4Stack" value="${java.net.preferIPv4Stack}"/> <sysproperty key="java.net.preferIPv4Stack" value="${java.net.preferIPv4Stack}"/>
<sysproperty key="hadoop.bin.path" value="${test.hadoop.bin.path}${junit.script.extension}"/> <sysproperty key="hadoop.bin.path" value="${test.hadoop.bin.path}${junit.script.extension}"/>
<sysproperty key="test.concurrency.num.threads" value="${test.concurrency.num.threads}"/>
<jvmarg line="${jvm.args}"/> <jvmarg line="${jvm.args}"/>


<classpath refid="test.local.classpath"/> <classpath refid="test.local.classpath"/>
Expand Down Expand Up @@ -486,6 +504,23 @@
</condition> </condition>
</target> </target>


<target name="thriftif" depends="check-thrift-home">
<echo message="Project: ${ant.project.name}"/>
<delete dir="${thrift.gen.dir}"/>
<mkdir dir="${thrift.gen.dir}"/>
<for param="thrift.file">
<path>
<fileset dir="." includes="if/*.thrift,if/test/*.thrift" />
</path>
<sequential>
<echo message="Generating Thrift code for @{thrift.file}"/>
<exec executable="${thrift.home}/bin/thrift" failonerror="true" dir=".">
<arg line="${thrift.args} -I ${basedir}/include -I ${basedir}/.. -o ${thrift.gen.dir} @{thrift.file} " />
</exec>
</sequential>
</for>
</target>

<target name="check-ivy" depends="ivy-init-settings"> <target name="check-ivy" depends="ivy-init-settings">
<echo message="Project: ${ant.project.name}"/> <echo message="Project: ${ant.project.name}"/>
<available file="${basedir}/ivy.xml" property="ivy.present"/> <available file="${basedir}/ivy.xml" property="ivy.present"/>
Expand Down
17 changes: 13 additions & 4 deletions build.properties
Expand Up @@ -72,8 +72,8 @@ jsp.test.jar=${hadoop.root}/lib/jetty-ext/jsp-api.jar
common.jar=${hadoop.root}/lib/commons-httpclient-3.0.1.jar common.jar=${hadoop.root}/lib/commons-httpclient-3.0.1.jar


# module names needed for build process # module names needed for build process
iterate.hive.all=ant,shims,common,serde,metastore,ql,contrib,service,cli,jdbc,hwi,hbase-handler,pdk,builtins iterate.hive.all=ant,shims,common,serde,metastore,ql,contrib,service,cli,jdbc,beeline,hwi,hbase-handler,pdk,builtins,testutils
iterate.hive.modules=shims,common,serde,metastore,ql,contrib,service,cli,jdbc,hwi,hbase-handler,pdk,builtins iterate.hive.modules=shims,common,serde,metastore,ql,contrib,service,cli,jdbc,beeline,hwi,hbase-handler,pdk,builtins,testutils
iterate.hive.tests=ql,contrib,hbase-handler,hwi,jdbc,metastore,odbc,serde,service iterate.hive.tests=ql,contrib,hbase-handler,hwi,jdbc,metastore,odbc,serde,service
iterate.hive.thrift=ql,service,metastore,serde iterate.hive.thrift=ql,service,metastore,serde
iterate.hive.protobuf=ql iterate.hive.protobuf=ql
Expand All @@ -92,7 +92,16 @@ test.junit.timeout=43200000
# Use this property to selectively disable tests from the command line: # Use this property to selectively disable tests from the command line:
# ant test -Dtest.junit.exclude="**/TestCliDriver.class" # ant test -Dtest.junit.exclude="**/TestCliDriver.class"
# ant test -Dtest.junit.exclude="**/Test*CliDriver.class,**/TestPartitions.class" # ant test -Dtest.junit.exclude="**/Test*CliDriver.class,**/TestPartitions.class"
test.junit.exclude= test.junit.exclude="**/TestBeeLineDriver.class, **/TestHiveServer2Concurrency.class"
test.continue.on.failure=false

test.submodule.exclude=
test.junit.maxmemory=512m

test.concurrency.num.threads=1
#test.beelinepositive.exclude=add_part_exist.q,alter1.q,alter2.q,alter4.q,alter5.q,alter_rename_partition.q,alter_rename_partition_authorization.q,archive.q,archive_corrupt.q,archive_multi.q,archive_mr_1806.q,archive_multi_mr_1806.q,authorization_1.q,authorization_2.q,authorization_4.q,authorization_5.q,authorization_6.q,authorization_7.q,ba_table1.q,ba_table2.q,ba_table3.q,ba_table_udfs.q,binary_table_bincolserde.q,binary_table_colserde.q,cluster.q,columnarserde_create_shortcut.q,combine2.q,constant_prop.q,create_nested_type.q,create_or_replace_view.q,create_struct_table.q,create_union_table.q,database.q,database_location.q,database_properties.q,ddltime.q,describe_database_json.q,drop_database_removes_partition_dirs.q,escape1.q,escape2.q,exim_00_nonpart_empty.q,exim_01_nonpart.q,exim_02_00_part_empty.q,exim_02_part.q,exim_03_nonpart_over_compat.q,exim_04_all_part.q,exim_04_evolved_parts.q,exim_05_some_part.q,exim_06_one_part.q,exim_07_all_part_over_nonoverlap.q,exim_08_nonpart_rename.q,exim_09_part_spec_nonoverlap.q,exim_10_external_managed.q,exim_11_managed_external.q,exim_12_external_location.q,exim_13_managed_location.q,exim_14_managed_location_over_existing.q,exim_15_external_part.q,exim_16_part_external.q,exim_17_part_managed.q,exim_18_part_external.q,exim_19_00_part_external_location.q,exim_19_part_external_location.q,exim_20_part_managed_location.q,exim_21_export_authsuccess.q,exim_22_import_exist_authsuccess.q,exim_23_import_part_authsuccess.q,exim_24_import_nonexist_authsuccess.q,global_limit.q,groupby_complex_types.q,groupby_complex_types_multi_single_reducer.q,index_auth.q,index_auto.q,index_auto_empty.q,index_bitmap.q,index_bitmap1.q,index_bitmap2.q,index_bitmap3.q,index_bitmap_auto.q,index_bitmap_rc.q,index_compact.q,index_compact_1.q,index_compact_2.q,index_compact_3.q,index_stale_partitioned.q,init_file.q,input16.q,input16_cc.q,input46.q,input_columnarserde.q,input_dynamicserde.q,input_lazyserde.q,input_testxpath3.q,input_testxpath4.q,insert2_overwrite_partitions.q,insertexternal1.q,join_thrift.q,lateral_view.q,load_binary_data.q,load_exist_part_authsuccess.q,load_nonpart_authsuccess.q,load_part_authsuccess.q,loadpart_err.q,lock1.q,lock2.q,lock3.q,lock4.q,merge_dynamic_partition.q,multi_insert.q,multi_insert_move_tasks_share_dependencies.q,null_column.q,ppd_clusterby.q,query_with_semi.q,rename_column.q,sample6.q,sample_islocalmode_hook.q,set_processor_namespaces.q,show_tables.q,source.q,split_sample.q,str_to_map.q,transform1.q,udaf_collect_set.q,udaf_context_ngrams.q,udaf_histogram_numeric.q,udaf_ngrams.q,udaf_percentile_approx.q,udf_array.q,udf_bitmap_and.q,udf_bitmap_or.q,udf_explode.q,udf_format_number.q,udf_map.q,udf_map_keys.q,udf_map_values.q,udf_max.q,udf_min.q,udf_named_struct.q,udf_percentile.q,udf_printf.q,udf_sentences.q,udf_sort_array.q,udf_split.q,udf_struct.q,udf_substr.q,udf_translate.q,udf_union.q,udf_xpath.q,udtf_stack.q,view.q,virtual_column.q




# #
# Ivy Properties # Ivy Properties
Expand All @@ -108,7 +117,7 @@ ivy.changingPattern=.*SNAPSHOT
ivy.publish.pattern=[artifact]-[revision].[ext] ivy.publish.pattern=[artifact]-[revision].[ext]
ivy.artifact.retrieve.pattern=[conf]/[artifact]-[revision](-[classifier]).[ext] ivy.artifact.retrieve.pattern=[conf]/[artifact]-[revision](-[classifier]).[ext]
ivysettings.xml=${ivy.conf.dir}/ivysettings.xml ivysettings.xml=${ivy.conf.dir}/ivysettings.xml
ivyresolvelog=download-only ivyresolvelog=default
ivy.mvn.repo=http://repo2.maven.org/maven2 ivy.mvn.repo=http://repo2.maven.org/maven2
ivy_repo_url=${ivy.mvn.repo}/org/apache/ivy/ivy/${ivy.version}/ivy-${ivy.version}.jar ivy_repo_url=${ivy.mvn.repo}/org/apache/ivy/ivy/${ivy.version}/ivy-${ivy.version}.jar
hive.ivy.org=org.apache.hive hive.ivy.org=org.apache.hive
Expand Down

0 comments on commit 6b27df5

Please sign in to comment.