Repository: bigtop Updated Branches: refs/heads/master c716b1272 -> 44aa7bb0f
BIGTOP-2968: Bump Hive version to 2.3.2 Project: http://git-wip-us.apache.org/repos/asf/bigtop/repo Commit: http://git-wip-us.apache.org/repos/asf/bigtop/commit/44aa7bb0 Tree: http://git-wip-us.apache.org/repos/asf/bigtop/tree/44aa7bb0 Diff: http://git-wip-us.apache.org/repos/asf/bigtop/diff/44aa7bb0 Branch: refs/heads/master Commit: 44aa7bb0f6abdd513f8a90fb78bf6ce9051cc487 Parents: c716b12 Author: YoungWoo Kim <[email protected]> Authored: Sun Nov 8 00:07:16 2015 +0900 Committer: Youngwoo Kim <[email protected]> Committed: Tue Dec 26 22:50:03 2017 +0900 ---------------------------------------------------------------------- .../src/common/hive/do-component-build | 14 +- .../src/common/hive/patch0-HIVE-16302.diff | 16 ++ bigtop-packages/src/common/hive/patch0.diff | 228 ------------------- .../src/common/hive/patch1-HIVE-12875.diff | 60 ----- .../src/common/tez/do-component-build | 2 +- bigtop-packages/src/deb/hive/hive-jdbc.install | 10 +- bigtop-packages/src/deb/hive/rules | 5 +- bigtop-packages/src/rpm/hive/SPECS/hive.spec | 29 +-- bigtop-packages/src/rpm/tez/SPECS/tez.spec | 2 +- bigtop.bom | 2 +- 10 files changed, 35 insertions(+), 333 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/bigtop/blob/44aa7bb0/bigtop-packages/src/common/hive/do-component-build ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/hive/do-component-build b/bigtop-packages/src/common/hive/do-component-build index f8be82c..9223e8f 100644 --- a/bigtop-packages/src/common/hive/do-component-build +++ b/bigtop-packages/src/common/hive/do-component-build @@ -20,19 +20,17 @@ set -ex HIVE_MAVEN_OPTS=" -Dhbase.version=$HBASE_VERSION \ -Dzookeeper.version=$ZOOKEEPER_VERSION \ --Dhadoop.mr.rev=23 \ --Dhadoop.security.version=$HADOOP_VERSION \ --Dhadoop-23.version=$HADOOP_VERSION \ --Dhbase.hadoop2.version=$HBASE_VERSION \ --Dmvn.hadoop.profile=hadoop23 \ +-Dhadoop.version=$HADOOP_VERSION \ -DskipTests \ --Dhbase.version.with.hadoop.version=$HBASE_VERSION \ -Dtez.version=${TEZ_VERSION} \ --Dspark.version=${SPARK1_VERSION} +-Dspark.version=${SPARK_VERSION} " export MAVEN_OPTS="-Xmx1500m -Xms1500m" -mvn ${HIVE_MAVEN_OPTS} clean install -Phadoop-2,dist "$@" +mvn ${HIVE_MAVEN_OPTS} clean install -Pdist "$@" mkdir -p build/dist tar -C build/dist --strip-components=1 -xzf packaging/target/apache-hive-${HIVE_VERSION}-bin.tar.gz + +# HIVE-13177 +sed -i "s#jdbcStandaloneJarPath=\`ls \${HIVE_LIB}/hive-jdbc-\*-standalone.jar\`#jdbcStandaloneJarPath=#" build/dist/bin/ext/beeline.sh http://git-wip-us.apache.org/repos/asf/bigtop/blob/44aa7bb0/bigtop-packages/src/common/hive/patch0-HIVE-16302.diff ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/hive/patch0-HIVE-16302.diff b/bigtop-packages/src/common/hive/patch0-HIVE-16302.diff new file mode 100644 index 0000000..96cd99e --- /dev/null +++ b/bigtop-packages/src/common/hive/patch0-HIVE-16302.diff @@ -0,0 +1,16 @@ +diff --git a/shims/common/pom.xml b/shims/common/pom.xml +index 19821cd..e2245df 100644 +--- a/shims/common/pom.xml ++++ b/shims/common/pom.xml +@@ -86,6 +86,11 @@ + </exclusion> + </exclusions> + </dependency> ++ <dependency> ++ <groupId>junit</groupId> ++ <artifactId>junit</artifactId> ++ <scope>test</scope> ++ </dependency> + </dependencies> + + <build> http://git-wip-us.apache.org/repos/asf/bigtop/blob/44aa7bb0/bigtop-packages/src/common/hive/patch0.diff ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/hive/patch0.diff b/bigtop-packages/src/common/hive/patch0.diff deleted file mode 100644 index d70e694..0000000 --- a/bigtop-packages/src/common/hive/patch0.diff +++ /dev/null @@ -1,228 +0,0 @@ -diff --git a/pom.xml b/pom.xml -index 3c0bc9b..9bbee65 100644 ---- a/pom.xml -+++ b/pom.xml -@@ -157,7 +157,7 @@ - <ST4.version>4.0.4</ST4.version> - <tez.version>0.5.2</tez.version> - <super-csv.version>2.2.0</super-csv.version> -- <spark.version>1.3.1</spark.version> -+ <spark.version>1.5.1</spark.version> - <scala.binary.version>2.10</scala.binary.version> - <scala.version>2.10.4</scala.version> - <tempus-fugit.version>1.1</tempus-fugit.version> -diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/impl/JobMetricsListener.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/impl/JobMetricsListener.java -index 51772cd..52f4b9c 100644 ---- a/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/impl/JobMetricsListener.java -+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/spark/status/impl/JobMetricsListener.java -@@ -23,29 +23,15 @@ - - import org.apache.commons.logging.Log; - import org.apache.commons.logging.LogFactory; -+import org.apache.spark.JavaSparkListener; - import org.apache.spark.executor.TaskMetrics; --import org.apache.spark.scheduler.SparkListener; --import org.apache.spark.scheduler.SparkListenerApplicationEnd; --import org.apache.spark.scheduler.SparkListenerApplicationStart; --import org.apache.spark.scheduler.SparkListenerBlockManagerAdded; --import org.apache.spark.scheduler.SparkListenerBlockManagerRemoved; --import org.apache.spark.scheduler.SparkListenerEnvironmentUpdate; --import org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate; --import org.apache.spark.scheduler.SparkListenerJobEnd; - import org.apache.spark.scheduler.SparkListenerJobStart; --import org.apache.spark.scheduler.SparkListenerStageCompleted; --import org.apache.spark.scheduler.SparkListenerStageSubmitted; - import org.apache.spark.scheduler.SparkListenerTaskEnd; --import org.apache.spark.scheduler.SparkListenerTaskGettingResult; --import org.apache.spark.scheduler.SparkListenerTaskStart; --import org.apache.spark.scheduler.SparkListenerUnpersistRDD; --import org.apache.spark.scheduler.SparkListenerExecutorRemoved; --import org.apache.spark.scheduler.SparkListenerExecutorAdded; - - import com.google.common.collect.Lists; - import com.google.common.collect.Maps; - --public class JobMetricsListener implements SparkListener { -+public class JobMetricsListener extends JavaSparkListener { - - private static final Log LOG = LogFactory.getLog(JobMetricsListener.class); - -@@ -54,36 +40,6 @@ - private final Map<Integer, Map<String, List<TaskMetrics>>> allJobMetrics = Maps.newHashMap(); - - @Override -- public void onExecutorRemoved(SparkListenerExecutorRemoved removed) { -- -- } -- -- @Override -- public void onExecutorAdded(SparkListenerExecutorAdded added) { -- -- } -- -- @Override -- public void onStageCompleted(SparkListenerStageCompleted stageCompleted) { -- -- } -- -- @Override -- public void onStageSubmitted(SparkListenerStageSubmitted stageSubmitted) { -- -- } -- -- @Override -- public void onTaskStart(SparkListenerTaskStart taskStart) { -- -- } -- -- @Override -- public void onTaskGettingResult(SparkListenerTaskGettingResult taskGettingResult) { -- -- } -- -- @Override - public synchronized void onTaskEnd(SparkListenerTaskEnd taskEnd) { - int stageId = taskEnd.stageId(); - int stageAttemptId = taskEnd.stageAttemptId(); -@@ -119,46 +75,6 @@ public synchronized void onJobStart(SparkListenerJobStart jobStart) { - jobIdToStageId.put(jobId, intStageIds); - } - -- @Override -- public synchronized void onJobEnd(SparkListenerJobEnd jobEnd) { -- -- } -- -- @Override -- public void onEnvironmentUpdate(SparkListenerEnvironmentUpdate environmentUpdate) { -- -- } -- -- @Override -- public void onBlockManagerAdded(SparkListenerBlockManagerAdded blockManagerAdded) { -- -- } -- -- @Override -- public void onBlockManagerRemoved(SparkListenerBlockManagerRemoved blockManagerRemoved) { -- -- } -- -- @Override -- public void onUnpersistRDD(SparkListenerUnpersistRDD unpersistRDD) { -- -- } -- -- @Override -- public void onApplicationStart(SparkListenerApplicationStart applicationStart) { -- -- } -- -- @Override -- public void onApplicationEnd(SparkListenerApplicationEnd applicationEnd) { -- -- } -- -- @Override -- public void onExecutorMetricsUpdate(SparkListenerExecutorMetricsUpdate executorMetricsUpdate) { -- -- } -- - public synchronized Map<String, List<TaskMetrics>> getJobMetric(int jobId) { - return allJobMetrics.get(jobId); - } -diff --git a/spark-client/src/main/java/org/apache/hive/spark/client/RemoteDriver.java b/spark-client/src/main/java/org/apache/hive/spark/client/RemoteDriver.java -index b77c9e8..f5b1e48 100644 ---- a/spark-client/src/main/java/org/apache/hive/spark/client/RemoteDriver.java -+++ b/spark-client/src/main/java/org/apache/hive/spark/client/RemoteDriver.java -@@ -43,26 +43,13 @@ - import org.apache.hive.spark.client.rpc.Rpc; - import org.apache.hive.spark.client.rpc.RpcConfiguration; - import org.apache.hive.spark.counter.SparkCounters; -+import org.apache.spark.JavaSparkListener; - import org.apache.spark.SparkConf; - import org.apache.spark.api.java.JavaFutureAction; - import org.apache.spark.api.java.JavaSparkContext; --import org.apache.spark.scheduler.SparkListener; --import org.apache.spark.scheduler.SparkListenerApplicationEnd; --import org.apache.spark.scheduler.SparkListenerApplicationStart; --import org.apache.spark.scheduler.SparkListenerBlockManagerAdded; --import org.apache.spark.scheduler.SparkListenerBlockManagerRemoved; --import org.apache.spark.scheduler.SparkListenerEnvironmentUpdate; --import org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate; - import org.apache.spark.scheduler.SparkListenerJobEnd; - import org.apache.spark.scheduler.SparkListenerJobStart; --import org.apache.spark.scheduler.SparkListenerStageCompleted; --import org.apache.spark.scheduler.SparkListenerStageSubmitted; - import org.apache.spark.scheduler.SparkListenerTaskEnd; --import org.apache.spark.scheduler.SparkListenerTaskGettingResult; --import org.apache.spark.scheduler.SparkListenerTaskStart; --import org.apache.spark.scheduler.SparkListenerUnpersistRDD; --import org.apache.spark.scheduler.SparkListenerExecutorRemoved; --import org.apache.spark.scheduler.SparkListenerExecutorAdded; - import org.slf4j.Logger; - import org.slf4j.LoggerFactory; - -@@ -438,21 +425,11 @@ private void monitorJob(JavaFutureAction<?> job, - - } - -- private class ClientListener implements SparkListener { -+ private class ClientListener extends JavaSparkListener { - - private final Map<Integer, Integer> stageToJobId = Maps.newHashMap(); - - @Override -- public void onExecutorRemoved(SparkListenerExecutorRemoved removed) { -- -- } -- -- @Override -- public void onExecutorAdded(SparkListenerExecutorAdded added) { -- -- } -- -- @Override - public void onJobStart(SparkListenerJobStart jobStart) { - synchronized (stageToJobId) { - for (int i = 0; i < jobStart.stageIds().length(); i++) { -@@ -500,39 +477,6 @@ public void onTaskEnd(SparkListenerTaskEnd taskEnd) { - } - } - -- @Override -- public void onStageCompleted(SparkListenerStageCompleted stageCompleted) { } -- -- @Override -- public void onStageSubmitted(SparkListenerStageSubmitted stageSubmitted) { } -- -- @Override -- public void onTaskStart(SparkListenerTaskStart taskStart) { } -- -- @Override -- public void onTaskGettingResult(SparkListenerTaskGettingResult taskGettingResult) { } -- -- @Override -- public void onEnvironmentUpdate(SparkListenerEnvironmentUpdate environmentUpdate) { } -- -- @Override -- public void onBlockManagerAdded(SparkListenerBlockManagerAdded blockManagerAdded) { } -- -- @Override -- public void onBlockManagerRemoved(SparkListenerBlockManagerRemoved blockManagerRemoved) { } -- -- @Override -- public void onUnpersistRDD(SparkListenerUnpersistRDD unpersistRDD) { } -- -- @Override -- public void onApplicationStart(SparkListenerApplicationStart applicationStart) { } -- -- @Override -- public void onApplicationEnd(SparkListenerApplicationEnd applicationEnd) { } -- -- @Override -- public void onExecutorMetricsUpdate(SparkListenerExecutorMetricsUpdate executorMetricsUpdate) { } -- - /** - * Returns the client job ID for the given Spark job ID. - * http://git-wip-us.apache.org/repos/asf/bigtop/blob/44aa7bb0/bigtop-packages/src/common/hive/patch1-HIVE-12875.diff ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/hive/patch1-HIVE-12875.diff b/bigtop-packages/src/common/hive/patch1-HIVE-12875.diff deleted file mode 100644 index 7a6acda..0000000 --- a/bigtop-packages/src/common/hive/patch1-HIVE-12875.diff +++ /dev/null @@ -1,60 +0,0 @@ ---- apache-hive-1.2.1-src/ql/src/java/org/apache/hadoop/hive/ql/Driver.java.orig 2015-06-18 22:51:23.000000000 +0200 -+++ apache-hive-1.2.1-src/ql/src/java/org/apache/hadoop/hive/ql/Driver.java 2016-01-27 14:34:20.179641745 +0100 -@@ -33,6 +33,7 @@ - import java.util.Queue; - import java.util.Set; - -+import com.google.common.collect.Sets; - import org.apache.commons.lang.StringUtils; - import org.apache.commons.logging.Log; - import org.apache.commons.logging.LogFactory; -@@ -557,12 +558,27 @@ - */ - public static void doAuthorization(BaseSemanticAnalyzer sem, String command) - throws HiveException, AuthorizationException { -- HashSet<ReadEntity> inputs = sem.getInputs(); -- HashSet<WriteEntity> outputs = sem.getOutputs(); - SessionState ss = SessionState.get(); - HiveOperation op = ss.getHiveOperation(); - Hive db = sem.getDb(); - -+ Set<ReadEntity> additionalInputs = new HashSet<ReadEntity>(); -+ for (Entity e : sem.getInputs()) { -+ if (e.getType() == Entity.Type.PARTITION) { -+ additionalInputs.add(new ReadEntity(e.getTable())); -+ } -+ } -+ -+ Set<WriteEntity> additionalOutputs = new HashSet<WriteEntity>(); -+ for (Entity e : sem.getOutputs()) { -+ if (e.getType() == Entity.Type.PARTITION) { -+ additionalOutputs.add(new WriteEntity(e.getTable(), WriteEntity.WriteType.DDL_NO_LOCK)); -+ } -+ } -+ -+ Set<ReadEntity> inputs = Sets.union(sem.getInputs(), additionalInputs); -+ Set<WriteEntity> outputs = Sets.union(sem.getOutputs(), additionalOutputs); -+ - if (ss.isAuthorizationModeV2()) { - // get mapping of tables to columns used - ColumnAccessInfo colAccessInfo = sem.getColumnAccessInfo(); -@@ -759,8 +775,8 @@ - - } - -- private static void doAuthorizationV2(SessionState ss, HiveOperation op, HashSet<ReadEntity> inputs, -- HashSet<WriteEntity> outputs, String command, Map<String, List<String>> tab2cols, -+ private static void doAuthorizationV2(SessionState ss, HiveOperation op, Set<ReadEntity> inputs, -+ Set<WriteEntity> outputs, String command, Map<String, List<String>> tab2cols, - Map<String, List<String>> updateTab2Cols) throws HiveException { - - /* comment for reviewers -> updateTab2Cols needed to be separate from tab2cols because if I -@@ -780,7 +796,7 @@ - } - - private static List<HivePrivilegeObject> getHivePrivObjects( -- HashSet<? extends Entity> privObjects, Map<String, List<String>> tableName2Cols) { -+ Set<? extends Entity> privObjects, Map<String, List<String>> tableName2Cols) { - List<HivePrivilegeObject> hivePrivobjs = new ArrayList<HivePrivilegeObject>(); - if(privObjects == null){ - return hivePrivobjs; http://git-wip-us.apache.org/repos/asf/bigtop/blob/44aa7bb0/bigtop-packages/src/common/tez/do-component-build ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/common/tez/do-component-build b/bigtop-packages/src/common/tez/do-component-build index f49f090..2a5ca0c 100644 --- a/bigtop-packages/src/common/tez/do-component-build +++ b/bigtop-packages/src/common/tez/do-component-build @@ -20,7 +20,7 @@ set -xe BUILD_TEZ_OPTS="clean package \ -Dtar -Dhadoop.version=${HADOOP_VERSION} \ --Phadoop27 \ +-Phadoop28 \ -DskipTests" mvn ${BUILD_TEZ_OPTS} "$@" http://git-wip-us.apache.org/repos/asf/bigtop/blob/44aa7bb0/bigtop-packages/src/deb/hive/hive-jdbc.install ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/deb/hive/hive-jdbc.install b/bigtop-packages/src/deb/hive/hive-jdbc.install index ccd2a26..64b8e6d 100644 --- a/bigtop-packages/src/deb/hive/hive-jdbc.install +++ b/bigtop-packages/src/deb/hive/hive-jdbc.install @@ -1,9 +1 @@ -/usr/lib/hive/lib/commons-logging-*.jar -/usr/lib/hive/lib/hive-exec*.jar -/usr/lib/hive/lib/hive-jdbc*.jar -/usr/lib/hive/lib/hive-metastore*.jar -/usr/lib/hive/lib/hive-serde*.jar -/usr/lib/hive/lib/hive-service*.jar -/usr/lib/hive/lib/libfb303-*.jar -/usr/lib/hive/lib/libthrift-*.jar -/usr/lib/hive/lib/log4j-*.jar +/usr/lib/hive/jdbc/hive-jdbc*.jar http://git-wip-us.apache.org/repos/asf/bigtop/blob/44aa7bb0/bigtop-packages/src/deb/hive/rules ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/deb/hive/rules b/bigtop-packages/src/deb/hive/rules index dfb2537..6ef9b85 100755 --- a/bigtop-packages/src/deb/hive/rules +++ b/bigtop-packages/src/deb/hive/rules @@ -50,8 +50,9 @@ override_dh_auto_install: server2 metastore hcatalog-server webhcat-server --doc-dir=debian/tmp/usr/share/doc/${hive_pkg_name} # We need to get rid of jars that happen to be shipped in other CDH packages rm -f debian/tmp/usr/lib/hive/lib/hbase-*.jar debian/tmp/usr/lib/hive/lib/zookeeper-*.jar - ln -s /usr/lib/hbase/hbase-common.jar /usr/lib/hbase/hbase-client.jar debian/tmp/usr/lib/hive/lib - ln -s /usr/lib/zookeeper/zookeeper.jar debian/tmp/usr/lib/hive/lib + ln -s /usr/lib/hbase/hbase-common.jar /usr/lib/hbase/hbase-client.jar /usr/lib/hbase/hbase-hadoop-compat.jar /usr/lib/hbase/hbase-hadoop2-compat.jar debian/tmp/usr/lib/hive/lib + ln -s /usr/lib/hbase/hbase-prefix-tree.jar /usr/lib/hbase/hbase-procedure.jar /usr/lib/hbase/hbase-protocol.jar /usr/lib/hbase/hbase-server.jar debian/tmp/usr/lib/hive/lib/ + ln -s /usr/lib/zookeeper/zookeeper.jar debian/tmp/usr/lib/hive/lib # Workaround for BIGTOP-583 rm -f debian/tmp/usr/lib/hive/lib/slf4j-log4j12-*.jar bash debian/build-hive-install-file.sh >> debian/hive.install http://git-wip-us.apache.org/repos/asf/bigtop/blob/44aa7bb0/bigtop-packages/src/rpm/hive/SPECS/hive.spec ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/rpm/hive/SPECS/hive.spec b/bigtop-packages/src/rpm/hive/SPECS/hive.spec index 3766eab..cada51c 100644 --- a/bigtop-packages/src/rpm/hive/SPECS/hive.spec +++ b/bigtop-packages/src/rpm/hive/SPECS/hive.spec @@ -263,7 +263,8 @@ cp $RPM_SOURCE_DIR/hive-site.xml . # We need to get rid of jars that happen to be shipped in other Bigtop packages %__rm -f $RPM_BUILD_ROOT/%{usr_lib_hive}/lib/hbase-*.jar $RPM_BUILD_ROOT/%{usr_lib_hive}/lib/zookeeper-*.jar %__ln_s /usr/lib/zookeeper/zookeeper.jar $RPM_BUILD_ROOT/%{usr_lib_hive}/lib/ -%__ln_s /usr/lib/hbase/hbase-common.jar /usr/lib/hbase/hbase-client.jar $RPM_BUILD_ROOT/%{usr_lib_hive}/lib/ +%__ln_s /usr/lib/hbase/hbase-common.jar /usr/lib/hbase/hbase-client.jar /usr/lib/hbase/hbase-hadoop-compat.jar /usr/lib/hbase/hbase-hadoop2-compat.jar $RPM_BUILD_ROOT/%{usr_lib_hive}/lib/ +%__ln_s /usr/lib/hbase/hbase-prefix-tree.jar /usr/lib/hbase/hbase-procedure.jar /usr/lib/hbase/hbase-protocol.jar /usr/lib/hbase/hbase-server.jar $RPM_BUILD_ROOT/%{usr_lib_hive}/lib/ # Workaround for BIGTOP-583 %__rm -f $RPM_BUILD_ROOT/%{usr_lib_hive}/lib/slf4j-log4j12-*.jar @@ -334,18 +335,8 @@ fi %doc %{doc_hive} %{man_dir}/man1/hive.1.* %exclude %dir %{usr_lib_hive} -%exclude %dir %{usr_lib_hive}/lib -%exclude %{usr_lib_hive}/lib/hive-jdbc-*.jar -%exclude %{usr_lib_hive}/lib/hive-metastore-*.jar -%exclude %{usr_lib_hive}/lib/hive-serde-*.jar -%exclude %{usr_lib_hive}/lib/hive-exec-*.jar -%exclude %{usr_lib_hive}/lib/libthrift-*.jar -%exclude %{usr_lib_hive}/lib/hive-service-*.jar -%exclude %{usr_lib_hive}/lib/libfb303-*.jar -%exclude %{usr_lib_hive}/lib/log4j-*.jar -%exclude %{usr_lib_hive}/lib/commons-logging-*.jar -%exclude %{usr_lib_hive}/lib/hbase-*.jar -%exclude %{usr_lib_hive}/lib/hive-hbase-handler*.jar +%exclude %dir %{usr_lib_hive}/jdbc +%exclude %{usr_lib_hive}/jdbc/hive-jdbc-*.jar %files hbase %defattr(-,root,root,755) @@ -355,16 +346,8 @@ fi %files jdbc %defattr(-,root,root,755) %dir %{usr_lib_hive} -%dir %{usr_lib_hive}/lib -%{usr_lib_hive}/lib/hive-jdbc-*.jar -%{usr_lib_hive}/lib/hive-metastore-*.jar -%{usr_lib_hive}/lib/hive-serde-*.jar -%{usr_lib_hive}/lib/hive-exec-*.jar -%{usr_lib_hive}/lib/libthrift-*.jar -%{usr_lib_hive}/lib/hive-service-*.jar -%{usr_lib_hive}/lib/libfb303-*.jar -%{usr_lib_hive}/lib/log4j-*.jar -%{usr_lib_hive}/lib/commons-logging-*.jar +%dir %{usr_lib_hive}/jdbc +%{usr_lib_hive}/jdbc/hive-jdbc-*.jar %files hcatalog %defattr(-,root,root,755) http://git-wip-us.apache.org/repos/asf/bigtop/blob/44aa7bb0/bigtop-packages/src/rpm/tez/SPECS/tez.spec ---------------------------------------------------------------------- diff --git a/bigtop-packages/src/rpm/tez/SPECS/tez.spec b/bigtop-packages/src/rpm/tez/SPECS/tez.spec index eb0f46a..e2fdd76 100644 --- a/bigtop-packages/src/rpm/tez/SPECS/tez.spec +++ b/bigtop-packages/src/rpm/tez/SPECS/tez.spec @@ -100,7 +100,7 @@ sh %{SOURCE2} \ %__ln_s -f /usr/lib/hadoop/hadoop-annotations.jar $RPM_BUILD_ROOT/%{lib_tez}/hadoop-annotations.jar %__ln_s -f /usr/lib/hadoop/hadoop-auth.jar $RPM_BUILD_ROOT/%{lib_tez}/hadoop-auth.jar %__ln_s -f /usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-common.jar $RPM_BUILD_ROOT/%{lib_tez}/hadoop-mapreduce-client-common.jar -%__ln_s -f /usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-core.jar $RPM_BUILD_ROOT/%{lib_tez}/hhadoop-mapreduce-client-core.jar +%__ln_s -f /usr/lib/hadoop-mapreduce/hadoop-mapreduce-client-core.jar $RPM_BUILD_ROOT/%{lib_tez}/hadoop-mapreduce-client-core.jar %__ln_s -f /usr/lib/hadoop-yarn/hadoop-yarn-server-web-proxy.jar $RPM_BUILD_ROOT/%{lib_tez}/hadoop-yarn-server-web-proxy.jar %pre http://git-wip-us.apache.org/repos/asf/bigtop/blob/44aa7bb0/bigtop.bom ---------------------------------------------------------------------- diff --git a/bigtop.bom b/bigtop.bom index 9c72a11..81837d2 100644 --- a/bigtop.bom +++ b/bigtop.bom @@ -187,7 +187,7 @@ bigtop { 'hive' { name = 'hive' relNotes = 'Apache Hive' - version { base = '1.2.1'; pkg = base; release = 1 } + version { base = '2.3.2'; pkg = base; release = 1 } tarball { destination = "apache-${name}-${version.base}-src.tar.gz" source = destination } url { download_path = "/$name/$name-${version.base}/"
