活动介绍
file-type

Hive常用参数配置与调优指南

MD文件

下载需积分: 8 | 7KB | 更新于2024-08-05 | 187 浏览量 | 0 下载量 举报 收藏
download 立即下载
"hive常用参数设置和调优" 在Hive中,参数调整是提升系统性能的关键环节。以下是一些常见的Hive参数设置及其作用,这些参数可以帮助优化Hive的执行效率,减少数据处理时间。 ### Hive 常用参数设置 1. **`mapred.job.queue.name`**: 这个参数用于指定任务应该被提交到哪个队列。在YARN环境中,队列管理资源分配,可以根据需求设置不同的队列来优先处理不同级别的任务。 2. **`mapreduce.job.max.split.locations`**: 这个参数定义了最大的Split数量。增大这个值可能会增加并行度,但也要考虑集群的处理能力。 3. **`mapred.max.split.size`**: 每个Map的最大输入大小,它决定合并后的文件大小。适当调整此值可以平衡数据处理的并行度和网络传输的开销。 4. **`mapred.min.split.size.per.node`** 和 **`mapred.min.split.size.per.rack`**: 这两个参数用于设置在单个节点或同一交换机下的最小Split大小,它们影响数据在不同节点间的分布和数据本地性。 5. **`hive.input.format`**: 设置为 `CombineHiveInputFormat` 可在执行Map任务前合并小文件,减少Map任务数量,提高效率。 6. **`hive.exec.dynamic.partition`** 和 **`hive.exec.dynamic.partition.mode`**: 动态分区参数允许在运行时创建动态分区,`nonstrict` 模式允许所有分区列都是动态的,而在 `strict` 模式下,至少有一个分区列必须是静态的。 7. **`hive.exec.max.dynamic.partitions.pernode`** 和 **`hive.exec.max.dynamic.partitions`**: 这两个参数控制动态分区的创建数量,避免过多分区导致的问题。 8. **`hive.exec.max.created.files`**: 设置全局可创建的最大文件数,防止分区过多导致的文件管理问题。 ### Hive 查询优化参数 1. **`Hive.mapred.mode=strict`**: 严格模式可以避免一些潜在的性能问题和数据不一致,如强制要求对分区表的查询包含分区字段的过滤,对ORDER BY语句要求伴随LIMIT等。 ### Hive 常用调优参数 1. **Join操作优化**: 使用MapJoin代替ReduceJoin,当小表可以完全加载到内存时,MapJoin可以显著提高性能。 2. **数据倾斜优化**: 数据倾斜可能导致某些Reducer处理的数据远多于其他Reducer,可以通过哈希分桶或预聚合减少倾斜。 3. **Bucketing和Sorting**: 对表进行分桶和排序,可以提高JOIN和GROUP BY操作的效率。 4. **使用Tez或Spark作为执行引擎**: Tez和Spark比传统的MapReduce更高效,可以提高查询速度。 5. **避免全表扫描**: 尽可能使用WHERE子句进行过滤,减少不必要的数据处理。 6. **使用EXPLAIN命令**: 通过EXPLAIN查看Hive查询的执行计划,分析优化策略。 7. **分区策略**: 对大量数据的表使用分区,可以显著提高查询速度。 8. **压缩配置**: 开启压缩可以减少磁盘I/O和网络传输,提高效率。例如,设置 `hive.exec.compress.output=true` 和 `hive.exec.compress.intermediate=true`。 9. **优化SQL语句**: 避免笛卡尔积,合理使用索引,优化JOIN顺序,使用子查询替代UNION ALL等。 10. **调整执行器参数**: 如 `hive.exec.reducers.bytes-per-reducer` 和 `hive.exec.reducers.max` 可以控制Reducer数量,避免过多或过少的Reducer。 以上就是一些常用的Hive参数设置和调优策略,根据实际的业务场景和集群资源状况,合理调整这些参数可以显著提升Hive的性能和查询效率。

相关推荐

filetype

insert overwrite table case_data_sample select * from case_data_sample_tmp; 2025-06-18 16:37:06,500 INFO [main] conf.HiveConf: Using the default value passed in for log id: 531f6207-2ea7-471a-9eac-9ce1e6a79910 2025-06-18 16:37:06,500 INFO [main] session.SessionState: Updating thread name to 531f6207-2ea7-471a-9eac-9ce1e6a79910 main 2025-06-18 16:37:06,503 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Compiling command(queryId=root_20250618163706_29623b7c-a221-4e37-9b54-e05709d7f990): insert overwrite table case_data_sample select * from case_data_sample_tmp 2025-06-18 16:37:06,546 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Concurrency mode is disabled, not creating a lock manager 2025-06-18 16:37:06,547 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Starting Semantic Analysis 2025-06-18 16:37:06,559 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Completed phase 1 of Semantic Analysis 2025-06-18 16:37:06,560 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for source tables 2025-06-18 16:37:06,588 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for subqueries 2025-06-18 16:37:06,588 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for destination tables 2025-06-18 16:37:06,627 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Completed getting MetaData in Semantic Analysis 2025-06-18 16:37:08,746 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for source tables 2025-06-18 16:37:08,784 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for subqueries 2025-06-18 16:37:08,784 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for destination tables 2025-06-18 16:37:08,884 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] common.FileUtils: Creating directory if it doesn't exist: hdfs://master:8020/user/hive/warehouse/ad_traffic.db/case_data_sample/.hive-staging_hive_2025-06-18_16-37-06_538_2993870488593298816-1 2025-06-18 16:37:09,012 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Generate an operator pipeline to autogather column stats for table ad_traffic.case_data_sample in query insert overwrite table case_data_sample select * from case_data_sample_tmp 2025-06-18 16:37:09,069 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for source tables 2025-06-18 16:37:09,098 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for subqueries 2025-06-18 16:37:09,098 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Get metadata for destination tables 2025-06-18 16:37:09,155 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Context: New scratch dir is hdfs://master:8020/user/hive/tmp/root/531f6207-2ea7-471a-9eac-9ce1e6a79910/hive_2025-06-18_16-37-09_012_5684077500801740374-1 2025-06-18 16:37:09,221 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] common.FileUtils: Creating directory if it doesn't exist: hdfs://master:8020/user/hive/tmp/root/531f6207-2ea7-471a-9eac-9ce1e6a79910/hive_2025-06-18_16-37-09_012_5684077500801740374-1/-mr-10000/.hive-staging_hive_2025-06-18_16-37-09_012_5684077500801740374-1 2025-06-18 16:37:09,234 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: CBO Succeeded; optimized logical plan. 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for FS(2) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for FS(9) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for SEL(8) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for GBY(7) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for RS(6) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for GBY(5) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for SEL(4) 2025-06-18 16:37:09,329 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for SEL(1) 2025-06-18 16:37:09,330 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ppd.OpProcFactory: Processing for TS(0) 2025-06-18 16:37:09,385 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] optimizer.ColumnPrunerProcFactory: RS 6 oldColExprMap: {VALUE._col20=Column[_col20], VALUE._col10=Column[_col10], VALUE._col21=Column[_col21], VALUE._col11=Column[_col11], VALUE._col12=Column[_col12], VALUE._col2=Column[_col2], VALUE._col3=Column[_col3], VALUE._col4=Column[_col4], VALUE._col5=Column[_col5], VALUE._col0=Column[_col0], VALUE._col1=Column[_col1], VALUE._col13=Column[_col13], VALUE._col14=Column[_col14], VALUE._col15=Column[_col15], VALUE._col16=Column[_col16], VALUE._col6=Column[_col6], VALUE._col17=Column[_col17], VALUE._col7=Column[_col7], VALUE._col18=Column[_col18], VALUE._col8=Column[_col8], VALUE._col19=Column[_col19], VALUE._col9=Column[_col9]} 2025-06-18 16:37:09,386 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] optimizer.ColumnPrunerProcFactory: RS 6 newColExprMap: {VALUE._col20=Column[_col20], VALUE._col10=Column[_col10], VALUE._col21=Column[_col21], VALUE._col11=Column[_col11], VALUE._col12=Column[_col12], VALUE._col2=Column[_col2], VALUE._col3=Column[_col3], VALUE._col4=Column[_col4], VALUE._col5=Column[_col5], VALUE._col0=Column[_col0], VALUE._col1=Column[_col1], VALUE._col13=Column[_col13], VALUE._col14=Column[_col14], VALUE._col15=Column[_col15], VALUE._col16=Column[_col16], VALUE._col6=Column[_col6], VALUE._col17=Column[_col17], VALUE._col7=Column[_col7], VALUE._col18=Column[_col18], VALUE._col8=Column[_col8], VALUE._col19=Column[_col19], VALUE._col9=Column[_col9]} 2025-06-18 16:37:09,500 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SetSparkReducerParallelism: Number of reducers for sink RS[6] was already determined to be: 1 2025-06-18 16:37:09,646 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Examining input format to see if vectorization is enabled. 2025-06-18 16:37:09,655 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Vectorization is enabled for input format(s) [org.apache.hadoop.mapred.TextInputFormat] 2025-06-18 16:37:09,655 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Validating and vectorizing MapWork... (vectorizedVertexNum 0) 2025-06-18 16:37:09,706 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map vectorization enabled: true 2025-06-18 16:37:09,706 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map vectorized: false 2025-06-18 16:37:09,707 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map notVectorizedReason: Aggregation Function expression for GROUPBY operator: UDF compute_stats not supported 2025-06-18 16:37:09,707 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map vectorizedVertexNum: 0 2025-06-18 16:37:09,707 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map enabledConditionsMet: [hive.vectorized.use.vector.serde.deserialize IS true] 2025-06-18 16:37:09,707 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Map inputFileFormatClassNameSet: [org.apache.hadoop.mapred.TextInputFormat] 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Validating and vectorizing ReduceWork... (vectorizedVertexNum 1) 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reduce vectorization enabled: true 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reduce vectorized: false 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reduce notVectorizedReason: Aggregation Function expression for GROUPBY operator: UDF compute_stats not supported 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reduce vectorizedVertexNum: 1 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reducer hive.vectorized.execution.reduce.enabled: true 2025-06-18 16:37:09,753 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] physical.Vectorizer: Reducer engine: spark 2025-06-18 16:37:09,784 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] parse.CalcitePlanner: Completed plan generation 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Semantic Analysis Completed (retrial = false) 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: , FieldSchema(name:case_data_sample_tmp.timestamps, type:int, comment:null), FieldSchema(name:case_data_sample_tmp.camp, type:int, comment:null), FieldSchema(name:case_data_sample_tmp.creativeid, type:int, comment:null), FieldSchema(name:case_data_sample_tmp.mobile_os, type:int, comment:null), FieldSchema(name:case_data_sample_tmp.mobile_type, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.app_key_md5, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.app_name_md5, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.placementid, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.useragent, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.mediaid, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.os_type, type:string, comment:null), FieldSchema(name:case_data_sample_tmp.born_time, type:int, comment:null), FieldSchema(name:case_data_sample_tmp.label, type:int, comment:null)], properties:null) 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Completed compiling command(queryId=root_20250618163706_29623b7c-a221-4e37-9b54-e05709d7f990); Time taken: 3.282 seconds 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] reexec.ReExecDriver: Execution #1 of query 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Concurrency mode is disabled, not creating a lock manager 2025-06-18 16:37:09,785 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Executing command(queryId=root_20250618163706_29623b7c-a221-4e37-9b54-e05709d7f990): insert overwrite table case_data_sample select * from case_data_sample_tmp 2025-06-18 16:37:09,786 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Query ID = root_20250618163706_29623b7c-a221-4e37-9b54-e05709d7f990 2025-06-18 16:37:09,786 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Total jobs = 1 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Launching Job 1 out of 1 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Starting task [Stage-1:MAPRED] in serial mode 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: In order to change the average load for a reducer (in bytes): 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: set hive.exec.reducers.bytes.per.reducer=<number> 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: In order to limit the maximum number of reducers: 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: set hive.exec.reducers.max=<number> 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: In order to set a constant number of reducers: 2025-06-18 16:37:09,813 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: set mapreduce.job.reduces=<number> 2025-06-18 16:37:09,834 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] session.SparkSessionManagerImpl: Setting up the session manager. 2025-06-18 16:37:10,327 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] session.SparkSession: Trying to open Spark session 4201eecc-977e-4e69-89ec-50403379b3d2 Failed to execute spark task, with exception 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create Spark client for Spark session 4201eecc-977e-4e69-89ec-50403379b3d2)' 2025-06-18 16:37:10,372 ERROR [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: .lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.util.RunJar.run(RunJar.java:318) at org.apache.hadoop.util.RunJar.main(RunJar.java:232) Caused by: java.lang.NoClassDefFoundError: org/apache/spark/SparkConf at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.generateSparkConf(HiveSparkClientFactory.java:263) at org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:98) at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:76) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:87) ... 24 more Caused by: java.lang.ClassNotFoundException: org.apache.spark.SparkConf at java.net.URLClassLoader.findClass(URLClassLoader.java:387) at java.lang.ClassLoader.loadClass(ClassLoader.java:418) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:359) at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ... 28 more 2025-06-18 16:37:10,378 ERROR [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] spark.SparkTask: Failed to execute spark task, with exception 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create Spark client for Spark session 4201eecc-977e-4e69-89ec-50403379b3d2)' org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create Spark client for Spark session 4201eecc-977e-4e69-89ec-50403379b3d2 at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.getHiveException(SparkSessionImpl.java:221) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:92) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.java:115) at org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUtilities.java:136) at org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:115) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821) at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.util.RunJar.run(RunJar.java:318) at org.apache.hadoop.util.RunJar.main(RunJar.java:232) Caused by: java.lang.NoClassDefFoundError: org/apache/spark/SparkConf at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.generateSparkConf(HiveSparkClientFactory.java:263) at org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:98) at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:76) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:87) ... 24 more Caused by: java.lang.ClassNotFoundException: org.apache.spark.SparkConf at java.net.URLClassLoader.findClass(URLClassLoader.java:387) at java.lang.ClassLoader.loadClass(ClassLoader.java:418) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:359) at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ... 28 more 2025-06-18 16:37:10,391 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] reexec.ReOptimizePlugin: ReOptimization: retryPossible: false FAILED: Execution Error, return code 30041 from org.apache.hadoop.hive.ql.exec.spark.SparkTask. Failed to create Spark client for Spark session 4201eecc-977e-4e69-89ec-50403379b3d2 2025-06-18 16:37:10,391 ERROR [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: FAILED: Execution Error, return code 30041 from org.apache.hadoop.hive.ql.exec.spark.SparkTask. Failed to create Spark client for Spark session 4201eecc-977e-4e69-89ec-50403379b3d2 2025-06-18 16:37:10,392 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Completed executing command(queryId=root_20250618163706_29623b7c-a221-4e37-9b54-e05709d7f990); Time taken: 0.607 seconds 2025-06-18 16:37:10,392 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] ql.Driver: Concurrency mode is disabled, not creating a lock manager 2025-06-18 16:37:10,433 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] conf.HiveConf: Using the default value passed in for log id: 531f6207-2ea7-471a-9eac-9ce1e6a79910 2025-06-18 16:37:10,433 INFO [531f6207-2ea7-471a-9eac-9ce1e6a79910 main] session.SessionState: Resetting thread name to main

霸气多宝鱼
  • 粉丝: 9
上传资源 快速赚钱