spark on yarn 读取hdfs拒绝连接
代码:
package dsy.read_hdfs
import org.apache.spark.SparkConf
import org.apache.spark.sql.{DataFrame, SparkSession}
object read_hdfs {
def main(args: Array[String]): Unit = {
System.setProperty("HADOOP_USER_NAME", "root");
System.setProperty("user.name", "root");
val value = this.getClass.getClassLoader.loadClass("org.apache.spark.scheduler.cluster.YarnClusterManager")
val spark: SparkSession = {
val conf: SparkConf = new SparkConf()
// 设置yarn-client模式提交
.setMaster("yarn")
//App名字
.set("spark.app.name", this.getClass.getSimpleName.stripSuffix("$"))
// 设置resourcemanager的ip
.set("yarn.resourcemanager.hostname", "dsy")
// 设置executor的个数
.set("spark.executor.instance", "2")
// 设置e