Could not find or load main class
Closed this issue · 6 comments
Code of Conduct
- I agree to follow this project's Code of Conduct
Search before asking
- I have searched in the issues and found no similar issues.
Describe the bug
./beeline -u "jdbc:hive2://hadoop01:2181,hadoop02:2181,hadoop03:2181/default;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=kyuubi" -n root1 --hiveconf kyuubi.engine.type=HIVE_SQL --hiveconf kyuubi.engine.hive.deploy.mode=local
Launching engine Manual(shell): it works well
/usr/java/jdk1.8.0_351-amd64/bin/java
-Xmx1g
-cp /usr/lib/kyuubi/externals/engines/hive/kyuubi-hive-sql-engine_2.12-1.9.1.jar:/usr/lib/hive/conf:/etc/hadoop/conf:/usr/lib/hadoop/etc/hadoop:/usr/lib/hive/lib/:/usr/lib/kyuubi/jars/commons-collections-3.2.2.jar:/usr/lib/kyuubi/jars/hadoop-client-runtime-3.3.6.jar:/usr/lib/kyuubi/jars/hadoop-client-api-3.3.6.jar: org.apache.kyuubi.engine.hive.HiveSQLEngine
--conf kyuubi.session.user=root1
--conf kyuubi.engine.id=b1313aae-c609-4698-bc70-6581168961f8
--conf hive.engine.name=kyuubi_USER_HIVE_SQL_root1_default_b1313aae-c609-4698-bc70-6581168961f8
--conf hive.server2.thrift.resultset.default.fetch.size=1000
--conf kyuubi.backend.engine.exec.pool.keepalive.time=PT1M
--conf kyuubi.backend.engine.exec.pool.shutdown.timeout=PT20S
--conf kyuubi.backend.engine.exec.pool.size=100
--conf kyuubi.backend.engine.exec.pool.wait.queue.size=100
--conf kyuubi.backend.server.exec.pool.keepalive.time=PT1M
--conf kyuubi.backend.server.exec.pool.shutdown.timeout=PT20S
--conf kyuubi.backend.server.exec.pool.size=100
--conf kyuubi.backend.server.exec.pool.wait.queue.size=100
--conf kyuubi.batch.application.check.interval=PT10S
--conf kyuubi.batch.application.starvation.timeout=PT3M
--conf kyuubi.batch.session.idle.timeout=PT6H
--conf kyuubi.client.ipAddress=192.168.1.110
--conf kyuubi.client.version=1.9.1
--conf kyuubi.engine.event.json.log.path=/var/lib/kyuubi/engine/event
--conf kyuubi.engine.flink.application.jars=
--conf kyuubi.engine.flink.extra.classpath=
--conf kyuubi.engine.flink.java.options=
--conf kyuubi.engine.flink.memory=1g
--conf kyuubi.engine.hive.deploy.mode=local
--conf kyuubi.engine.hive.event.loggers=JSON
--conf kyuubi.engine.hive.extra.classpath=
--conf kyuubi.engine.hive.java.options=
--conf kyuubi.engine.hive.memory=1g
--conf kyuubi.engine.pool.name=kyuubi-engine-pool
--conf kyuubi.engine.pool.selectPolicy=RANDOM
--conf kyuubi.engine.pool.size=-1
--conf kyuubi.engine.session.initialize.sql=
--conf kyuubi.engine.share.level=USER
--conf kyuubi.engine.spark.event.loggers=SPARK
--conf kyuubi.engine.submit.time=1721366841647
--conf kyuubi.engine.submit.timeout=PT30S
--conf kyuubi.engine.type=HIVE_SQL
--conf kyuubi.engine.ui.retainedSessions=200
--conf kyuubi.engine.ui.retainedStatements=200
--conf kyuubi.engine.ui.stop.enabled=true
--conf kyuubi.engine.yarn.cores=1
--conf kyuubi.engine.yarn.java.options=
--conf kyuubi.engine.yarn.memory=1024
--conf kyuubi.engine.yarn.queue=default
--conf kyuubi.engine.yarn.submit.timeout=PT1M
--conf kyuubi.event.async.pool.keepalive.time=PT1M
--conf kyuubi.event.async.pool.size=8
--conf kyuubi.event.async.pool.wait.queue.size=100
--conf kyuubi.frontend.connection.url.use.hostname=true
--conf kyuubi.frontend.max.message.size=104857600
--conf kyuubi.frontend.max.worker.threads=999
--conf kyuubi.frontend.min.worker.threads=9
--conf kyuubi.frontend.protocols=THRIFT_BINARY,REST
--conf kyuubi.frontend.proxy.http.client.ip.header=X-Real-IP
--conf kyuubi.frontend.rest.jetty.stopTimeout=PT10S
--conf kyuubi.frontend.rest.max.worker.threads=999
--conf kyuubi.frontend.thrift.binary.ssl.disallowed.protocols=SSLv2,SSLv3
--conf kyuubi.frontend.thrift.binary.ssl.enabled=false
--conf kyuubi.frontend.thrift.max.message.size=104857600
--conf kyuubi.frontend.thrift.max.worker.threads=999
--conf kyuubi.frontend.thrift.min.worker.threads=9
--conf kyuubi.frontend.thrift.worker.keepalive.time=PT1M
--conf kyuubi.ha.addresses=hadoop01:2181,hadoop02:2181,hadoop03:2181
--conf kyuubi.ha.client.class=org.apache.kyuubi.ha.client.zookeeper.ZookeeperDiscoveryClient
--conf kyuubi.ha.engine.ref.id=b1313aae-c609-4698-bc70-6581168961f8
--conf kyuubi.ha.namespace=/kyuubi_1.9.1_USER_HIVE_SQL/root1/default
--conf kyuubi.ha.zookeeper.acl.enabled=false
--conf kyuubi.ha.zookeeper.auth.type=NONE
--conf kyuubi.ha.zookeeper.connection.base.retry.wait=1000
--conf kyuubi.ha.zookeeper.connection.max.retries=3
--conf kyuubi.ha.zookeeper.connection.max.retry.wait=30000
--conf kyuubi.ha.zookeeper.connection.retry.policy=EXPONENTIAL_BACKOFF
--conf kyuubi.ha.zookeeper.connection.timeout=15000
--conf kyuubi.ha.zookeeper.engine.auth.type=NONE
--conf kyuubi.ha.zookeeper.node.creation.timeout=PT2M
--conf kyuubi.ha.zookeeper.session.timeout=60000
--conf kyuubi.metadata.cleaner.enabled=true
--conf kyuubi.metadata.cleaner.interval=PT30M
--conf kyuubi.metadata.max.age=PT128H
--conf kyuubi.metadata.recovery.threads=10
--conf kyuubi.metadata.request.async.retry.enabled=true
--conf kyuubi.metadata.request.async.retry.queue.size=65536
--conf kyuubi.metadata.request.async.retry.threads=10
--conf kyuubi.metadata.request.retry.interval=PT5S
--conf kyuubi.metadata.store.class=org.apache.kyuubi.server.metadata.jdbc.JDBCMetadataStore
--conf kyuubi.metrics.console.interval=PT20S
--conf kyuubi.metrics.enabled=false
--conf kyuubi.metrics.reporters=
--conf kyuubi.operation.query.timeout=3600000
--conf kyuubi.operation.scheduler.pool=fair
--conf kyuubi.server.info.provider=ENGINE
--conf kyuubi.server.ipAddress=192.168.1.110
--conf kyuubi.session.check.interval=PT5M
--conf kyuubi.session.close.on.disconnect=true
--conf kyuubi.session.connection.url=hadoop01:10009
--conf kyuubi.session.engine.alive.timeout=PT2M
--conf kyuubi.session.engine.check.interval=PT1M
--conf kyuubi.session.engine.idle.timeout=PT30M
--conf kyuubi.session.engine.initialize.timeout=PT5M
--conf kyuubi.session.engine.launch.async=true
--conf kyuubi.session.engine.log.timeout=PT24H
--conf kyuubi.session.idle.timeout=PT6H
--conf kyuubi.session.real.user=root1
--conf spark.cleaner.periodicGC.interval=5min
--conf spark.driver.cores=1
--conf spark.driver.maxResultSize=1g
--conf spark.dynamicAllocation.cachedExecutorIdleTimeout=30min
--conf spark.dynamicAllocation.enabled=true
--conf spark.dynamicAllocation.executorAllocationRatio=0.5
--conf spark.dynamicAllocation.executorIdleTimeout=60s
--conf spark.dynamicAllocation.initialExecutors=2
--conf spark.dynamicAllocation.maxExecutors=25
--conf spark.dynamicAllocation.minExecutors=2
--conf spark.dynamicAllocation.schedulerBacklogTimeout=1s
--conf spark.dynamicAllocation.shuffleTracking.enabled=false
--conf spark.dynamicAllocation.shuffleTracking.timeout=30min
--conf spark.dynamicAllocation.sustainedSchedulerBacklogTimeout=1s
--conf spark.hadoop.cacheConf=false
--conf spark.io.compression.lz4.blockSize=128kb
--conf spark.master=yarn
--conf spark.scheduler.allocation.file=hdfs:///user/spark/conf/kyuubi-fairscheduler.xml
--conf spark.scheduler.mode=FAIR
--conf spark.shuffle.file.buffer=1m
--conf spark.shuffle.io.backLog=8192
--conf spark.shuffle.push.enabled=true
--conf spark.shuffle.service.enabled=true
--conf spark.shuffle.service.index.cache.size=100m
--conf spark.shuffle.service.port=17337
--conf spark.shuffle.service.removeShuffle=false
--conf spark.sql.adaptive.advisoryPartitionSizeInBytes=128M
--conf spark.sql.adaptive.autoBroadcastJoinThreshold=10MB
--conf spark.sql.adaptive.coalescePartitions.enabled=true
--conf spark.sql.adaptive.coalescePartitions.initialPartitionNum=8192
--conf spark.sql.adaptive.coalescePartitions.minPartitionSize=1MB
--conf spark.sql.adaptive.coalescePartitions.parallelismFirst=true
--conf spark.sql.adaptive.enabled=true
--conf spark.sql.adaptive.forceOptimizeSkewedJoin=false
--conf spark.sql.adaptive.localShuffleReader.enabled=true
--conf spark.sql.adaptive.optimizeSkewsInRebalancePartitions.enabled=true
--conf spark.sql.adaptive.rebalancePartitionsSmallPartitionFactor=0.2
--conf spark.sql.adaptive.skewJoin.enabled=true
--conf spark.sql.adaptive.skewJoin.skewedPartitionFactor=5
--conf spark.sql.adaptive.skewJoin.skewedPartitionThresholdInBytes=256MB
--conf spark.sql.autoBroadcastJoinThreshold=10MB
--conf spark.sql.hive.convertMetastoreOrc=true
--conf spark.sql.hive.metastore.jars=/usr/lib/hive/lib/
--conf spark.sql.hive.metastore.version=3.1.3
--conf spark.sql.orc.filterPushdown=true
--conf spark.sql.statistics.fallBackToHdfs=true
--conf spark.submit.deployMode=client
kyuubi.engine.hive.deploy.mode=yarn it works well
./beeline -u "jdbc:hive2://hadoop01:2181,hadoop02:2181,hadoop03:2181/default;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=kyuubi" -n root1 --hiveconf kyuubi.engine.type=HIVE_SQL --hiveconf kyuubi.engine.hive.deploy.mode=yarn
Affects Version(s)
1.9.1
Kyuubi Server Log Output
2024-07-19 13:27:21.651 INFO KyuubiSessionManager-exec-pool: Thread-630 org.apache.kyuubi.engine.EngineRef: Launching engine:
/usr/java/jdk1.8.0_351-amd64/bin/java \
-Xmx1g \
-cp /usr/lib/kyuubi/externals/engines/hive/kyuubi-hive-sql-engine_2.12-1.9.1.jar:/usr/lib/hive/conf:/etc/hadoop/conf:/usr/lib/hadoop/etc/hadoop:/usr/lib/hive/lib/*:/usr/lib/kyuubi/jars/commons-collections-3.2.2.jar:/usr/lib/kyuubi/jars/hadoop-client-runtime-3.3.6.jar:/usr/lib/kyuubi/jars/hadoop-client-api-3.3.6.jar: org.apache.kyuubi.engine.hive.HiveSQLEngine \
--conf kyuubi.session.user=root1 \
--conf kyuubi.engine.id=b1313aae-c609-4698-bc70-6581168961f8 \
--conf hive.engine.name=kyuubi_USER_HIVE_SQL_root1_default_b1313aae-c609-4698-bc70-6581168961f8 \
--conf hive.server2.thrift.resultset.default.fetch.size=1000 \
--conf kyuubi.backend.engine.exec.pool.keepalive.time=PT1M \
--conf kyuubi.backend.engine.exec.pool.shutdown.timeout=PT20S \
--conf kyuubi.backend.engine.exec.pool.size=100 \
--conf kyuubi.backend.engine.exec.pool.wait.queue.size=100 \
--conf kyuubi.backend.server.exec.pool.keepalive.time=PT1M \
--conf kyuubi.backend.server.exec.pool.shutdown.timeout=PT20S \
--conf kyuubi.backend.server.exec.pool.size=100 \
--conf kyuubi.backend.server.exec.pool.wait.queue.size=100 \
--conf kyuubi.batch.application.check.interval=PT10S \
--conf kyuubi.batch.application.starvation.timeout=PT3M \
--conf kyuubi.batch.session.idle.timeout=PT6H \
--conf kyuubi.client.ipAddress=192.168.1.110 \
--conf kyuubi.client.version=1.9.1 \
--conf kyuubi.engine.event.json.log.path=/var/lib/kyuubi/engine/event \
--conf kyuubi.engine.flink.application.jars= \
--conf kyuubi.engine.flink.extra.classpath= \
--conf kyuubi.engine.flink.java.options= \
--conf kyuubi.engine.flink.memory=1g \
--conf kyuubi.engine.hive.deploy.mode=local \
--conf kyuubi.engine.hive.event.loggers=JSON \
--conf kyuubi.engine.hive.extra.classpath= \
--conf kyuubi.engine.hive.java.options= \
--conf kyuubi.engine.hive.memory=1g \
--conf kyuubi.engine.pool.name=kyuubi-engine-pool \
--conf kyuubi.engine.pool.selectPolicy=RANDOM \
--conf kyuubi.engine.pool.size=-1 \
--conf kyuubi.engine.session.initialize.sql= \
--conf kyuubi.engine.share.level=USER \
--conf kyuubi.engine.spark.event.loggers=SPARK \
--conf kyuubi.engine.submit.time=1721366841647 \
--conf kyuubi.engine.submit.timeout=PT30S \
--conf kyuubi.engine.type=HIVE_SQL \
--conf kyuubi.engine.ui.retainedSessions=200 \
--conf kyuubi.engine.ui.retainedStatements=200 \
--conf kyuubi.engine.ui.stop.enabled=true \
--conf kyuubi.engine.yarn.cores=1 \
--conf kyuubi.engine.yarn.java.options= \
--conf kyuubi.engine.yarn.memory=1024 \
--conf kyuubi.engine.yarn.queue=default \
--conf kyuubi.engine.yarn.submit.timeout=PT1M \
--conf kyuubi.event.async.pool.keepalive.time=PT1M \
--conf kyuubi.event.async.pool.size=8 \
--conf kyuubi.event.async.pool.wait.queue.size=100 \
--conf kyuubi.frontend.connection.url.use.hostname=true \
--conf kyuubi.frontend.max.message.size=104857600 \
--conf kyuubi.frontend.max.worker.threads=999 \
--conf kyuubi.frontend.min.worker.threads=9 \
--conf kyuubi.frontend.protocols=THRIFT_BINARY,REST \
--conf kyuubi.frontend.proxy.http.client.ip.header=X-Real-IP \
--conf kyuubi.frontend.rest.jetty.stopTimeout=PT10S \
--conf kyuubi.frontend.rest.max.worker.threads=999 \
--conf kyuubi.frontend.thrift.binary.ssl.disallowed.protocols=SSLv2,SSLv3 \
--conf kyuubi.frontend.thrift.binary.ssl.enabled=false \
--conf kyuubi.frontend.thrift.max.message.size=104857600 \
--conf kyuubi.frontend.thrift.max.worker.threads=999 \
--conf kyuubi.frontend.thrift.min.worker.threads=9 \
--conf kyuubi.frontend.thrift.worker.keepalive.time=PT1M \
--conf kyuubi.ha.addresses=hadoop01:2181,hadoop02:2181,hadoop03:2181 \
--conf kyuubi.ha.client.class=org.apache.kyuubi.ha.client.zookeeper.ZookeeperDiscoveryClient \
--conf kyuubi.ha.engine.ref.id=b1313aae-c609-4698-bc70-6581168961f8 \
--conf kyuubi.ha.namespace=/kyuubi_1.9.1_USER_HIVE_SQL/root1/default \
--conf kyuubi.ha.zookeeper.acl.enabled=false \
--conf kyuubi.ha.zookeeper.auth.type=NONE \
--conf kyuubi.ha.zookeeper.connection.base.retry.wait=1000 \
--conf kyuubi.ha.zookeeper.connection.max.retries=3 \
--conf kyuubi.ha.zookeeper.connection.max.retry.wait=30000 \
--conf kyuubi.ha.zookeeper.connection.retry.policy=EXPONENTIAL_BACKOFF \
--conf kyuubi.ha.zookeeper.connection.timeout=15000 \
--conf kyuubi.ha.zookeeper.engine.auth.type=NONE \
--conf kyuubi.ha.zookeeper.node.creation.timeout=PT2M \
--conf kyuubi.ha.zookeeper.session.timeout=60000 \
--conf kyuubi.metadata.cleaner.enabled=true \
--conf kyuubi.metadata.cleaner.interval=PT30M \
--conf kyuubi.metadata.max.age=PT128H \
--conf kyuubi.metadata.recovery.threads=10 \
--conf kyuubi.metadata.request.async.retry.enabled=true \
--conf kyuubi.metadata.request.async.retry.queue.size=65536 \
--conf kyuubi.metadata.request.async.retry.threads=10 \
--conf kyuubi.metadata.request.retry.interval=PT5S \
--conf kyuubi.metadata.store.class=org.apache.kyuubi.server.metadata.jdbc.JDBCMetadataStore \
--conf kyuubi.metrics.console.interval=PT20S \
--conf kyuubi.metrics.enabled=false \
--conf kyuubi.metrics.reporters= \
--conf kyuubi.operation.query.timeout=3600000 \
--conf kyuubi.operation.scheduler.pool=fair \
--conf kyuubi.server.info.provider=ENGINE \
--conf kyuubi.server.ipAddress=192.168.1.110 \
--conf kyuubi.session.check.interval=PT5M \
--conf kyuubi.session.close.on.disconnect=true \
--conf kyuubi.session.connection.url=hadoop01:10009 \
--conf kyuubi.session.engine.alive.timeout=PT2M \
--conf kyuubi.session.engine.check.interval=PT1M \
--conf kyuubi.session.engine.idle.timeout=PT30M \
--conf kyuubi.session.engine.initialize.timeout=PT5M \
--conf kyuubi.session.engine.launch.async=true \
--conf kyuubi.session.engine.log.timeout=PT24H \
--conf kyuubi.session.idle.timeout=PT6H \
--conf kyuubi.session.real.user=root1 \
--conf spark.cleaner.periodicGC.interval=5min \
--conf spark.driver.cores=1 \
--conf spark.driver.maxResultSize=1g \
--conf spark.dynamicAllocation.cachedExecutorIdleTimeout=30min \
--conf spark.dynamicAllocation.enabled=true \
--conf spark.dynamicAllocation.executorAllocationRatio=0.5 \
--conf spark.dynamicAllocation.executorIdleTimeout=60s \
--conf spark.dynamicAllocation.initialExecutors=2 \
--conf spark.dynamicAllocation.maxExecutors=25 \
--conf spark.dynamicAllocation.minExecutors=2 \
--conf spark.dynamicAllocation.schedulerBacklogTimeout=1s \
--conf spark.dynamicAllocation.shuffleTracking.enabled=false \
--conf spark.dynamicAllocation.shuffleTracking.timeout=30min \
--conf spark.dynamicAllocation.sustainedSchedulerBacklogTimeout=1s \
--conf spark.hadoop.cacheConf=false \
--conf spark.io.compression.lz4.blockSize=128kb \
--conf spark.master=yarn \
--conf spark.scheduler.allocation.file=hdfs:///user/spark/conf/kyuubi-fairscheduler.xml \
--conf spark.scheduler.mode=FAIR \
--conf spark.shuffle.file.buffer=1m \
--conf spark.shuffle.io.backLog=8192 \
--conf spark.shuffle.push.enabled=true \
--conf spark.shuffle.service.enabled=true \
--conf spark.shuffle.service.index.cache.size=100m \
--conf spark.shuffle.service.port=17337 \
--conf spark.shuffle.service.removeShuffle=false \
--conf spark.sql.adaptive.advisoryPartitionSizeInBytes=128M \
--conf spark.sql.adaptive.autoBroadcastJoinThreshold=10MB \
--conf spark.sql.adaptive.coalescePartitions.enabled=true \
--conf spark.sql.adaptive.coalescePartitions.initialPartitionNum=8192 \
--conf spark.sql.adaptive.coalescePartitions.minPartitionSize=1MB \
--conf spark.sql.adaptive.coalescePartitions.parallelismFirst=true \
--conf spark.sql.adaptive.enabled=true \
--conf spark.sql.adaptive.forceOptimizeSkewedJoin=false \
--conf spark.sql.adaptive.localShuffleReader.enabled=true \
--conf spark.sql.adaptive.optimizeSkewsInRebalancePartitions.enabled=true \
--conf spark.sql.adaptive.rebalancePartitionsSmallPartitionFactor=0.2 \
--conf spark.sql.adaptive.skewJoin.enabled=true \
--conf spark.sql.adaptive.skewJoin.skewedPartitionFactor=5 \
--conf spark.sql.adaptive.skewJoin.skewedPartitionThresholdInBytes=256MB \
--conf spark.sql.autoBroadcastJoinThreshold=10MB \
--conf spark.sql.hive.convertMetastoreOrc=true \
--conf spark.sql.hive.metastore.jars=/usr/lib/hive/lib/* \
--conf spark.sql.hive.metastore.version=3.1.3 \
--conf spark.sql.orc.filterPushdown=true \
--conf spark.sql.statistics.fallBackToHdfs=true \
--conf spark.submit.deployMode=client
2024-07-19 13:27:21.652 INFO KyuubiSessionManager-exec-pool: Thread-630 org.apache.kyuubi.engine.ProcBuilder: Logging to /var/lib/kyuubi/root1/kyuubi-hive-sql-engine.log.2
2024-07-19 13:27:22.763 INFO Curator-Framework-0 org.apache.kyuubi.shaded.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting
2024-07-19 13:27:22.872 INFO KyuubiSessionManager-exec-pool: Thread-630-EventThread org.apache.kyuubi.shaded.zookeeper.ClientCnxn: EventThread shut down for session: 0x200000139453541
2024-07-19 13:27:22.872 INFO KyuubiSessionManager-exec-pool: Thread-630 org.apache.kyuubi.shaded.zookeeper.ZooKeeper: Session: 0x200000139453541 closed
2024-07-19 13:27:22.876 INFO KyuubiSessionManager-exec-pool: Thread-630 org.apache.kyuubi.operation.LaunchEngine: Processing root1's query[d093f68e-7596-4c81-a0e0-281b1d9c3c71]: RUNNING_STATE -> ERROR_STATE, time taken: 1.344 seconds
2024-07-19 13:27:23.008 INFO KyuubiTBinaryFrontendHandler-Pool: Thread-571 org.apache.kyuubi.server.KyuubiTBinaryFrontendService: Received request of closing SessionHandle [b1313aae-c609-4698-bc70-6581168961f8]
2024-07-19 13:27:23.009 INFO KyuubiTBinaryFrontendHandler-Pool: Thread-571 org.apache.kyuubi.session.KyuubiSessionManager: root1's KyuubiSessionImpl with SessionHandle [b1313aae-c609-4698-bc70-6581168961f8] is closed, current opening sessions 2
2024-07-19 13:27:23.009 INFO KyuubiTBinaryFrontendHandler-Pool: Thread-571 org.apache.kyuubi.operation.LaunchEngine: Processing root1's query[d093f68e-7596-4c81-a0e0-281b1d9c3c71]: ERROR_STATE -> CLOSED_STATE, time taken: 1.477 seconds
2024-07-19 13:27:23.013 INFO KyuubiTBinaryFrontendHandler-Pool: Thread-571 org.apache.kyuubi.server.KyuubiTBinaryFrontendService: Finished closing SessionHandle [b1313aae-c609-4698-bc70-6581168961f8]
Kyuubi Engine Log Output
2024-07-19 13:27:21.651 INFO KyuubiSessionManager-exec-pool: Thread-630 org.apache.kyuubi.engine.EngineRef: Launching engine:
/usr/java/jdk1.8.0_351-amd64/bin/java \
-Xmx1g \
-cp /usr/lib/kyuubi/externals/engines/hive/kyuubi-hive-sql-engine_2.12-1.9.1.jar:/usr/lib/hive/conf:/etc/hadoop/conf:/usr/lib/hadoop/etc/hadoop:/usr/lib/hive/lib/*:/usr/lib/kyuubi/jars/commons-collections-3.2.2.jar:/usr/lib/kyuubi/jars/hadoop-client-runtime-3.3.6.jar:/usr/lib/kyuubi/jars/hadoop-client-api-3.3.6.jar: org.apache.kyuubi.engine.hive.HiveSQLEngine \
--conf kyuubi.session.user=root1 \
--conf kyuubi.engine.id=b1313aae-c609-4698-bc70-6581168961f8 \
--conf hive.engine.name=kyuubi_USER_HIVE_SQL_root1_default_b1313aae-c609-4698-bc70-6581168961f8 \
--conf hive.server2.thrift.resultset.default.fetch.size=1000 \
--conf kyuubi.backend.engine.exec.pool.keepalive.time=PT1M \
--conf kyuubi.backend.engine.exec.pool.shutdown.timeout=PT20S \
--conf kyuubi.backend.engine.exec.pool.size=100 \
--conf kyuubi.backend.engine.exec.pool.wait.queue.size=100 \
--conf kyuubi.backend.server.exec.pool.keepalive.time=PT1M \
--conf kyuubi.backend.server.exec.pool.shutdown.timeout=PT20S \
--conf kyuubi.backend.server.exec.pool.size=100 \
--conf kyuubi.backend.server.exec.pool.wait.queue.size=100 \
--conf kyuubi.batch.application.check.interval=PT10S \
--conf kyuubi.batch.application.starvation.timeout=PT3M \
--conf kyuubi.batch.session.idle.timeout=PT6H \
--conf kyuubi.client.ipAddress=192.168.1.110 \
--conf kyuubi.client.version=1.9.1 \
--conf kyuubi.engine.event.json.log.path=/var/lib/kyuubi/engine/event \
--conf kyuubi.engine.flink.application.jars= \
--conf kyuubi.engine.flink.extra.classpath= \
--conf kyuubi.engine.flink.java.options= \
--conf kyuubi.engine.flink.memory=1g \
--conf kyuubi.engine.hive.deploy.mode=local \
--conf kyuubi.engine.hive.event.loggers=JSON \
--conf kyuubi.engine.hive.extra.classpath= \
--conf kyuubi.engine.hive.java.options= \
--conf kyuubi.engine.hive.memory=1g \
--conf kyuubi.engine.pool.name=kyuubi-engine-pool \
--conf kyuubi.engine.pool.selectPolicy=RANDOM \
--conf kyuubi.engine.pool.size=-1 \
--conf kyuubi.engine.session.initialize.sql= \
--conf kyuubi.engine.share.level=USER \
--conf kyuubi.engine.spark.event.loggers=SPARK \
--conf kyuubi.engine.submit.time=1721366841647 \
--conf kyuubi.engine.submit.timeout=PT30S \
--conf kyuubi.engine.type=HIVE_SQL \
--conf kyuubi.engine.ui.retainedSessions=200 \
--conf kyuubi.engine.ui.retainedStatements=200 \
--conf kyuubi.engine.ui.stop.enabled=true \
--conf kyuubi.engine.yarn.cores=1 \
--conf kyuubi.engine.yarn.java.options= \
--conf kyuubi.engine.yarn.memory=1024 \
--conf kyuubi.engine.yarn.queue=default \
--conf kyuubi.engine.yarn.submit.timeout=PT1M \
--conf kyuubi.event.async.pool.keepalive.time=PT1M \
--conf kyuubi.event.async.pool.size=8 \
--conf kyuubi.event.async.pool.wait.queue.size=100 \
--conf kyuubi.frontend.connection.url.use.hostname=true \
--conf kyuubi.frontend.max.message.size=104857600 \
--conf kyuubi.frontend.max.worker.threads=999 \
--conf kyuubi.frontend.min.worker.threads=9 \
--conf kyuubi.frontend.protocols=THRIFT_BINARY,REST \
--conf kyuubi.frontend.proxy.http.client.ip.header=X-Real-IP \
--conf kyuubi.frontend.rest.jetty.stopTimeout=PT10S \
--conf kyuubi.frontend.rest.max.worker.threads=999 \
--conf kyuubi.frontend.thrift.binary.ssl.disallowed.protocols=SSLv2,SSLv3 \
--conf kyuubi.frontend.thrift.binary.ssl.enabled=false \
--conf kyuubi.frontend.thrift.max.message.size=104857600 \
--conf kyuubi.frontend.thrift.max.worker.threads=999 \
--conf kyuubi.frontend.thrift.min.worker.threads=9 \
--conf kyuubi.frontend.thrift.worker.keepalive.time=PT1M \
--conf kyuubi.ha.addresses=hadoop01:2181,hadoop02:2181,hadoop03:2181 \
--conf kyuubi.ha.client.class=org.apache.kyuubi.ha.client.zookeeper.ZookeeperDiscoveryClient \
--conf kyuubi.ha.engine.ref.id=b1313aae-c609-4698-bc70-6581168961f8 \
--conf kyuubi.ha.namespace=/kyuubi_1.9.1_USER_HIVE_SQL/root1/default \
--conf kyuubi.ha.zookeeper.acl.enabled=false \
--conf kyuubi.ha.zookeeper.auth.type=NONE \
--conf kyuubi.ha.zookeeper.connection.base.retry.wait=1000 \
--conf kyuubi.ha.zookeeper.connection.max.retries=3 \
--conf kyuubi.ha.zookeeper.connection.max.retry.wait=30000 \
--conf kyuubi.ha.zookeeper.connection.retry.policy=EXPONENTIAL_BACKOFF \
--conf kyuubi.ha.zookeeper.connection.timeout=15000 \
--conf kyuubi.ha.zookeeper.engine.auth.type=NONE \
--conf kyuubi.ha.zookeeper.node.creation.timeout=PT2M \
--conf kyuubi.ha.zookeeper.session.timeout=60000 \
--conf kyuubi.metadata.cleaner.enabled=true \
--conf kyuubi.metadata.cleaner.interval=PT30M \
--conf kyuubi.metadata.max.age=PT128H \
--conf kyuubi.metadata.recovery.threads=10 \
--conf kyuubi.metadata.request.async.retry.enabled=true \
--conf kyuubi.metadata.request.async.retry.queue.size=65536 \
--conf kyuubi.metadata.request.async.retry.threads=10 \
--conf kyuubi.metadata.request.retry.interval=PT5S \
--conf kyuubi.metadata.store.class=org.apache.kyuubi.server.metadata.jdbc.JDBCMetadataStore \
--conf kyuubi.metrics.console.interval=PT20S \
--conf kyuubi.metrics.enabled=false \
--conf kyuubi.metrics.reporters= \
--conf kyuubi.operation.query.timeout=3600000 \
--conf kyuubi.operation.scheduler.pool=fair \
--conf kyuubi.server.info.provider=ENGINE \
--conf kyuubi.server.ipAddress=192.168.1.110 \
--conf kyuubi.session.check.interval=PT5M \
--conf kyuubi.session.close.on.disconnect=true \
--conf kyuubi.session.connection.url=hadoop01:10009 \
--conf kyuubi.session.engine.alive.timeout=PT2M \
--conf kyuubi.session.engine.check.interval=PT1M \
--conf kyuubi.session.engine.idle.timeout=PT30M \
--conf kyuubi.session.engine.initialize.timeout=PT5M \
--conf kyuubi.session.engine.launch.async=true \
--conf kyuubi.session.engine.log.timeout=PT24H \
--conf kyuubi.session.idle.timeout=PT6H \
--conf kyuubi.session.real.user=root1 \
--conf spark.cleaner.periodicGC.interval=5min \
--conf spark.driver.cores=1 \
--conf spark.driver.maxResultSize=1g \
--conf spark.dynamicAllocation.cachedExecutorIdleTimeout=30min \
--conf spark.dynamicAllocation.enabled=true \
--conf spark.dynamicAllocation.executorAllocationRatio=0.5 \
--conf spark.dynamicAllocation.executorIdleTimeout=60s \
--conf spark.dynamicAllocation.initialExecutors=2 \
--conf spark.dynamicAllocation.maxExecutors=25 \
--conf spark.dynamicAllocation.minExecutors=2 \
--conf spark.dynamicAllocation.schedulerBacklogTimeout=1s \
--conf spark.dynamicAllocation.shuffleTracking.enabled=false \
--conf spark.dynamicAllocation.shuffleTracking.timeout=30min \
--conf spark.dynamicAllocation.sustainedSchedulerBacklogTimeout=1s \
--conf spark.hadoop.cacheConf=false \
--conf spark.io.compression.lz4.blockSize=128kb \
--conf spark.master=yarn \
--conf spark.scheduler.allocation.file=hdfs:///user/spark/conf/kyuubi-fairscheduler.xml \
--conf spark.scheduler.mode=FAIR \
--conf spark.shuffle.file.buffer=1m \
--conf spark.shuffle.io.backLog=8192 \
--conf spark.shuffle.push.enabled=true \
--conf spark.shuffle.service.enabled=true \
--conf spark.shuffle.service.index.cache.size=100m \
--conf spark.shuffle.service.port=17337 \
--conf spark.shuffle.service.removeShuffle=false \
--conf spark.sql.adaptive.advisoryPartitionSizeInBytes=128M \
--conf spark.sql.adaptive.autoBroadcastJoinThreshold=10MB \
--conf spark.sql.adaptive.coalescePartitions.enabled=true \
--conf spark.sql.adaptive.coalescePartitions.initialPartitionNum=8192 \
--conf spark.sql.adaptive.coalescePartitions.minPartitionSize=1MB \
--conf spark.sql.adaptive.coalescePartitions.parallelismFirst=true \
--conf spark.sql.adaptive.enabled=true \
--conf spark.sql.adaptive.forceOptimizeSkewedJoin=false \
--conf spark.sql.adaptive.localShuffleReader.enabled=true \
--conf spark.sql.adaptive.optimizeSkewsInRebalancePartitions.enabled=true \
--conf spark.sql.adaptive.rebalancePartitionsSmallPartitionFactor=0.2 \
--conf spark.sql.adaptive.skewJoin.enabled=true \
--conf spark.sql.adaptive.skewJoin.skewedPartitionFactor=5 \
--conf spark.sql.adaptive.skewJoin.skewedPartitionThresholdInBytes=256MB \
--conf spark.sql.autoBroadcastJoinThreshold=10MB \
--conf spark.sql.hive.convertMetastoreOrc=true \
--conf spark.sql.hive.metastore.jars=/usr/lib/hive/lib/* \
--conf spark.sql.hive.metastore.version=3.1.3 \
--conf spark.sql.orc.filterPushdown=true \
--conf spark.sql.statistics.fallBackToHdfs=true \
--conf spark.submit.deployMode=client
2024-07-19 13:27:21.652 INFO KyuubiSessionManager-exec-pool: Thread-630 org.apache.kyuubi.engine.ProcBuilder: Logging to /var/lib/kyuubi/root1/kyuubi-hive-sql-engine.log.2
2024-07-19 13:27:22.763 INFO Curator-Framework-0 org.apache.kyuubi.shaded.curator.framework.imps.CuratorFrameworkImpl: backgroundOperationsLoop exiting
2024-07-19 13:27:22.872 INFO KyuubiSessionManager-exec-pool: Thread-630-EventThread org.apache.kyuubi.shaded.zookeeper.ClientCnxn: EventThread shut down for session: 0x200000139453541
2024-07-19 13:27:22.872 INFO KyuubiSessionManager-exec-pool: Thread-630 org.apache.kyuubi.shaded.zookeeper.ZooKeeper: Session: 0x200000139453541 closed
2024-07-19 13:27:22.876 INFO KyuubiSessionManager-exec-pool: Thread-630 org.apache.kyuubi.operation.LaunchEngine: Processing root1's query[d093f68e-7596-4c81-a0e0-281b1d9c3c71]: RUNNING_STATE -> ERROR_STATE, time taken: 1.344 seconds
Error: Could not find or load main class
Error: org.apache.kyuubi.KyuubiSQLException: Failed to detect the root cause, please check /var/lib/kyuubi/root1/kyuubi-hive-sql-engine.log.2 at server side if necessary. The last 10 line(s) of log are:
Error: Could not find or load main class
at org.apache.kyuubi.KyuubiSQLException$.apply(KyuubiSQLException.scala:69)
at org.apache.kyuubi.engine.ProcBuilder.getError(ProcBuilder.scala:277)
at org.apache.kyuubi.engine.ProcBuilder.getError$(ProcBuilder.scala:270)
at org.apache.kyuubi.engine.hive.HiveProcessBuilder.getError(HiveProcessBuilder.scala:38)
at org.apache.kyuubi.engine.EngineRef.$anonfun$create$1(EngineRef.scala:236)
at org.apache.kyuubi.ha.client.zookeeper.ZookeeperDiscoveryClient.tryWithLock(ZookeeperDiscoveryClient.scala:166)
at org.apache.kyuubi.engine.EngineRef.tryWithLock(EngineRef.scala:178)
at org.apache.kyuubi.engine.EngineRef.create(EngineRef.scala:183)
at org.apache.kyuubi.engine.EngineRef.$anonfun$getOrCreate$1(EngineRef.scala:317)
at scala.Option.getOrElse(Option.scala:189)
at org.apache.kyuubi.engine.EngineRef.getOrCreate(EngineRef.scala:317)
at org.apache.kyuubi.session.KyuubiSessionImpl.$anonfun$openEngineSession$2(KyuubiSessionImpl.scala:159)
at org.apache.kyuubi.session.KyuubiSessionImpl.$anonfun$openEngineSession$2$adapted(KyuubiSessionImpl.scala:133)
at org.apache.kyuubi.ha.client.DiscoveryClientProvider$.withDiscoveryClient(DiscoveryClientProvider.scala:36)
at org.apache.kyuubi.session.KyuubiSessionImpl.$anonfun$openEngineSession$1(KyuubiSessionImpl.scala:133)
at org.apache.kyuubi.session.KyuubiSession.handleSessionException(KyuubiSession.scala:49)
at org.apache.kyuubi.session.KyuubiSessionImpl.openEngineSession(KyuubiSessionImpl.scala:133)
at org.apache.kyuubi.operation.LaunchEngine.$anonfun$runInternal$1(LaunchEngine.scala:60)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750) (state=,code=0)
Kyuubi Server Configurations
1 ##Authentication
2 kyuubi.authentication NONE
3 ## Backend
4 kyuubi.backend.engine.exec.pool.keepalive.time PT1M
5 kyuubi.backend.engine.exec.pool.shutdown.timeout PT20S
6 kyuubi.backend.engine.exec.pool.size 100
7 kyuubi.backend.engine.exec.pool.wait.queue.size 100
8 kyuubi.backend.server.exec.pool.keepalive.time PT1M
9 kyuubi.backend.server.exec.pool.shutdown.timeout PT20S
10 kyuubi.backend.server.exec.pool.size 100
11 kyuubi.backend.server.exec.pool.wait.queue.size 100
12 kyuubi.backend.server.event.json.log.path /var/lib/kyuubi/server/event
13 kyuubi.backend.server.event.loggers JSON
14 ## Batch
15 kyuubi.batch.application.check.interval PT10S
16 kyuubi.batch.application.starvation.timeout PT3M
17 kyuubi.batch.session.idle.timeout PT6H
18 ## Engine
19 kyuubi.engine.event.json.log.path /var/lib/kyuubi/engine/event
20 #yarn application mode.
21 kyuubi.engine.flink.application.jars
22 #yarn session mode.
23 kyuubi.engine.flink.extra.classpath
24 kyuubi.engine.flink.java.options
25 kyuubi.engine.flink.memory 1g
26 # locla:like hive-cli yarn: Application Maste
27 kyuubi.engine.hive.deploy.mode LOCAL
28 kyuubi.engine.hive.event.loggers JSON
29 kyuubi.engine.hive.extra.classpath
30 kyuubi.engine.hive.java.options
31 kyuubi.engine.hive.memory 1g
32 kyuubi.engine.pool.name kyuubi-engine-pool
33 kyuubi.engine.pool.selectPolicy RANDOM
34 kyuubi.engine.pool.size -1
35 kyuubi.engine.pool.size.threshold 9
36 kyuubi.engine.session.initialize.sql
37 kyuubi.engine.share.level USER
38 kyuubi.engine.spark.event.loggers SPARK
39 kyuubi.engine.submit.timeout PT30S
40 kyuubi.engine.type SPARK_SQL
41 kyuubi.engine.ui.retainedSessions 200
42 kyuubi.engine.ui.retainedStatements 200
43 kyuubi.engine.ui.stop.enabled true
44 kyuubi.engine.yarn.cores 1
45 kyuubi.engine.yarn.java.options
46 kyuubi.engine.yarn.memory 1024
47 kyuubi.engine.yarn.queue default
48 kyuubi.engine.yarn.submit.timeout PT1M
49 ## Event
50 kyuubi.event.async.pool.keepalive.time PT1M
51 kyuubi.event.async.pool.size 8
52 kyuubi.event.async.pool.wait.queue.size 100
53 ##Frontend
54 kyuubi.frontend.advertised.host hadoop01
55 kyuubi.frontend.connection.url.use.hostname true
56 kyuubi.frontend.max.message.size 104857600
57 kyuubi.frontend.max.worker.threads 999
58 kyuubi.frontend.min.worker.threads 9
59 kyuubi.frontend.protocols THRIFT_BINARY,REST
60 kyuubi.frontend.proxy.http.client.ip.header X-Real-IP
61 kyuubi.frontend.rest.bind.port 10099
62 kyuubi.frontend.rest.jetty.stopTimeout PT10S
63 kyuubi.frontend.rest.max.worker.threads 999
64 kyuubi.frontend.thrift.binary.bind.port 10009
65 kyuubi.frontend.thrift.binary.ssl.disallowed.protocols SSLv2,SSLv3
66 kyuubi.frontend.thrift.binary.ssl.enabled false
67 kyuubi.frontend.thrift.max.message.size 104857600
68 kyuubi.frontend.thrift.max.worker.threads 999
69 kyuubi.frontend.thrift.min.worker.threads 9
70 kyuubi.frontend.thrift.worker.keepalive.time PT1M
71 ## ha
72 kyuubi.ha.addresses hadoop01:2181,hadoop02:2181,hadoop03:2181
73 kyuubi.ha.client.class org.apache.kyuubi.ha.client.zookeeper.ZookeeperDiscoveryClient
74 kyuubi.ha.namespace kyuubi
75 kyuubi.ha.zookeeper.acl.enabled false
76 kyuubi.ha.zookeeper.auth.type NONE
77 kyuubi.ha.zookeeper.connection.base.retry.wait 1000
78 kyuubi.ha.zookeeper.connection.max.retries 3
79 kyuubi.ha.zookeeper.connection.max.retry.wait 30000
80 kyuubi.ha.zookeeper.connection.retry.policy EXPONENTIAL_BACKOFF
81 kyuubi.ha.zookeeper.connection.timeout 15000
82 kyuubi.ha.zookeeper.engine.auth.type NONE
83 kyuubi.ha.zookeeper.node.creation.timeout PT2M
84 kyuubi.ha.zookeeper.session.timeout 60000
85 ##Metadata
86 kyuubi.metadata.cleaner.enabled true
87 kyuubi.metadata.cleaner.interval PT30M
88 kyuubi.metadata.max.age PT128H
89 kyuubi.metadata.recovery.threads 10
90 kyuubi.metadata.request.async.retry.enabled true
91 kyuubi.metadata.request.async.retry.queue.size 65536
92 kyuubi.metadata.request.async.retry.threads 10
93 kyuubi.metadata.request.retry.interval PT5S
94 kyuubi.metadata.store.class org.apache.kyuubi.server.metadata.jdbc.JDBCMetadataStore
95 kyuubi.metadata.store.jdbc.database.schema.init true
96 kyuubi.metadata.store.jdbc.database.type MYSQL
97 kyuubi.metadata.store.jdbc.driver com.mysql.cj.jdbc.Driver
98 kyuubi.metadata.store.jdbc.password qishu1@3
99 kyuubi.metadata.store.jdbc.user root
100 kyuubi.metadata.store.jdbc.url jdbc:mysql://192.168.1.110:13306/kyuubi_metadata
101 ##Metrics
102 kyuubi.metrics.console.interval PT20S
103 kyuubi.metrics.enabled false
104 kyuubi.metrics.reporters
105 ##operation
106 kyuubi.operation.query.timeout 3600000
107 kyuubi.operation.scheduler.pool fair
## server
109 kyuubi.server.administrators kyuubi
110 kyuubi.server.info.provider ENGINE
111 ## session
112 kyuubi.session.check.interval PT5M
113 kyuubi.session.close.on.disconnect true
114 kyuubi.session.engine.alive.timeout PT2M
115 kyuubi.session.engine.check.interval PT1M
116 kyuubi.session.engine.idle.timeout PT30M
117 kyuubi.session.engine.initialize.timeout PT5M
118 kyuubi.session.engine.launch.async true
119 kyuubi.session.engine.log.timeout PT24H
120 kyuubi.session.idle.timeout PT6H
121 ##spark
122 spark.cleaner.periodicGC.interval 5min
123 spark.driver.cores 1
124 spark.driver.maxResultSize 1g
125 spark.dynamicAllocation.cachedExecutorIdleTimeout 30min
126 spark.dynamicAllocation.enabled true
127 spark.dynamicAllocation.executorAllocationRatio 0.5
128 spark.dynamicAllocation.executorIdleTimeout 60s
129 spark.dynamicAllocation.initialExecutors 2
130 spark.dynamicAllocation.maxExecutors 25
131 spark.dynamicAllocation.minExecutors 2
132 spark.dynamicAllocation.schedulerBacklogTimeout 1s
133 spark.dynamicAllocation.shuffleTracking.enabled false
134 spark.dynamicAllocation.shuffleTracking.timeout 30min
135 spark.dynamicAllocation.sustainedSchedulerBacklogTimeout 1s
136 spark.hadoop.cacheConf false
137 spark.io.compression.lz4.blockSize 128kb
138 spark.master yarn
139 spark.scheduler.allocation.file hdfs:///user/spark/conf/kyuubi-fairscheduler.xml
140 spark.scheduler.mode FAIR
141 spark.shuffle.file.buffer 1m
142 spark.shuffle.io.backLog 8192
143 spark.shuffle.push.enabled true
144 spark.shuffle.service.enabled true
145 spark.shuffle.service.index.cache.size 100m
146 spark.shuffle.service.port 17337
147 spark.shuffle.service.removeShuffle false
148 spark.sql.adaptive.advisoryPartitionSizeInBytes 128M
149 spark.sql.adaptive.autoBroadcastJoinThreshold 10MB
150 spark.sql.adaptive.coalescePartitions.enabled true
151 spark.sql.adaptive.coalescePartitions.initialPartitionNum 8192
152 spark.sql.adaptive.coalescePartitions.minPartitionSize 1MB
153 spark.sql.adaptive.coalescePartitions.parallelismFirst true
154 spark.sql.adaptive.enabled true
155 spark.sql.adaptive.forceOptimizeSkewedJoin false
156 spark.sql.adaptive.localShuffleReader.enabled true
157 spark.sql.adaptive.optimizeSkewsInRebalancePartitions.enabled true
158 spark.sql.adaptive.rebalancePartitionsSmallPartitionFactor 0.2
159 spark.sql.adaptive.skewJoin.enabled true
160 spark.sql.adaptive.skewJoin.skewedPartitionFactor 5
161 spark.sql.adaptive.skewJoin.skewedPartitionThresholdInBytes 256MB
161 spark.sql.adaptive.skewJoin.skewedPartitionThresholdInBytes 256MB
162 spark.sql.autoBroadcastJoinThreshold 10MB
163 spark.sql.hive.convertMetastoreOrc true
164 spark.sql.hive.metastore.jars /usr/lib/hive/lib/*
165 spark.sql.hive.metastore.version 3.1.3
166 spark.sql.orc.filterPushdown true
167 spark.sql.statistics.fallBackToHdfs true
168 spark.submit.deployMode client
Kyuubi Engine Configurations
as Kyuubi Server Configurations
Additional context
./beeline -u "jdbc:hive2://hadoop01:2181,hadoop02:2181,hadoop03:2181/default;serviceDiscoveryMode=zooKeeper;zooKeeperNamespace=kyuubi" -n root --hiveconf kyuubi.engine.type=FLINK_SQL
flink engine also have same problem
Are you willing to submit PR?
- Yes. I would be willing to submit a PR with guidance from the Kyuubi community to fix.
- No. I cannot submit a PR at this time.
Hello @CrazyBeeline,
Thanks for finding the time to report the issue!
We really appreciate the community's efforts to improve Apache Kyuubi.
Looks like classpath issues, we have an out-of-box sandbox env for testing purposes, you can try and apply your configurations gradually to find the bad ones https://github.com/awesome-kyuubi/hadoop-testing
kyuubi.engine.hive.java.options and kyuubi.engine.flink.java.options is empty remove it from kyuubi-defaults.conf work well
kyuubi.engine.hive.java.options and kyuubi.engine.flink.java.options is empty remove it from kyuubi-defaults.conf work well
Does this result in an extra :
at the end of the classpath? Can you submit a PR to fix this?
hadoop-client-api-3.3.6.jar: org.apache.kyuubi.engine.hive.HiveSQLEngine
Does this result in an extra : at the end of the classpath? Can you submit a PR to fix this?
Sorry, my understanding was wrong.
The root cause is that the empty string configured in kyuubi.engine.flink.java.options is used as the main class, resulting in the Could not find or load main class
error.
Test: