confluentinc/kafka-connect-storage-cloud

host exception while reading hive table from pyspark..it is showing list of tables but not able to read data from hive table

Opened this issue · 2 comments

answer = self.gateway_client. send_command command) -> 1321 return_value = get_return_value 1322 answer, self.gateway_client, self.target_id, self.name) 1323 /ui/jupyterhub_data/anaconda3/lib/python3.8/site-packages/pyspark/sql/utils.py in deco(*a, **kw) 194 #
Hide where the exception came from that shows a non-Pythonic 195 # JVM exception message. --> 196 raise converted from None 197 else: raise 198

IllegalArgumentException: java.net.UnknownHostException: bdacdh-ns

@masseyke @jbaiera the issue is i am not able to integrate spark with hive on 6 node hadoop cluster. While running this pyspark script i am able to see the tables described on hive but when i try to execute select * query it is giving an error ....just wanted to know the required steps for the integration...like what changes to do with the configuration files and all..
the exception it is showing ...unknownhostException- (bdacdh-ns)---this is the name of the name node...cant understand why it is showing that..
it would be really helpful ...if you put light on all the procedure for hive spark integration

UnknownHostException is caused by your DNS config.

Unfortunately, this isn't the correct repo to report Spark issues. Please try Stackoverflow.