我正在试图阅读Parquet文件使用Spark。spark版本是1.3.0。当我试图阅读时,我得到以下例外。非常感谢您的帮助。cdh版本:5.4.3
parquet.format.converter.parquetmetadataconverter.fromparquetstatistics(parquetmetadataconverter)处的java.lang.nullpointerexception。java:249)在parquet.format.converter.parquetmetadataconverter.fromparquetmetadata(parquetmetadataconverter。java:543)在parquet.format.converter.parquetmetadataconverter.readparquetmetadata(parquetmetadataconverter。java:520)在parquet.hadoop.parquetfilereader.readfooter(parquetfilereader。java:426)在parquet.hadoop.parquetfilereader.readfooter(parquetfilereader。java:381)在parquet.hadoop.parquetrecordreader.initializeinternalreader(parquetrecordreader。java:155) 在parquet.hadoop.parquetrecordreader.initialize(parquetrecordreader。java:138)在org.apache.spark.sql.sources.sqlnewhadooprdd$$anon$1。scala:153)位于org.apache.spark.sql.sources.sqlnewhadooprdd.compute(sqlnewhadooprdd)。scala:124)位于org.apache.spark.sql.sources.sqlnewhadooprdd.compute(sqlnewhadooprdd)。scala:66)在org.apache.spark.rdd.rdd.computeorreadcheckpoint(rdd。scala:277)在org.apache.spark.rdd.rdd.iterator(rdd。scala:244)在org.apache.spark.rdd.mappartitionsrdd.compute(mappartitionsrdd。scala:35)在org.apache.spark.rdd.rdd.computeorreadcheckpoint(rdd。scala:277)在org.apache.spark.rdd.rdd.iterator(rdd。scala:244)在org.apache.spark.rdd.mappartitionsrdd.compute(mappartitionsrdd。scala:35)在org.apache.spark.rdd.rdd.computeorreadcheckpoint(rdd。scala:277)在org.apache.spark.rdd.rdd.iterator(rdd。scala:244)在org.apache.spark.scheduler.shufflemaptask.runtask(shufflemaptask。scala:70)在org.apache.spark.scheduler.shufflemaptask.runtask(shufflemaptask。scala:41) 在org.apache.spark.scheduler.task.run(task。scala:70)在org.apache.spark.executor.executor$taskrunner.run(executor。scala:213)位于java.util.concurrent.threadpoolexecutor.runworker(threadpoolexecutor。java:1145)在java.util.concurrent.threadpoolexecutor$worker.run(threadpoolexecutor。java:615)在java.lang.thread.run(线程。java:745)
暂无答案!
目前还没有任何答案,快来回答吧!