spark sql failed to parse parquet -
spark sql parse parquet file, failed. cmd
create temporary table testparquet using org.apache.spark.sql.parquet options ( path "file:////tmp/table_generic.parquet" );
this error log:
java.lang.nullpointerexception @ parquet.format.converter.parquetmetadataconverter.fromparquetstatistics(parquetmetadataconverter.java:249) @ parquet.format.converter.parquetmetadataconverter.fromparquetmetadata(parquetmetadataconverter.java:543) @ parquet.format.converter.parquetmetadataconverter.readparquetmetadata(parquetmetadataconverter.java:520) @ parquet.hadoop.parquetfilereader.readfooter(parquetfilereader.java:426) @ org.apache.spark.sql.parquet.parquetrelation2$metadatacache$$anonfun$refresh$6.apply(newparquet.scala:298) @ org.apache.spark.sql.parquet.parquetrelation2$metadatacache$$anonfun$refresh$6.apply(newparquet.scala:297) @ scala.collection.parallel.mutable.pararray$map.leaf(pararray.scala:658) @ scala.collection.parallel.task$$anonfun$tryleaf$1.apply$mcv$sp(tasks.scala:54) @ scala.collection.parallel.task$$anonfun$tryleaf$1.apply(tasks.scala:53) @ scala.collection.parallel.task$$anonfun$tryleaf$1.apply(tasks.scala:53) @ scala.collection.parallel.task$class.tryleaf(tasks.scala:56) @ scala.collection.parallel.mutable.pararray$map.tryleaf(pararray.scala:650) @ scala.collection.parallel.adaptiveworkstealingtasks$wrappedtask$class.compute(tasks.scala:165) @ scala.collection.parallel.adaptiveworkstealingforkjointasks$wrappedtask.compute(tasks.scala:514) @ scala.concurrent.forkjoin.recursiveaction.exec(recursiveaction.java:160) @ scala.concurrent.forkjoin.forkjointask.doexec(forkjointask.java:260) @ scala.concurrent.forkjoin.forkjoinpool$workqueue.runtask(forkjoinpool.java:1339) @ scala.concurrent.forkjoin.forkjoinpool.runworker(forkjoinpool.java:1979) @ scala.concurrent.forkjoin.forkjoinworkerthread.run(forkjoinworkerthread.java:107)
the parquet's format has no errors drill has parsed correctly. why that?
Comments
Post a Comment