asfimport
asfimport
There was a previous attempt to fix avro logical-type support in avro in The proposed solution is not complete and missed the case of nested logical-types for avro
Setting up a specialized HadoopParquetReader allows for deprecating the Hadoop specific methods in ParquetReader as in PARQUET-2381. This could allow for an easier and more gentle transition away from the...
Setting up a specialized HadoopParquetWriter allows for deprecating the Hadoop specific methods in ParquetWriter as in PARQUET-2381. This could allow for an easier and more gentle transition away from the...
Deprecate methods relying on Hadoop classes when alternatives using Parquet interfaces are available. This would allow for fully decoupling from Hadoop for simple read/write (and potentially rewrite) operations in the...
Currently the codecs implemented by Parquet implement the Hadoop Configurable and CompressionCodec interfaces. As part of the effort to decouple from Hadoop there need to be alternatives to these Hadoop...
**Reporter**: [Chao Sun](https://issues.apache.org/jira/secure/ViewProfile.jspa?name=csun) / @sunchao #### Related issues: - [Improve Parquet IO Performance within cloud datalakes](https://github.com/apache/parquet-java/issues/2912) (is depended upon by) #### PRs and other links: - [GitHub Pull Request #1010](https://github.com/apache/parquet-mr/pull/1010)...
Hi Team, While exploring parquet encryption, it is found that, if a field in nested column is encrypted , and If I want to read this parquet directory from other...
remove the deprecated PathGlobPattern class and its uses from parquet-thrift The return types from the hadoop GlobPattern code changed in HADOOP-12436; in the class as is will not compile against...
Parquet uses reflection to load a hadoop2 input stream, falling back to a hadoop-1 compatible client if not found. All hadoop 2.0.2+ releases work with H2SeekableInputStream, so the binding to...
ParquetFileReader's implementation has the following flow (simplified) - - For every column -> Read from storage in 8MB blocks -> Read all uncompressed pages into output queue - From output...