Martin Mauch

Results 403 comments of Martin Mauch

Ok, so it fails during schema inference. Are you able to specify a schema manually?

Did you try the combination of specifying a schema and using `maxRowsInMemory`?

Can you try the newest version (`0.20.3`) and also the version using `.format("excel")`?

Does that combination of Spark and spark-excel even work?? Please always try the newest version of spark-excel when posting issues. I don't think this solves the problem in this case,...

@williamdphillips in [this line](https://github.com/crealytics/spark-excel/blob/35a5b9b8d67d133345040ab642a897cba1f6b519/src/main/scala/com/crealytics/spark/excel/DefaultSource.scala#L38) we're passing the `HadoopConfiguration` to the `WorkbookReader` and then use it [here](https://github.com/crealytics/spark-excel/blob/35a5b9b8d67d133345040ab642a897cba1f6b519/src/main/scala/com/crealytics/spark/excel/WorkbookReader.scala#L65) to actually read from the filesystem. There might be another way to read from...

There was a [similar issue in Hudi](https://github.com/apache/hudi/issues/817). @johnboyer have you tried building spark-excel with 2.12.15? Does it fix the problem?

Hi @EnverOsmanov, thanks for the PR! I'm slightly worried that `.takeWhile` slightly modifies the semantics to stop after having found the first non-matching index. At the moment this doesn't matter...

Hmm, maybe it is to be able to do the ```scala r.getCell(_, MissingCellPolicy.CREATE_NULL_AS_BLANK) ``` @quanghgx could you chime in here?

The documentation reads like this is only supported for a few specific file formats: https://docs.databricks.com/ingestion/auto-loader/options.html#file-format-options Not sure if they are hard-coded somewhere, or one would need to implement a special...