hyperleaup
hyperleaup copied to clipboard
Create and manipulate Tableau Hyper files from Apache Spark DataFrames and Spark SQL
https://community.databricks.com/s/question/0D53f00001Tgle2CAB/hyperleaup-to-push-data-to-tableau-server ` --------------------------------------------------------------------------- TypeError Traceback (most recent call last) in 7 8 # Publish the Hyper File! ----> 9 luid = hf.publish(tableau_server_url=tableau_server, 10 username=username, 11 password=password, /local_disk0/.ephemeral_nfs/envs/pythonEnv-9ce478c3-3c92-4e39-8fb0-576418e91fb7/lib/python3.8/site-packages/hyperleaup/hyper_file.py in publish(self, tableau_server_url,...
@goodwillpunning this is a great library. I have been using this for about 3 months to refresh my tableau extracts. Sometimes an analyst might wanna track the past state of...
@goodwillpunning Thank you for creating such a wonderful tool. When I use hf.save function, I get path of hyperfile. However, when I download the hyperfile it just shows headers and...
ServerResponseError: 403007: Forbidden The datasource 'xxxxxxxxx' already exists and may not be overwritten without the 'overwrite' flag set to 'true'.
Bumps [pyspark](https://github.com/apache/spark) from 3.1.2 to 3.1.3. Commits d1f8a50 Preparing Spark release v3.1.3-rc4 7540421 Preparing development version 3.1.4-SNAPSHOT b8c0799 Preparing Spark release v3.1.3-rc3 0a7eda3 [SPARK-38075][SQL][3.1] Fix hasNext in HiveScriptTransformationExec's pro... 91db9a3...
Issue when creating hyper file from spark dataframe. Issue with file types, decimal and timestamp Column 'xxx' in Parquet file '/tmp/hyperleaup/hyper_files/testtesttest.snappy.parquet' cannot be scanned by Hyper due to its data...
Right now the generated parquet files have to be saved into local tmp folder first and then call .save() to save hyper into Volumes (a public folder). There is no...
https://github.com/goodwillpunning/hyperleaup/blob/bf2a4e6bce6b613ee36fefbc14c68f87ce218c6c/src/hyperleaup/creator.py#L306 The type for the attribute _parquet_paths_ is written as **list[str]** but it must have been used the type class List as **List[str]** (with capital L)
- Add databricks-sdk dependency since it is not included on DBR 11.3LTS or lower. - Add Dustin Vannoy to author list and readme.
A Boolean value that is used to publish data sources asynchronously. If you set this value to false (the default), the publishing process runs as a synchronous process. If a...