lakeFS
lakeFS copied to clipboard
lakeFS - Data version control for your data lake | Git for data
Allow to safely replace the database credentials used by lakeFS to connect to Postgres. This is useful, for example, when connecting to AWS RDS using IAM roles: this method provides...
GC is a Spark job and our users must use spark-submit to run it. Databricks users who usually use databricks notebooks, don't necessarily know how to run spark-submit from databricks....
It accepts a repository uri and commit as an argument which is a misalignment with the rest of the system.
Currently, the [docs](https://docs.lakefs.io/integrations/spark.html#access-lakefs-using-the-lakefs-specific-hadoop-filesystem) do not state which versions of Spark, Databricks, Scala are supported
Create table or schema on lakeFS fails when: - Using Hive metastore - The location is the root of the repository - The repository used in the location is empty...
To run lakeFS Garbage collection, a lakeFS user needs to have the following lakeFS policy attached: ``` { "statement": [ { "action": [ "fs:ListObjects", "fs:ReadCommit", "fs:ReadRepository", "fs:ReadObject", "fs:ListRepositories", "retention:GetGarbageCollectionRules", "retention:PrepareGarbageCollectionCommits"...