great_expectations
great_expectations copied to clipboard
Always know what to expect from your data.
pyspark.sql.utils.AnalysisException: "Table or view not found: when creating an expectations suite
**Describe the bug** Hive metastore is not found, and so the hive table is not found. **To Reproduce** When running ``` great_expectations suite scaffold ge_demo_tbl.warning ``` I always get `pyspark.sql.utils.AnalysisException:...
**Describe the bug** Data Docs are created in the Google Cloud Storage (bucket). The Data Docs get overwritten each time a new test is executed via Cronjob. i.e. the output...
**Describe the bug** A clear and concise description of what the bug is. **To Reproduce** Steps to reproduce the behavior: 1. Create a datasource accessing a Teradata DB 2. Create...
**Describe the bug** After getting BatchRequest and ExpectationSuite, I tried to programmatically create a `SimpleCheckpoint`, then persistent it to config file. There was an error when convert Checkpoint to YAML...
**Is your feature request related to a problem? Please describe.** I really need to be able to do a 1 to 1 comparison on my batch/sample of data against another...
**Describe the bug** Issue related to storing Validation Result to PostgreSQL. I cannot create ge_validations_store tabel in the schema that I want. **To Reproduce** Follow the steps here: https://docs.greatexpectations.io/docs/guides/setup/configuring_metadata_stores/how_to_configure_a_validation_result_store_to_postgresql/ Everything...
**Describe the bug** When checkpoint ran with request_format COMPLETE with include_unexpected_rows e.g. result_format={'result_format': 'COMPLETE', 'partial_unexpected_count': 20, 'include_unexpected_rows': True} It returns either Error or None. **To Reproduce** For expectation_type = expect_column_values_to_not_be_null,...
**Describe the bug** Generation of expectations using historic data does not for column type TIMESTAMP using BigQuery as source. In BigQuery there is 2 types for date: - DATETIME -...
Description: When running a checkpoint within the V3 API, first I need to validate table metadata, and I only want to proceed to column validations if all the table validation...
**Reference** [Avanade team using great_expectations in Azure Machine Learning pipeline for data expectation tests](https://greatexpectations.io/case-studies/avanade-case-study/) My team is also implementing an ML solution using Azure Machine Learning pipeline. I would like...