Cleanup: update the name and information in files
Fixes #ISSUE_Number
What does this PR do?
This PR includes four commits, which are used to rename the old project name to "Apache Cloudberry" and update the contact information. See the commit messages for details. These four commits contain so many changes, so let be patient for the review.
Type of Change
- [ ] Bug fix (non-breaking change)
- [ ] New feature (non-breaking change)
- [ ] Breaking change (fix or feature with breaking changes)
- [x] Documentation update
Breaking Changes
Test Plan
- [ ] Unit tests added/updated
- [ ] Integration tests added/updated
- [ ] Passed
make installcheck - [ ] Passed
make -C src/test installcheck-cbdb-parallel
Impact
Performance:
User-facing changes:
Dependencies:
Checklist
- [x] Followed contribution guide
- [x] Added/updated documentation
- [x] Reviewed code for security implications
- [x] Requested review from cloudberry committers
Additional Context
⚠️ To skip CI: Add [skip ci] to your PR title. Only use when necessary! ⚠️
@tuhaihe How were these changes tested?
Hi @edespino thanks for your feedback and support. I will update my PR again to address the related comments.
The terms cloudberrydb and cbdb still appear in certain files, such as deploy/k8s and hd-ci/compile_cbdb.bash. For now, I would like to leave them unchanged and see how to update them later.
I can build my changes on one docker:
make[2]: Leaving directory '/cloudberry/gpcontrib/pxf_fdw'
make[1]: Leaving directory '/cloudberry/gpcontrib'
Apache Cloudberry (Incubating) installation complete.
[root@679abf66a2bb cloudberry]# git branch
main
* tuhaihe/cleanup-mails-and-names1126
[root@679abf66a2bb cloudberry]# cat /etc/os-release
NAME="Rocky Linux"
VERSION="9.4 (Blue Onyx)"
ID="rocky"
ID_LIKE="rhel centos fedora"
VERSION_ID="9.4"
PLATFORM_ID="platform:el9"
PRETTY_NAME="Rocky Linux 9.4 (Blue Onyx)"
ANSI_COLOR="0;32"
LOGO="fedora-logo-icon"
CPE_NAME="cpe:/o:rocky:rocky:9::baseos"
HOME_URL="https://rockylinux.org/"
BUG_REPORT_URL="https://bugs.rockylinux.org/"
SUPPORT_END="2032-05-31"
ROCKY_SUPPORT_PRODUCT="Rocky-Linux-9"
ROCKY_SUPPORT_PRODUCT_VERSION="9.4"
REDHAT_SUPPORT_PRODUCT="Rocky Linux"
REDHAT_SUPPORT_PRODUCT_VERSION="9.4"
However, when run the command make installcheck:
(using postmaster on Unix socket, default port)
============== dropping database "regression" ==============
psql: error: connection to server on socket "/tmp/.s.PGSQL.5432" failed: No such file or directory
Is the server running locally and accepting connections on that socket?
command failed: "/usr/local/cbdb/bin/psql" -X -c "DROP DATABASE IF EXISTS \"regression\"" "postgres"
make[1]: *** [GNUmakefile:208: installcheck-good] Error 2
make[1]: Leaving directory '/cloudberry/src/test/regress'
make: *** [GNUmakefile:146: installcheck] Error 2
I can build my changes on one docker:
make[2]: Leaving directory '/cloudberry/gpcontrib/pxf_fdw' make[1]: Leaving directory '/cloudberry/gpcontrib' Apache Cloudberry (Incubating) installation complete. [root@679abf66a2bb cloudberry]# git branch main * tuhaihe/cleanup-mails-and-names1126 [root@679abf66a2bb cloudberry]# cat /etc/os-release NAME="Rocky Linux" VERSION="9.4 (Blue Onyx)" ID="rocky" ID_LIKE="rhel centos fedora" VERSION_ID="9.4" PLATFORM_ID="platform:el9" PRETTY_NAME="Rocky Linux 9.4 (Blue Onyx)" ANSI_COLOR="0;32" LOGO="fedora-logo-icon" CPE_NAME="cpe:/o:rocky:rocky:9::baseos" HOME_URL="https://rockylinux.org/" BUG_REPORT_URL="https://bugs.rockylinux.org/" SUPPORT_END="2032-05-31" ROCKY_SUPPORT_PRODUCT="Rocky-Linux-9" ROCKY_SUPPORT_PRODUCT_VERSION="9.4" REDHAT_SUPPORT_PRODUCT="Rocky Linux" REDHAT_SUPPORT_PRODUCT_VERSION="9.4"However, when run the command
make installcheck:(using postmaster on Unix socket, default port) ============== dropping database "regression" ============== psql: error: connection to server on socket "/tmp/.s.PGSQL.5432" failed: No such file or directory Is the server running locally and accepting connections on that socket? command failed: "/usr/local/cbdb/bin/psql" -X -c "DROP DATABASE IF EXISTS \"regression\"" "postgres" make[1]: *** [GNUmakefile:208: installcheck-good] Error 2 make[1]: Leaving directory '/cloudberry/src/test/regress' make: *** [GNUmakefile:146: installcheck] Error 2
do u have make create-demo-cluster? also need export the GPPORT/PGPORT
Now I run make installcheck returns:
=========================
13 of 658 tests failed.
=========================
The differences that caused some tests to fail can be viewed in the
file "/home/gpadmin/cloudberry/src/test/regress/regression.diffs". A copy of the test summary that you see
above is saved in the file "/home/gpadmin/cloudberry/src/test/regress/regression.out".
make[1]: *** [GNUmakefile:208: installcheck-good] Error 1
make[1]: Leaving directory '/home/gpadmin/cloudberry/src/test/regress'
make: *** [GNUmakefile:146: installcheck] Error 2
regression.out.txt regression.diffs.txt
I will see how to fix them next week.
Now I run
make installcheckreturns:========================= 13 of 658 tests failed. ========================= The differences that caused some tests to fail can be viewed in the file "/home/gpadmin/cloudberry/src/test/regress/regression.diffs". A copy of the test summary that you see above is saved in the file "/home/gpadmin/cloudberry/src/test/regress/regression.out". make[1]: *** [GNUmakefile:208: installcheck-good] Error 1 make[1]: Leaving directory '/home/gpadmin/cloudberry/src/test/regress' make: *** [GNUmakefile:146: installcheck] Error 2regression.out.txt regression.diffs.txt
I will see how to fix them next week.
Feel free to ask engineers if you need any help.
The latest make installcheck command:
========================
4 of 659 tests failed.
========================
The differences that caused some tests to fail can be viewed in the file "/home/gpadmin/cloudberry/src/test/regress/regression.diffs". The diffs are the following:
diff -I HINT: -I CONTEXT: -I GP_IGNORE: -U3 /home/gpadmin/cloudberry/src/test/regress/expected/explain_optimizer.out /home/gpadmin/cloudberry/src/test/regress/results/explain.out
--- /home/gpadmin/cloudberry/src/test/regress/expected/explain_optimizer.out 2024-12-11 00:06:32.043419735 -0800
+++ /home/gpadmin/cloudberry/src/test/regress/results/explain.out 2024-12-11 00:06:32.068420210 -0800
@@ -395,7 +395,6 @@
}, +
"Settings": { +
"Optimizer": "Pivotal Optimizer (GPORCA)", +
- "optimizer": "on", +
"enable_parallel": "off", +
"parallel_setup_cost": "0", +
"parallel_tuple_cost": "0", +
diff -I HINT: -I CONTEXT: -I GP_IGNORE: -U3 /home/gpadmin/cloudberry/src/test/regress/expected/gpcopy_encoding.out /home/gpadmin/cloudberry/src/test/regress/results/gpcopy_encoding.out
--- /home/gpadmin/cloudberry/src/test/regress/expected/gpcopy_encoding.out 2024-12-11 00:12:53.291486702 -0800
+++ /home/gpadmin/cloudberry/src/test/regress/results/gpcopy_encoding.out 2024-12-11 00:12:53.293486751 -0800
@@ -21,7 +21,7 @@
copy enctest to '/tmp/enctest_utf_to_latin1-1' encoding 'latin1';
set client_encoding='latin1';
copy enctest to stdout;
-�
+�
copy enctest to '/tmp/enctest_utf_to_latin1-2';
-- Connect to 'latin1' database, and load back the files we just created.
-- This is to check that they were created correctly, and that the ENCODING
diff -I HINT: -I CONTEXT: -I GP_IGNORE: -U3 /home/gpadmin/cloudberry/src/test/regress/expected/dispatch_encoding.out /home/gpadmin/cloudberry/src/test/regress/results/dispatch_encoding.out
--- /home/gpadmin/cloudberry/src/test/regress/expected/dispatch_encoding.out 2024-12-11 00:15:14.095364710 -0800
+++ /home/gpadmin/cloudberry/src/test/regress/results/dispatch_encoding.out 2024-12-11 00:15:14.097364748 -0800
@@ -55,14 +55,16 @@
select raise_error(t) from enctest;
ERROR: raise_error called on "funny char Ä"
+CONTEXT: PL/pgSQL function raise_error(text) line 3 at RAISE
-- now do it again with latin1
set client_encoding='latin1';
select raise_notice(t) from enctest;
-NOTICE: raise_notice called on "funny char �"
+NOTICE: raise_notice called on "funny char �"
raise_notice
--------------
(1 row)
select raise_error(t) from enctest;
-ERROR: raise_error called on "funny char �"
+ERROR: raise_error called on "funny char �"
+CONTEXT: PL/pgSQL function raise_error(text) line 3 at RAISE
diff -I HINT: -I CONTEXT: -I GP_IGNORE: -U3 /home/gpadmin/cloudberry/src/test/regress/expected/cbdb_parallel.out /home/gpadmin/cloudberry/src/test/regress/results/cbdb_parallel.out
--- /home/gpadmin/cloudberry/src/test/regress/expected/cbdb_parallel.out 2024-12-11 00:40:32.554101669 -0800
+++ /home/gpadmin/cloudberry/src/test/regress/results/cbdb_parallel.out 2024-12-11 00:40:32.628103091 -0800
@@ -536,11 +536,8 @@
GP_IGNORE:(20 rows)
select count(*) from ao1, ao2 where ao1.x = ao2.x;
- count
----------
- 1080000
-(1 row)
-
+ERROR: could not resize shared memory segment "/PostgreSQL.3871811390" to 8388608 bytes: No space left on device
+ERROR: current transaction is aborted, commands ignored until end of transaction block
commit;
--
-- test parallel with indices
@@ -1563,45 +1560,13 @@
GP_IGNORE:(30 rows)
select * from rt1 join (select count(*) as c, sum(t1.a) as a from t1 join t2 using(a)) t3 on t3.c = rt1.a;
- a | b | c | a
----+---+---+---
-(0 rows)
-
+ERROR: could not resize shared memory segment "/PostgreSQL.288487486" to 4194304 bytes: No space left on device
set local enable_parallel = off;
+ERROR: current transaction is aborted, commands ignored until end of transaction block
explain(costs off, locus) select * from rt1 join (select count(*) as c, sum(t1.a) as a from t1 join t2 using(a)) t3 on t3.c = rt1.a;
-QUERY PLAN
-___________
- Hash Join
- Locus: Entry
- Hash Cond: (rt1.a = (count(*)))
- -> Gather Motion 1:1 (slice1; segments: 1)
- Locus: Entry
- -> Seq Scan on rt1
- Locus: SegmentGeneral
- -> Hash
- Locus: Entry
- -> Finalize Aggregate
- Locus: Entry
- -> Gather Motion 3:1 (slice2; segments: 3)
- Locus: Entry
- -> Partial Aggregate
- Locus: Hashed
- -> Hash Join
- Locus: Hashed
- Hash Cond: (t1.a = t2.a)
- -> Seq Scan on t1
- Locus: Hashed
- -> Hash
- Locus: Hashed
- -> Seq Scan on t2
- Locus: Hashed
-GP_IGNORE:(25 rows)
-
+ERROR: current transaction is aborted, commands ignored until end of transaction block
select * from rt1 join (select count(*) as c, sum(t1.a) as a from t1 join t2 using(a)) t3 on t3.c = rt1.a;
- a | b | c | a
----+---+---+---
-(0 rows)
-
+ERROR: current transaction is aborted, commands ignored until end of transaction block
abort;
--
-- Test final join path's parallel_workers should be same with join_locus whose
@@ -2247,44 +2212,24 @@
GP_IGNORE:(10 rows)
select sum(t1.c1) from t1 where c1 not in (select c2 from t2);
- sum
------
- 1
-(1 row)
-
+ERROR: could not resize shared memory segment "/PostgreSQL.3126407118" to 8388608 bytes: No space left on device
explain(costs off) select * from t1 where c1 not in (select c2 from t3_null);
-QUERY PLAN
-___________
- Gather Motion 6:1 (slice1; segments: 6)
- -> Parallel Hash Left Anti Semi (Not-In) Join
- Hash Cond: (t1.c1 = t3_null.c2)
- -> Parallel Seq Scan on t1
- -> Parallel Hash
- -> Broadcast Workers Motion 6:6 (slice2; segments: 6)
- -> Parallel Seq Scan on t3_null
-GP_IGNORE:(8 rows)
-
+ERROR: current transaction is aborted, commands ignored until end of transaction block
select * from t1 where c1 not in (select c2 from t3_null);
- c1 | c2
-----+----
-(0 rows)
-
+ERROR: current transaction is aborted, commands ignored until end of transaction block
-- non-parallel results.
set local enable_parallel = off;
+ERROR: current transaction is aborted, commands ignored until end of transaction block
select sum(t1.c1) from t1 where c1 not in (select c2 from t2);
- sum
------
- 1
-(1 row)
-
+ERROR: current transaction is aborted, commands ignored until end of transaction block
select * from t1 where c1 not in (select c2 from t3_null);
- c1 | c2
-----+----
-(0 rows)
-
+ERROR: current transaction is aborted, commands ignored until end of transaction block
drop table t1;
+ERROR: current transaction is aborted, commands ignored until end of transaction block
drop table t2;
+ERROR: current transaction is aborted, commands ignored until end of transaction block
drop table t3_null;
+ERROR: current transaction is aborted, commands ignored until end of transaction block
end;
--
-- End of Test Parallel-aware Hash Left Anti Semi (Not-In) Join.
@@ -2545,19 +2490,15 @@
begin;
set local max_parallel_workers_per_gather = 8;
select * from refresh_compare(true, false);
- parallel_is_better
---------------------
- t
-(1 row)
-
+ERROR: could not resize shared memory segment "/PostgreSQL.1207446260" to 8388608 bytes: No space left on device
+CONTEXT: SQL statement "refresh materialized view matv"
+PL/pgSQL function refresh_compare(boolean,boolean) line 30 at SQL statement
select * from refresh_compare(false, false);
- parallel_is_better
---------------------
- t
-(1 row)
-
+ERROR: current transaction is aborted, commands ignored until end of transaction block
drop function refresh_compare;
+ERROR: current transaction is aborted, commands ignored until end of transaction block
reset max_parallel_workers_per_gather;
+ERROR: current transaction is aborted, commands ignored until end of transaction block
end;
--
-- Parallel Create AO/AOCO Table AS
- Disk Space Issue
Some tests failed with the error message: ‘No space left on device’. This issue seems unusual, as the test computer’s hard drive has 600GB of space, which should be sufficient. The disk usage details are as follows:
[gpadmin@cdw cloudberry]$ df -H
Filesystem Size Used Avail Use% Mounted on
overlay 529G 41G 467G 8% /
tmpfs 68M 0 68M 0% /dev
shm 68M 1.9M 66M 3% /dev/shm
/dev/vda1 529G 41G 467G 8% /etc/hosts
tmpfs 68G 0 68G 0% /proc/acpi
tmpfs 68G 0 68G 0% /proc/scsi
tmpfs 68G 0 68G 0% /sys/firmware
- Encoding Errors
Some tests returned errors related to the latin1 encoding. These files were not modified in this pull request (PR).
- Other Errors
For other errors, I will work with the engineering team to determine the appropriate fixes. If the errors are not introduced by this PR, I will create an issue to document them for future resolution.
+ERROR: could not resize shared memory segment "/PostgreSQL.3126407118" to 8388608 bytes: No space left on device
here need to enlarge shm size .
+ERROR: could not resize shared memory segment "/PostgreSQL.3126407118" to 8388608 bytes: No space left on device
here need to enlarge shm size .
Thanks @yjhnupt for your suggestions. I use the command to run the build docker image:
docker run -it --rm -h cdw --shm-size=2gb apache/incubator-cloudberry:cbdb-build-rocky9-latest
Now, after running make installcheck, it returns:
========================
3 of 659 tests failed.
========================
diff -I HINT: -I CONTEXT: -I GP_IGNORE: -U3 /home/gpadmin/cloudberry/src/test/regress/expected/explain_optimizer.out /home/gpadmin/cloudberry/src/test/regress/results/explain.out
--- /home/gpadmin/cloudberry/src/test/regress/expected/explain_optimizer.out 2024-12-11 02:06:36.538896201 -0800
+++ /home/gpadmin/cloudberry/src/test/regress/results/explain.out 2024-12-11 02:06:36.562896655 -0800
@@ -395,7 +395,6 @@
}, +
"Settings": { +
"Optimizer": "Pivotal Optimizer (GPORCA)", +
- "optimizer": "on", +
"enable_parallel": "off", +
"parallel_setup_cost": "0", +
"parallel_tuple_cost": "0", +
diff -I HINT: -I CONTEXT: -I GP_IGNORE: -U3 /home/gpadmin/cloudberry/src/test/regress/expected/gpcopy_encoding.out /home/gpadmin/cloudberry/src/test/regress/results/gpcopy_encoding.out
--- /home/gpadmin/cloudberry/src/test/regress/expected/gpcopy_encoding.out 2024-12-11 02:12:57.455824245 -0800
+++ /home/gpadmin/cloudberry/src/test/regress/results/gpcopy_encoding.out 2024-12-11 02:12:57.457824282 -0800
@@ -21,7 +21,7 @@
copy enctest to '/tmp/enctest_utf_to_latin1-1' encoding 'latin1';
set client_encoding='latin1';
copy enctest to stdout;
-�
+�
copy enctest to '/tmp/enctest_utf_to_latin1-2';
-- Connect to 'latin1' database, and load back the files we just created.
-- This is to check that they were created correctly, and that the ENCODING
diff -I HINT: -I CONTEXT: -I GP_IGNORE: -U3 /home/gpadmin/cloudberry/src/test/regress/expected/dispatch_encoding.out /home/gpadmin/cloudberry/src/test/regress/results/dispatch_encoding.out
--- /home/gpadmin/cloudberry/src/test/regress/expected/dispatch_encoding.out 2024-12-11 02:15:40.607127519 -0800
+++ /home/gpadmin/cloudberry/src/test/regress/results/dispatch_encoding.out 2024-12-11 02:15:40.610127575 -0800
@@ -55,14 +55,16 @@
select raise_error(t) from enctest;
ERROR: raise_error called on "funny char Ä"
+CONTEXT: PL/pgSQL function raise_error(text) line 3 at RAISE
-- now do it again with latin1
set client_encoding='latin1';
select raise_notice(t) from enctest;
-NOTICE: raise_notice called on "funny char �"
+NOTICE: raise_notice called on "funny char �"
raise_notice
--------------
(1 row)
select raise_error(t) from enctest;
-ERROR: raise_error called on "funny char �"
+ERROR: raise_error called on "funny char �"
+CONTEXT: PL/pgSQL function raise_error(text) line 3 at RAISE
Getting closer! When running manually, I would run with one of the following. The first diff explicitly expects the orca optimizer to be set on (regardless of it being on by default).
With Orca query optimizer on:
PGOPTIONS='-c optimizer=on' make installcheck
With Orca query optimizer off:
PGOPTIONS='-c optimizer=off' make installcheck
NOTICE: raise_notice called on "funny char �"
locale refer to : https://www.cnblogs.com/williamjie/p/9303115.html
Getting closer! When running manually, I would run with one of the following. The first diff explicitly expects the orca optimizer to be set on (regardless of it being on by default).
With Orca query optimizer on:
PGOPTIONS='-c optimizer=on' make installcheckWith Orca query optimizer off:
PGOPTIONS='-c optimizer=off' make installcheck
Thanks @edespino! It works.
I ran the following two commands:
- No errors
$ PGCLIENTENCODING=UTF8 PGOPTIONS='-c optimizer=on' make -C src/test installcheck-cbdb-parallel
...
=======================
All 259 tests passed.
=======================
make[1]: Leaving directory '/home/gpadmin/cloudberry/src/test/isolation2'
make: Leaving directory '/home/gpadmin/cloudberry/src/test'
- Still has the encoding errors for latin1
$ PGCLIENTENCODING=UTF8 PGOPTIONS='-c optimizer=on' make installcheck
...
========================
2 of 659 tests failed.
========================
[gpadmin@cdw cloudberry]$ cat /home/gpadmin/cloudberry/src/test/regress/regression.diffs
diff -I HINT: -I CONTEXT: -I GP_IGNORE: -U3 /home/gpadmin/cloudberry/src/test/regress/expected/gpcopy_encoding.out /home/gpadmin/cloudberry/src/test/regress/results/gpcopy_encoding.out
--- /home/gpadmin/cloudberry/src/test/regress/expected/gpcopy_encoding.out 2024-12-12 03:03:35.691378647 -0800
+++ /home/gpadmin/cloudberry/src/test/regress/results/gpcopy_encoding.out 2024-12-12 03:03:35.693378680 -0800
@@ -21,7 +21,7 @@
copy enctest to '/tmp/enctest_utf_to_latin1-1' encoding 'latin1';
set client_encoding='latin1';
copy enctest to stdout;
-�
+�
copy enctest to '/tmp/enctest_utf_to_latin1-2';
-- Connect to 'latin1' database, and load back the files we just created.
-- This is to check that they were created correctly, and that the ENCODING
diff -I HINT: -I CONTEXT: -I GP_IGNORE: -U3 /home/gpadmin/cloudberry/src/test/regress/expected/dispatch_encoding.out /home/gpadmin/cloudberry/src/test/regress/results/dispatch_encoding.out
--- /home/gpadmin/cloudberry/src/test/regress/expected/dispatch_encoding.out 2024-12-12 03:06:09.112219102 -0800
+++ /home/gpadmin/cloudberry/src/test/regress/results/dispatch_encoding.out 2024-12-12 03:06:09.114219140 -0800
@@ -55,14 +55,16 @@
select raise_error(t) from enctest;
ERROR: raise_error called on "funny char Ä"
+CONTEXT: PL/pgSQL function raise_error(text) line 3 at RAISE
-- now do it again with latin1
set client_encoding='latin1';
select raise_notice(t) from enctest;
-NOTICE: raise_notice called on "funny char �"
+NOTICE: raise_notice called on "funny char �"
raise_notice
--------------
(1 row)
select raise_error(t) from enctest;
-ERROR: raise_error called on "funny char �"
+ERROR: raise_error called on "funny char �"
+CONTEXT: PL/pgSQL function raise_error(text) line 3 at RAISE
~For now, I can continue testing by setting up the correct encoding environment for Latin1. However, this issue doesn’t seem critical since our CI/CD pipeline has passed successfully.~
~Would it be possible to merge this PR now, or would you prefer that I ensure all tests pass locally before proceeding? Personally, I lean towards merging the PR now, as the failed tests are limited to my local environment settings. Otherwise, with new PRs being introduced, I would need to rebase multiple times to address potential failures caused by the latest commits.~
Updates: I find how to fix them. See my commit 9cf2835. Will test it locally tomorrow!
Finally got it done (nearly)!
I reviewed all the changes file by file this week and found a few issues then fixed them. Now all the local tests (only one failed now) based on my latest changes have passed completely!
As for that latin1 encoding issue, we can add PGCLIENTENCODING=UTF8 when running the test commands.
Screenshots:
PGCLIENTENCODING=UTF8 PGOPTIONS='-c optimizer=on' make -C src/test installcheck-cbdb-parallel
PGCLIENTENCODING=UTF8 PGOPTIONS='-c optimizer=on' make installcheck
PGCLIENTENCODING=UTF8 PGOPTIONS='-c optimizer=off' make installcheck-world
for the failed test, haven't changed the file src/test/regress/expected/uaocs_compaction/stats.out. I'm not sure about it.
Update: The above failed test cannot happen again. I tested this command again on my branch and the latest main branch, another new test failed - see issue #781
My tests
My testing was based on the latest building Docker image:
docker run -it --rm -h cdw --shm-size=30gb apache/incubator-cloudberry:cbdb-build-rocky9-latest
Then pulling my branch:
git clone --branch cleanup-mails-and-names1126 https://github.com/tuhaihe/cloudberrydb.git ~/cloudberry
After that, compiled and installed the Cloudberry following this guide (https://github.com/edespino/cloudberry/blob/rocky9-dev-readme/deploy/build/README-rockylinux9.md).
Lastly, ran the three tests mentioned above. FYI.
Plan
Except for the first two commits in this PR, most of the others are quite similar. I’m planning to squash them down into a single epic commit, and then we can start merging this PR.
FYI: I removed the skip ci token in the PR body. This will allow the CI to run automatically on this PR. It was previously skipping.
FYI: I removed the skip ci token in the PR body. This will allow the CI to run automatically on this PR. It was previously skipping.
Thanks Ed! Now all the CI checks have passed. Do we have any plans to merge this PR?
We are currently cherry-picking commits from the archived Greenplum repository. Since more files with legacy brand names might surface, there’s no urgency to proceed with the PR review at this stage. Once the cherry-picking is complete, I will review all the files and update the PR accordingly.
Will rebase this PR once #925 and #924 are merged.
Updates - 0225, 2025
- Rebase on the main branch.
- All the tests pass in the GitHub workflow and my local machine.
Now ready for approval.
Hey @edespino could you help review this again? I would like to proceed more before our release 2.0. Thanks!