feat: Add OASEstimator class with oneDAL support and corresponding tests
- Implemented OASEstimator class in oas_estimator.py inheriting from EmpiricalCovariance.
- Added methods _sklearn_fit, _onedal_fit, score, and _onedal_score to support both scikit-learn and oneDAL backends.
- Integrated dispatch mechanism to switch between scikit-learn and oneDAL implementations.
- Added tests for OASEstimator in est_oas_estimator.py to verify fit and score functionalities.
Description
Add a comprehensive description of proposed changes
List associated issue number(s) if exist(s): #6 (for example)
Documentation PR (if needed): #1340 (for example)
Benchmarks PR (if needed): https://github.com/IntelPython/scikit-learn_bench/pull/155 (for example)
PR should start as a draft, then move to ready for review state after CI is passed and all applicable checkboxes are closed. This approach ensures that reviewers don't spend extra time asking for regular requirements.
You can remove a checkbox as not applicable only if it doesn't relate to this PR in any way. For example, PR with docs update doesn't require checkboxes for performance while PR with any change in actual code should have checkboxes and justify how this code change is expected to affect performance (or justification should be self-evident).
Checklist to comply with before moving PR from draft:
PR completeness and readability
- [ ] I have reviewed my changes thoroughly before submitting this pull request.
- [ ] I have commented my code, particularly in hard-to-understand areas.
- [ ] I have updated the documentation to reflect the changes or created a separate PR with update and provided its number in the description, if necessary.
- [ ] Git commit message contains an appropriate signed-off-by string (see CONTRIBUTING.md for details).
- [ ] I have added a respective label(s) to PR if I have a permission for that.
- [ ] I have resolved any merge conflicts that might occur with the base branch.
Testing
- [ ] I have run it locally and tested the changes extensively.
- [ ] All CI jobs are green or I have provided justification why they aren't.
- [ ] I have extended testing suite if new functionality was introduced in this PR.
Performance
- [ ] I have measured performance for affected algorithms using scikit-learn_bench and provided at least summary table with measured data, if performance change is expected.
- [ ] I have provided justification why performance has changed or why changes are not expected.
- [ ] I have provided justification why quality metrics have changed or why changes are not expected.
- [ ] I have extended benchmarking suite and provided corresponding scikit-learn_bench PR if new measurable functionality was introduced in this PR.
It looks like the sklearn equivalent is named just 'OAS': https://scikit-learn.org/stable/modules/generated/sklearn.covariance.OAS.html
It is also exposed as a function that returns arrays: https://scikit-learn.org/stable/modules/generated/oas-function.html
Would be ideal to follow those same interfaces so that patching could be used. As a reference, here's the other sklearn function that gets 'patched' by this library: https://github.com/uxlfoundation/scikit-learn-intelex/blob/main/sklearnex/model_selection/split.py
ref https://github.com/uxlfoundation/scikit-learn-intelex/issues/2305
Hello @marcialouis! Thank you so much for your contribution. Just to add to the great suggestions by @david-cortes-intel , you can find what additional code is necessary here: https://github.com/scikit-learn/scikit-learn/blob/main/sklearn/covariance/_shrunk_covariance.py#L46 . Essentially, a small amount of additional python commands can be used on top of our implementation of EmpericialCovariance.
Hi @marcialouis ! Do you feel like addressing comments?
Closing as it can't be merged in current state