transformers icon indicating copy to clipboard operation
transformers copied to clipboard

[WIP] Add H3

Open NielsRogge opened this issue 1 year ago • 1 comments

What does this PR do?

This PR adds the H3 model by Hazy Research (Stanford University).

I've removed the Flash Attention dependency, and main author @DanFu09 has removed the einops dependency (🙏 ).

I've kept an optional soft dependency on pykeops, to allow for speedups. The model runs fine if the user doesn't have this library installed.

NielsRogge avatar Mar 20 '23 08:03 NielsRogge

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint.

This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.

Please note that issues that do not follow the contributing guidelines are likely to be ignored.

github-actions[bot] avatar Apr 19 '23 15:04 github-actions[bot]

Hey @NielsRogge ✋,

Nice work! Do you know if this model would be integrated in the near future inside HuggingFace? Was this PR staled given complexities with custom ops?

Could you give an overview of the missing steps needed in this PR to have a functional H3 model integrated into HF? 🙏 Thanks for your work! 🙌

gaceladri avatar May 26 '23 16:05 gaceladri

Hi @gaceladri the PR is actually totally ready, the only thing that needs to done is perhaps make this function more like the other Attention classes in the library (like this class).

NielsRogge avatar May 27 '23 08:05 NielsRogge

This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.

Please note that issues that do not follow the contributing guidelines are likely to be ignored.

github-actions[bot] avatar Dec 29 '23 08:12 github-actions[bot]