Open-Assistant icon indicating copy to clipboard operation
Open-Assistant copied to clipboard

Inference Documentation

Open alando46 opened this issue 2 years ago • 4 comments

This is a mostly done (although not totally complete) PR with a technical overview of the inference architecture. I'm looking forward to high level feedback (general layout, flow of documentation) or specific suggestions (I'm sure I made some errors or missed some details.) I will try to wrap up the final section soon.

See related discussion on the issue: https://github.com/LAION-AI/Open-Assistant/issues/1473#issuecomment-1537083924

alando46 avatar May 10 '23 15:05 alando46

:x: pre-commit failed. Please run pre-commit run --all-files locally and commit the changes. Find more information in the repository's CONTRIBUTING.md

github-actions[bot] avatar May 10 '23 15:05 github-actions[bot]

Nice work!

It looks like you went for a quite detailed explanation of the code. As a suggestion, I think that a diagram of how all the components are connected, and a brief description of each one's goal, would help.

Also, I think it's important to try to use always the same word for the different components. A couple of examples of things I've seen which I think they are the same, but I'm not sure:

  • In the architecture document, it talks about a text client, that it's referred to as Text Client, text-client, and a REPL client in the inference document.

  • In the architecture document, it talks about a FastAPI webserver, that it's referred to as OA server in the mermaid diagram and OA Inference Server in the inference document.

Anyway, these are minor things. Thank you again for the work!

GuilleHoardings avatar May 11 '23 18:05 GuilleHoardings

thanks a lot for this :)

yk avatar May 11 '23 20:05 yk

:x: pre-commit failed. Please run pre-commit run --all-files locally and commit the changes. Find more information in the repository's CONTRIBUTING.md

github-actions[bot] avatar May 14 '23 19:05 github-actions[bot]

@alando46 this PR is currently still in draft status. Is it ready for review?

andreaskoepf avatar May 29 '23 14:05 andreaskoepf

@andreaskoepf, would it be possible to raise this as-is (w/o last section)? I keep trying to get back to this but things have gotten unexpectedly busy.

I'd propose raising this as-is, and then down the road I or someone else can update the final section. The plugin addition (very cool) will likely change some things on the server/worker inference system so might make sense to update the inference documentation once again anyways once that api stabilizes a bit.

Thoughts?

alando46 avatar Jun 12 '23 01:06 alando46