OpenAdapt
OpenAdapt copied to clipboard
Open Source Generative Process Automation (i.e. Generative RPA). AI-First Process Automation with Large ([Language (LLMs) / Action (LAMs) / Multimodal (LMMs)] / Visual Language (VLMs)) Models
How can we enable loss adjusters (and their managers) to automate repetitive tasks? This task involves creating recordings involving one or more relevant applications, and implementing tests for corresponding completions:...
### Feature request How can we incorporate the ideas in https://craftjarvis-jarvis1.github.io/ ? ### Motivation > We introduce JARVIS-1, an open-world agent that can perceive multimodal input (visual observations and human...
### Feature request How can we take advantage of https://github.com/haotian-liu/LLaVA ? https://llava-vl.github.io/ ### Motivation > LLaVA represents a novel end-to-end trained large multimodal model that combines a vision encoder and...
### PR Feedback: 👎 ## Description This PR adds documentation to the README.md file on how to use `pyenv` with `poetry` in the OpenAdapt project. It provides instructions for users...
### Feature request Currently we have: ``` _DEFAULTS = { "FOO": "bar", ... } ``` This requires lots of linter pragmas. This task involves refactoring simply: ``` FOO = "bar"...
### Feature request We want to get completions using https://huggingface.co/NousResearch/Nous-Hermes-13b ### Motivation https://twitter.com/Yampeleg/status/1668337702440165376
### Feature request How can we incorporate https://mbzuai-oryx.github.io/groundingLMM/ ? ### Motivation > Grounding Large Multimodal Model (GLaMM) is an end-to-end trained LMM which provides visual grounding capabilities with the flexibility...
### Feature request How can we incorporate the ideas in https://react-lm.github.io/ ? ### Motivation > In this paper, we explore the use of LLMs to generate both reasoning traces and...
### Feature request We would like to implement https://minigpt-v2.github.io/ as a `CompletionProvider`. See also https://www.reddit.com/r/MachineLearning/comments/17b7iyy/r_minigptv2_large_language_model_as_a_unified/ ### Motivation > After our three-stage training, the experimental results show that MiniGPT-v2 achieves strong...
### Feature request How can we take incorporate the ideas in https://github.com/ddupont808/GPT-4V-Act, e.g. Set-of-Mark prompting (https://arxiv.org/abs/2310.11441 / https://github.com/microsoft/SoM)? See also https://www.reddit.com/r/MachineLearning/comments/17cy0j7/d_p_web_browsing_uibased_ai_agent_gpt4vact/ ### Motivation > GPT-4V-Act serves as an eloquent multimodal...