manifest icon indicating copy to clipboard operation
manifest copied to clipboard

Prompt programming with FMs.

Results 7 manifest issues
Sort by recently updated
recently updated
newest added

## Description of the bug can not load model with half precision. And haven't figured out how to transfer model to CPU or GPU? ## To Reproduce run model gpt-j-6B...

```python def get_generation_url(self) -> str: """Get generation URL.""" engine = getattr(self, "engine") deployment_name = AZURE_DEPLOYMENT_NAME_MAPPING.get(engine, engine) return ( self.host + "/openai/deployments/" + deployment_name + "/chat/completions?api-version=2023-05-15" ) ``` The api-version is...

I was wondering if there was a good way to specify which GPUs deepspeed, accelerate, etc. should utilize. Right now I've been doing something like: ``` CUDA_VISIBLE_DEVICES=3,4,5 python -m manifest.api.app...

## Description of the bug When loading a local model using this command : ``` python3 -m manifest.api.app \ --model_type huggingface \ --model_name_or_path /workspace/models/minotaur-15b \ --device 0 \ --model_generation_type text-generation...

It seems like HuggingFace is only supported for local models. It would be nice to support calling model on the Hub with the InferenceAPI. I think this is relatively straightforward...

- add support for OpenRouter ([https://openrouter.ai/](https://openrouter.ai/)) - add `gpt-4o` and `gpt-4o-mini` to `OpenAIChatClient`