adapters icon indicating copy to clipboard operation
adapters copied to clipboard

How to use multiple prefix tuned models with parallel inference.

Open santhoshkolloju opened this issue 2 years ago • 1 comments

Hi Library error when I use muliptle prefix tuned models for parallel inference. What are the adpaters which supports parallel inference and how to know their configuration.

Thanks

santhoshkolloju avatar Jul 17 '22 03:07 santhoshkolloju

Hey, parallel inference (as well as all other adapter composition blocks) are currently only supported with the "original" bottleneck adapter configuration (all adapters using AdapterConfig). Support for other methods such as prefix tuning may be added in the future.

calpt avatar Jul 26 '22 09:07 calpt