lightseq icon indicating copy to clipboard operation
lightseq copied to clipboard

Can i use lightseq to speed up the model of fairseq Transformer Decoder ?

Open ismymajia opened this issue 4 years ago • 3 comments

Can i use lightseq to speed up the model of fairseq Transformer Decoder ?

I already export the Transformer Decoder language model trained by fairseq , now i want to speed up the model by light seq c++ api , how should i do ? Have any c++ demo? @Taka152

ismymajia avatar Aug 12 '21 04:08 ismymajia

You can check this example https://github.com/bytedance/lightseq/blob/master/examples/inference/cpp/gptlm_example.cc.cu to use in cpp. If you want to serve in a more powerful server, check here. To use the triton server, you may need to do exploration, we haven't provided an end2end example for now

Taka152 avatar Aug 12 '21 08:08 Taka152

this example only have code. What is the model like ? What is the difference between your model with my model? Can u give me a commpletly run example ?@Taka152

ismymajia avatar Aug 13 '21 03:08 ismymajia

./v1.0.0_libs/transformer_example.fp32 ./v0.0.1_gptlm.pkg/gpt.pb ./v0.0.1_gptlm.pkg/test_case Segmentation fault (core dumped)

i test your example, but error occurs.

ismymajia avatar Aug 16 '21 06:08 ismymajia