dalai icon indicating copy to clipboard operation
dalai copied to clipboard

I've done all of the above fixes. and now when I do the prompt I get:

Open RIAZAHAMMED opened this issue 1 year ago • 13 comments

          I've done all of the above fixes. and now when I do the prompt I get:

[2J[m[HWindows PowerShell

Copyright (C) Microsoft Corporation. All rights reserved.

Try the new cross-platform PowerShell https://aka.ms/pscore6]0;C:\WINDOWS\System32\WindowsPowerShell\v1.0\powershell.exe[?25h[25l[6;1H[?25hPS C:\Users\Adam\dalai\alpaca> [?25l[93mC[?25h[m[?25l PS C:\Users\Adam\dalai\alpaca> [93mC:\Users\Adam\dalai\alpaca\build\Release\main[m [90m--seed[m -1 [90m--threads[m [97m4[m [90m--n_predict[m [97m200[m [90m--model[m models/7B/ggml-model-q4_0.bin [90m--top_k[m [97m40[m [90m--top_p[m [97m0.9[m [90m--temp[m [97m0.1[m [90m--repeat_last

_n[m [97m64[m [90m--repeat_penalty[m [97m1.3[m [90m-p[m [36m"where is rome"[m[K

[?25hmain: seed = 1679321497

llama_model_load: loading model from 'models/7B/ggml-model-q4_0.bin' - please wait ...

llama_model_load: ggml ctx size = 6065.34 MB[14X

llama_model_load: memory_size = 2048.00 MB, n_mem = 65536

llama_model_load: loading model part 1/1 from 'models/7B/ggml-model-q4_0.bin'

llama_model_load: [59X.................................... done

[25lllama_model_load: model size = 4017.27 MB / num tensors = 291[113X

[175X

system_info: n_threads = 4 / 4 | AVX = 1 | AVX2 = 1 | AVX512 = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 0 | VSX = 0 |

[175X

main: prompt: 'where is rome'[146X

main: number of tokens in prompt = 5[139X

 1 -> ''[163X

3062 -> 'where'[158X

338 -> ' is'[160X

29871 -> ' '[162X

4871 -> 'rome'[159X

[175X

sampling parameters: temp = 0.100000, top_k = 40, top_p = 0.900000, repeat_last_n = 64, repeat_penalty = 1.300000[62X[29;1H[?25hPS C:\Users\Adam\dalai\alpaca> [93me[?25l[m[92mexit

[?25h

Originally posted by @nigelp in https://github.com/cocktailpeanut/dalai/issues/128#issuecomment-1476310763

I am cracking my head on the wall with this issue.... and I can't seem to get the llama files...

RIAZAHAMMED avatar Mar 20 '23 15:03 RIAZAHAMMED

It seems like you are using PowerShell. Do not use PowerShell.

ghost avatar Mar 20 '23 17:03 ghost

I have the exact same issue, and I'm not using Power Shell at all, just the regular command prompt.

I uninstalled everything and repeated the whole process twice. I always get that same weird result described above

aofalcao avatar Mar 20 '23 19:03 aofalcao

Okay after running through the logs and the programs yargs was missing in the bin these commands sorted it out for me good luck

cd %USERPROFILE%\node_modules\dalai\bin npm install --save yargs

mattcurr-sas avatar Mar 20 '23 21:03 mattcurr-sas

C:\Users&&&&\dalai-main\bin>npx dalai serve [#########.........] | idealTree: timing idealTree Completed in 326ms

Now I am getting this

RIAZAHAMMED avatar Mar 21 '23 00:03 RIAZAHAMMED

Same results here, not using powershell.

image

ImLunaHey avatar Mar 21 '23 01:03 ImLunaHey

I think I've found the issue, even though I'm launching this with cmd it's using PowerShell when it runs the autocomplete. Checking the debug option shows this.

ImLunaHey avatar Mar 21 '23 01:03 ImLunaHey

Actually it might not be powershell that's the issue.

Running the command that debug spat out I get the same result in cmd myself.

C:\Users\xo\dalai\llama>C:\Users\xo\dalai\llama\build\Release\main --seed -1 --threads 4 --n_predict 20 --model models/7B/ggml-model-q4_0.bin --top_k 40 --top_p 0.9 --temp 0.1 --repeat_last_n 64 --repeat_penalty 1.3 -p " a small wizard lives in a forest"
main: seed = 1679363493
llama_model_load: loading model from 'models/7B/ggml-model-q4_0.bin' - please wait ...
llama_model_load: n_vocab = 32000
llama_model_load: n_ctx   = 512
llama_model_load: n_embd  = 4096
llama_model_load: n_mult  = 256
llama_model_load: n_head  = 32
llama_model_load: n_layer = 32
llama_model_load: n_rot   = 128
llama_model_load: f16     = 2
llama_model_load: n_ff    = 11008
llama_model_load: n_parts = 1
llama_model_load: ggml ctx size = 4529.34 MB
llama_model_load: memory_size =   512.00 MB, n_mem = 16384
llama_model_load: loading model part 1/1 from 'models/7B/ggml-model-q4_0.bin'
llama_model_load:  done
llama_model_load: model size =    78.12 MB / num tensors = 1

system_info: n_threads = 4 / 12 | AVX = 1 | AVX2 = 1 | AVX512 = 0 | FMA = 0 | NEON = 0 | ARM_FMA = 0 | F16C = 0 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 0 | VSX = 0 |

main: prompt: '  a small wizard lives in a forest'
main: number of tokens in prompt = 10
     1 -> ''
 29871 -> ' '
   263 -> ' a'
  2319 -> ' small'
   281 -> ' w'
 17909 -> 'izard'
 12080 -> ' lives'
   297 -> ' in'
   263 -> ' a'
 13569 -> ' forest'

sampling parameters: temp = 0.100000, top_k = 40, top_p = 0.900000, repeat_last_n = 64, repeat_penalty = 1.300000


  a small wizard lives in a forest
¶¶↓∟↑↓ ◄ ‼▬

main: mem per token = 14368644 bytes
main:     load time =    30.53 ms
main:   sample time =    17.84 ms
main:  predict time =  5949.02 ms / 205.14 ms per token
main:    total time =  6820.01 ms

ImLunaHey avatar Mar 21 '23 01:03 ImLunaHey

Same results here, not using powershell.

image

Aha.. you got Llama 7B working.. I couldn't even get that in the Dalai... all I get is the Alpaca 7B... the other ones give me a timeout

RIAZAHAMMED avatar Mar 21 '23 01:03 RIAZAHAMMED

I have the same issue. Only alpaca works for me. Weirdly enough the first time I tried it worked. Then I tried installing other models and it never worked again.

carlucio7 avatar Mar 21 '23 03:03 carlucio7

this worked for me: https://github.com/cocktailpeanut/dalai/issues/147#issuecomment-1476831396

SarahDunlap avatar Mar 21 '23 16:03 SarahDunlap

this worked for me: #147 (comment)

Yes, I had found that post as well. I guess it messed up the files when I tried to download the 13b model and the download failed.

Thanks.

carlucio7 avatar Mar 21 '23 16:03 carlucio7

this worked for me: #147 (comment)

Yes. This was the thing. Thank you! I tried the same procedure even on different machines, and that finally solved it. I have no idea how could other windows users get these models running [could it be a different VS version? I'm using VS2022] By the way some people are mentioning that the alpaca model is working. That makes sense because in the build process, for alpaca, the final .exe is correctly called main.exe

aofalcao avatar Mar 21 '23 18:03 aofalcao

this worked for me: #147 (comment)

Yes. This was the thing. Thank you! I tried the same procedure even on different machines, and that finally solved it. I have no idea how could other windows users get these models running [could it be a different VS version? I'm using VS2022] By the way some people are mentioning that the alpaca model is working. That makes sense because in the build process, for alpaca, the final .exe is correctly called main.exe

I don't think anyone got this working on Windows... I got Alpaca working on Zorin Linux on my laptop but the ram on that is only 8GB so llama doesn't load at all

RIAZAHAMMED avatar Mar 22 '23 03:03 RIAZAHAMMED