FFAMax
FFAMax
Hello, Team! I have RTX 4090 and GTX 1080 Ti, run on both ``` CUDA_VISIBLE_DEVICES=0,1 python3 examples/llama3.py --download_model --shard 2 --size 8B # data loaded on both GPU Hit Ctrl-C...
``` (.venv) user@b250:~/tinygrad$ AMD=1 VISIBLE_DEVICES=0 python3 examples/llama3.py --download_model seed = 1730624168 0%| | 0/292 [00:00
Should we add it to https://github.com/tinygrad/tinygrad/blob/master/docs/env_vars.md ?
Hello, Team. Tried with CLANG - it worked, but with GPU just no progress: ``` python3 examples/llama3.py --host 0.0.0.0 --port 8080 --model /home/user/.cache/tinygrad/downloads/llama3-1b-instruct/Llama-3.2-1B-Instruct-Q6_K.gguf seed = 1730621540 ``` I see on...
For now CPU, AMD, Nvidia can be used as OpenCL, but limited due no option to pick the right ones. For example combination of AMD+Nvidia can work via OpenCL Example...
Hey Alex. We have wps* and num8* dicts, they cover entire range 0000000-99999999, I didn't found missed numbers in the dicts, but recently found several submissions was coming regularly, like...
I’ve noticed some inefficiencies in hashcat when working with powerful GPUs or multi-GPU setups: - **Dictionary Loading Overhead**: When using large or numerous dictionaries, a significant amount of time is...
Hi, Alex. Noticed, that for `WPA*01` to proceed one dict may take 10 seconds, while for `WPA*02` it may take 20-35 with the same dict. The idea is: If we...
FYI ``` # cat wpskey0.txt | grep -E '^[0-9]{8,8}$' | wc -l 1000000 # cat *.txt | grep -E '^[0-9]{8,8}$' | wc -l 171998 # cat *.txt | strings |...
Hey Alex, to optimize how dict's data loaded to GPU, should be great to give more data in bulk to reduce moving data back and forth and reduce time lost...