terafo

Results 9 comments of terafo

@LovecraftianHorror can't reproduce on ```master``` branch. Not sure if I should close this issue since it isn't fixed in released version.

Bigger model - generally better results, but uses more ram and slower. If you want fast results - 7B, best results - use whatever fits in your RAM.

You would need GPU with tens of gigs of VRAM and use another fork.

llama.cpp is made only for inference, it doesn't have training functionality. It wouldn't make sense to do that on CPU for model of that size anyways. Meta didn't release LLaMa...

Found some holes in my fix that weren't in original code, dot product still needs fixing.

Actually fixed that, a bit messy, but works with everything I threw on it

Theoretically it can be done in runtime and not in the separate file, but that would take a lot of time, more RAM and would require loading of full set...

I'm making progress towards generic vector support in a separate [branch](https://github.com/terafo/tinygrad/tree/vector), since it requires changing too much stuff that isn't related to hexagon. I'll open a separate draft PR for...

Just a friendly reminder that it's still outdated.