You probably want 48gb of vram or more to run the good stuff. I recommend renting GPU time instead of using your own hardware, via AWS or other vendors - runpod.io is pretty good.
You probably want 48gb of vram or more to run the good stuff. I recommend renting GPU time instead of using your own hardware, via AWS or other vendors - runpod.io is pretty good.
Imo it’s worthwhile to just run the biggest model available and rent expensive GPU time. It still amounts to very little overall and you get much better results. Project dependent of course
Uncensored models are so much better, too. chatGPT is like one of those plastic children’s toy hammers vs real models are titanium hammers
For anyone doing a serious project, it’s much more cost effective to rent a node and run your own models on it. You can spin them up and down as needed, cache often-used queries, etc.
I loved this phone! I had a jelly 2. It didn’t work with my train ticket app, so I had to switch to something bigger. Great conversation piece too.
Waiting on Brighter Shores to come out!