As a newcomer to the area and a noob in most things ML, one think that I still have a hard time evaluating is the minimal type of hardware to run these LLMs either locally or at scale.
I often see a lot of detail on how the models were trained but do not see much information on what is needed to actually run it. I found some information on /r/LocalLLaMA/ but it still very sparse.
Anyone has tips on how to figure this out besides actually running them (and needing to spend $$)?
I often see a lot of detail on how the models were trained but do not see much information on what is needed to actually run it. I found some information on /r/LocalLLaMA/ but it still very sparse.
Anyone has tips on how to figure this out besides actually running them (and needing to spend $$)?