• wonderingwanderer@sopuli.xyz
      link
      fedilink
      arrow-up
      2
      ·
      18 hours ago

      Yeah, more people should know about it. There’s really no reason to pay for an API for these giant 200 billion parameter commercial models sucking up intense resources in data centers.

      A quantized 24-32 billion parameter model works just fine, can be self-hosted, and can be fine-tuned on ethically-sourced datasets to suit your specific purposes. Bonus points for running your home lab on solar power.

      Not only are the commercial models trained on stolen data, but they’re so generalized that they’re basically worthless for any specialized purpose. A 12 billion parameter model with Retrieval-Augmented Generation is far less likely to hallucinate.