![](https://sh.itjust.works/pictrs/image/eff0bc9f-29b4-4a29-b078-b880ee1a09f3.webp)
![](https://fry.gs/pictrs/image/c6832070-8625-4688-b9e5-5d519541e092.png)
Exactly! If you only want to use a Large Language Model (LLM) to run your own local chatbot, then using a quantized version will dramatically improve speed and performance. It also allows consumer hardware to run larger models which would otherwise be prohibitively resource intensive.
Targets: the Dirty Dozen