GPT-2 (small): The 117M parameter model is lightweight and can run on low-power hardware, though its accuracy is lower than Llama. However, it’s capable of handling many text generation tasks decently when finetuned.
DistilGPT-2: A distilled version of GPT-2, it’s about 60% faster while keeping around 95% of the accuracy. This may still be limited, but it’s very efficient on low-power devices.
You are viewing a single comment's thread from: