• @[email protected]OP
    link
    fedilink
    English
    424 months ago

    That’s a very toxic attitude.

    Inference is in principle the process of generation of the AI response. So when you run locally and LLM you are using your GPU only for inference.