When it comes to inference using cloud services, Tesla T4 GPUs are often used, but they are not cheap.
It seems that creating a server room with A4000 workstations or RTX 3060 notes might be more cost-effective, but this could potentially violate Nvidia's terms and conditions.<p>On the other hand, there are cloud services that advertise that using the A4000 for inference is acceptable.
Does this mean that while support from Nvidia might not be available, it is implicitly tolerated by Nvidia?