from concrete.compiler import check_gpu_available,
Can we use GPU for parallel batched inference on encrypted inputs?
Concrete ML models will executed some layers with batched GPU computation. However, the batching is done on layer breadth and Concrete ML will not batch multiple examples like PyTorch.
1 Like