File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed
full_gpu_inference_pipeline Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Original file line number Diff line number Diff line change @@ -28,7 +28,7 @@ Before starting, I highly recommand you to read the the following two links to g
28
28
## Prepare the model repository
29
29
The full pipeline is as below:
30
30
31
- <img src =" https://github.com/Project-MONAI/tutorials/raw/master /full_gpu_inference_pipeline/pics/Picture3.png " >
31
+ <img src =" https://github.com/Project-MONAI/tutorials/raw/main /full_gpu_inference_pipeline/pics/Picture3.png " >
32
32
33
33
### Prepare the model repository file directories
34
34
The Triton model repository of the experiment can be fast set up by:
@@ -176,9 +176,9 @@ Since 3D medical images are generally big, the overhead brought by protocols can
176
176
Note that all the processes (pre/post and AI inference) are on GPU.
177
177
From the result, we can come to a conclusion that using shared memory will greatly reduce the latency when data transfer is huge.
178
178
179
- ![ ] ( https://github.com/Project-MONAI/tutorials/raw/master /full_gpu_inference_pipeline/pics/Picture2.png )
179
+ ![ ] ( https://github.com/Project-MONAI/tutorials/raw/main /full_gpu_inference_pipeline/pics/Picture2.png )
180
180
181
181
### Pre/Post-processing on GPU vs. CPU
182
182
After doing pre and post-processing on GPU, we can get a 12x speedup for the full pipeline.
183
183
184
- ![ ] ( https://github.com/Project-MONAI/tutorials/raw/master /full_gpu_inference_pipeline/pics/Picture1.png )
184
+ ![ ] ( https://github.com/Project-MONAI/tutorials/raw/main /full_gpu_inference_pipeline/pics/Picture1.png )
You can’t perform that action at this time.
0 commit comments