You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
cd torchpipe/ && git submodule update --init --recursive
38
38
```
39
39
@@ -144,7 +144,7 @@ For more examples, see [Showcase](./showcase/showcase.mdx).
144
144
145
145
## Customizing Dockerfile {#selfdocker}
146
146
147
-
Refer to the [example Dockerfile](https://g.hz.netease.com/deploy/torchpipe/-/blob/master/docker/torchpipe.base). After downloading TensorRT and OpenCV in advance, you can compile the corresponding base image.
147
+
Refer to the [example Dockerfile](https://github.com/torchpipe/torchpipe/-/blob/master/docker/torchpipe.base). After downloading TensorRT and OpenCV in advance, you can compile the corresponding base image.
148
148
```
149
149
# put TensorRT-8.6.1.6.Linux.x86_64-gnu.cuda-11.8.tar.gz into thirdparty/
In the example provided [here](https://g.hz.netease.com/deploy/torchpipe/-/blob/master/examples/yolox/yolox_multithreads_test.py), we use ten clients, each requesting different amounts of data per request, ranging from 1 to 10. We validate the consistency of the results in this case.
395
+
In the example provided [here](https://github.com/torchpipe/torchpipe/-/blob/master/examples/yolox/yolox_multithreads_test.py), we use ten clients, each requesting different amounts of data per request, ranging from 1 to 10. We validate the consistency of the results in this case.
396
396
397
397
Typically, users can iterate through all the data in a directory and repeatedly send requests to verify the stability and consistency of the results.
Copy file name to clipboardExpand all lines: docs/quick_start_new_user.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,7 +7,7 @@ type: explainer
7
7
8
8
# Trial in 30mins(new users)
9
9
10
-
TorchPipe is a multi-instance pipeline parallel library that provides a seamless integration between lower-level acceleration libraries (such as TensorRT and OpenCV) and RPC frameworks. It guarantees high service throughput while meeting latency requirements. This document is mainly for new users, that is, users who are in the introductory stage of acceleration-related theoretical knowledge, know some python grammar, and can read simple codes. This content mainly includes the use of torchpipe for accelerating service deployment, complemented by performance and effect comparisons. The complete code of this document can be found at [resnet50_thrift](https://g.hz.netease.com/deploy/torchpipe/-/blob/develop/examples/resnet50_thrift/)。
10
+
TorchPipe is a multi-instance pipeline parallel library that provides a seamless integration between lower-level acceleration libraries (such as TensorRT and OpenCV) and RPC frameworks. It guarantees high service throughput while meeting latency requirements. This document is mainly for new users, that is, users who are in the introductory stage of acceleration-related theoretical knowledge, know some python grammar, and can read simple codes. This content mainly includes the use of torchpipe for accelerating service deployment, complemented by performance and effect comparisons. The complete code of this document can be found at [resnet50_thrift](https://github.com/torchpipe/torchpipe/-/blob/develop/examples/resnet50_thrift/)。
The overall online service deployment can be found at [main_trt.py](https://g.hz.netease.com/deploy/torchpipe/-/blob/develop/examples/resnet50_thrift/main_trt.py)
87
+
The overall online service deployment can be found at [main_trt.py](https://github.com/torchpipe/torchpipe/-/blob/develop/examples/resnet50_thrift/main_trt.py)
88
88
89
89
:::tip
90
90
Since TensorRT is not thread-safe, when using this method for model acceleration, it is necessary to handle locking (with self.lock:) during the service deployment process.
@@ -104,7 +104,7 @@ From the above process, it's clear that when accelerating a single model, the fo
104
104
105
105

106
106
107
-
We've made adjustments to the deployment of our service using TorchPipe.The overall online service deployment can be found at [main_torchpipe.py](https://g.hz.netease.com/deploy/torchpipe/-/blob/develop/examples/resnet50_thrift/main_torchpipe.py).
107
+
We've made adjustments to the deployment of our service using TorchPipe.The overall online service deployment can be found at [main_torchpipe.py](https://github.com/torchpipe/torchpipe/-/blob/develop/examples/resnet50_thrift/main_torchpipe.py).
The specific test code can be found at [client_qps.py](https://g.hz.netease.com/deploy/torchpipe/-/blob/develop/examples/resnet50_thrift/client_qps.py)
222
+
The specific test code can be found at [client_qps.py](https://github.com/torchpipe/torchpipe/-/blob/develop/examples/resnet50_thrift/client_qps.py)
223
223
224
224
With the same Thrift service interface, testing on a machine with NIDIA-3080 GPU, 36-core CPU, and concurrency of 10, we have the following results:
0 commit comments