You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+11-2Lines changed: 11 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,9 +2,18 @@
2
2
3
3
For users experiencing the "Tensor in" & "Tensor out" approach to Deep Learning Inference, getting started with Triton can lead to many questions. The goal of this repository is to familiarize users with Triton's features and provide guides and examples to ease migration.
4
4
5
-
|[Conceptual Guide](Conceptual_Guide/README.md)|[Quick Overview](https://www.youtube.com/watch?v=NQDtfSi5QF4)|[Documentation](https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html)|[Just help me deploy](Quick_Deploy/README.md)|
5
+
|[Conceptual Guide](Conceptual_Guide/README.md)|[Quick Overview](https://www.youtube.com/watch?v=NQDtfSi5QF4)|[Documentation](https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html)|[Help me deploy](Quick_Deploy/README.md)|
The Triton Inference Server GitHub organization contains multiple repositories housing different features of the Triton Inference Server. The following is not a complete description of all the repositories, but just a simple guide to build intuitive understanding.
11
+
12
+
*[Server](https://github.com/triton-inference-server/server) is the main Triton Inference Server Repository.
13
+
*[Client](https://github.com/triton-inference-server/client) contains the libraries and examples needed to create Triton Clients
14
+
*[Backend](https://github.com/triton-inference-server/backend) contains the core scripts and utilities to build a new Triton Backend. Any repository containing the word "backend" is either a framework backend or an example for how to create a backend.
15
+
* Tools like [Model Analyzer](https://github.com/triton-inference-server/model_analyzer) and [Model Navigator](https://github.com/triton-inference-server/model_navigator) provide the tooling to either measure performance, or to simplify model acceleration.
16
+
17
+
## Adding Requests
18
+
19
+
Open an issue and specify details for adding a request for an example. Want to make a contribution? Open a pull request and tag an Admin.
0 commit comments