Skip to content

yuhodots/triton

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Triton

An Example for Python BackEnd

You can start triton server and client containers simutaneouly with docker-compose

docker compose up

You can pass a request with interactive mode after attaching the client container.

curl -X POST tritonserver:8000/v2/models/falcon7b/infer -d '{"inputs": [{"name":"text_input","datatype":"BYTES","shape":[1],"data":["How can you be"]}]}'

Then you can get a response from triton server!

{
    "model_name": "falcon7b",
    "model_version": "1",
    "outputs": [
        {
            "name": "text_output",
            "datatype": "BYTES",
            "shape": [1],
            "data": ["How can you be sure that you are getting the best deal on your car"]
        }
    ]
}

If you want to pass a request without creat a client container, plase add network_mode: host in docker-compose.yaml.

PyTorch BackEnd

If you want to deploy your own model with Pytorch BackEnd, refer this architecture.

model-registry
|
+-- <model_name>
    |
    +-- config.pbtxt
    +-- 1
        |
        +-- model.pt

Warning

You should not have empty directories in your Triton model repository.

Refer this issue: triton-inference-server/server#5786 (comment)

Reference

About

NVIDIA Triton Example Codes

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published