@@ -32,13 +32,13 @@ class TritonPythonModel:
3232 def execute (self , requests ):
3333 responses = []
3434 for request in requests :
35- inp = pb_utils .get_input_tensor_by_name (request , "model_2_input_string " )
36- inp2 = pb_utils .get_input_tensor_by_name (request , "model_2_input_UINT8_array " )
37- inp3 = pb_utils .get_input_tensor_by_name (request , "model_2_input_INT8_array " )
38- inp4 = pb_utils .get_input_tensor_by_name (request , "model_2_input_FP32_image " )
39- inp5 = pb_utils .get_input_tensor_by_name (request , "model_2_input_bool " )
35+ inp = pb_utils .get_input_tensor_by_name (request , "model_1_input_string " )
36+ inp2 = pb_utils .get_input_tensor_by_name (request , "model_1_input_UINT8_array " )
37+ inp3 = pb_utils .get_input_tensor_by_name (request , "model_1_input_INT8_array " )
38+ inp4 = pb_utils .get_input_tensor_by_name (request , "model_1_input_FP32_image " )
39+ inp5 = pb_utils .get_input_tensor_by_name (request , "model_1_input_bool " )
4040
41- print ("Model 2 received" , flush = True )
41+ print ("Model 1 received" , flush = True )
4242 print (inp .as_numpy (), flush = True )
4343 print (inp2 .as_numpy (), flush = True )
4444 print (inp3 .as_numpy (), flush = True )
@@ -47,23 +47,23 @@ def execute(self, requests):
4747
4848 inference_response = pb_utils .InferenceResponse (output_tensors = [
4949 pb_utils .Tensor (
50- "model_2_output_string " ,
50+ "model_1_output_string " ,
5151 inp .as_numpy (),
5252 ),
5353 pb_utils .Tensor (
54- "model_2_output_UINT8_array " ,
54+ "model_1_output_UINT8_array " ,
5555 inp2 .as_numpy (),
5656 ),
5757 pb_utils .Tensor (
58- "model_2_output_INT8_array " ,
58+ "model_1_output_INT8_array " ,
5959 inp3 .as_numpy (),
6060 ),
6161 pb_utils .Tensor (
62- "model_2_output_FP32_image " ,
62+ "model_1_output_FP32_image " ,
6363 inp4 .as_numpy (),
6464 ),
6565 pb_utils .Tensor (
66- "model_2_output_bool " ,
66+ "model_1_output_bool " ,
6767 inp5 .as_numpy (),
6868 )
6969 ])
0 commit comments