@@ -66,7 +66,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
6666# model_bytes = persist_torch_model.read_bytes()
6767# input_tensor = torch.randn(2)
6868
69- # expected_device = "cpu"
7069# expected_callback_channel = b"faux_channel_descriptor_bytes"
7170# callback_channel = mli.DragonCommChannel.find(expected_callback_channel)
7271
@@ -77,7 +76,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
7776# request = MessageHandler.build_request(
7877# reply_channel=callback_channel.descriptor,
7978# model=model_bytes,
80- # device=expected_device,
8179# inputs=[message_tensor_input],
8280# outputs=[],
8381# custom_attributes=None,
@@ -86,7 +84,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
8684# msg_bytes = MessageHandler.serialize_request(request)
8785
8886# inference_request = worker.deserialize(msg_bytes)
89- # assert inference_request.device == expected_device
9087# assert inference_request.callback._descriptor == expected_callback_channel
9188
9289
@@ -104,7 +101,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
104101# # input_tensor = torch.randn(2)
105102# # feature_store[input_key] = input_tensor
106103
107- # expected_device = "cpu"
108104# expected_callback_channel = b"faux_channel_descriptor_bytes"
109105# callback_channel = mli.DragonCommChannel.find(expected_callback_channel)
110106
@@ -117,7 +113,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
117113# request = MessageHandler.build_request(
118114# reply_channel=callback_channel.descriptor,
119115# model=message_model_key,
120- # device=expected_device,
121116# inputs=[message_tensor_input_key],
122117# outputs=[message_tensor_output_key],
123118# custom_attributes=None,
@@ -126,7 +121,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
126121# msg_bytes = MessageHandler.serialize_request(request)
127122
128123# inference_request = worker.deserialize(msg_bytes)
129- # assert inference_request.device == expected_device
130124# assert inference_request.callback._descriptor == expected_callback_channel
131125
132126
@@ -147,7 +141,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
147141# # input_tensor = torch.randn(2)
148142# # feature_store[input_key] = input_tensor
149143
150- # expected_device = "cpu"
151144# expected_callback_channel = b"faux_channel_descriptor_bytes"
152145# callback_channel = mli.DragonCommChannel.find(expected_callback_channel)
153146
@@ -160,7 +153,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
160153# request = MessageHandler.build_request(
161154# reply_channel=callback_channel.descriptor,
162155# model=model_bytes,
163- # device=expected_device,
164156# inputs=[message_tensor_input_key],
165157# # outputs=[message_tensor_output_key],
166158# outputs=[],
@@ -170,7 +162,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
170162# msg_bytes = MessageHandler.serialize_request(request)
171163
172164# inference_request = worker.deserialize(msg_bytes)
173- # assert inference_request.device == expected_device
174165# assert inference_request.callback._descriptor == expected_callback_channel
175166
176167
@@ -191,7 +182,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
191182# input_tensor = torch.randn(2)
192183# # feature_store[input_key] = input_tensor
193184
194- # expected_device = "cpu"
195185# expected_callback_channel = b"faux_channel_descriptor_bytes"
196186# callback_channel = mli.DragonCommChannel.find(expected_callback_channel)
197187
@@ -207,7 +197,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
207197# request = MessageHandler.build_request(
208198# reply_channel=callback_channel.descriptor,
209199# model=model_bytes,
210- # device=expected_device,
211200# inputs=[message_tensor_input],
212201# # outputs=[message_tensor_output_key],
213202# outputs=[message_tensor_output_key],
@@ -217,7 +206,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
217206# msg_bytes = MessageHandler.serialize_request(request)
218207
219208# inference_request = worker.deserialize(msg_bytes)
220- # assert inference_request.device == expected_device
221209# assert inference_request.callback._descriptor == expected_callback_channel
222210
223211
@@ -238,7 +226,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
238226# input_tensor = torch.randn(2)
239227# # feature_store[input_key] = input_tensor
240228
241- # expected_device = "cpu"
242229# expected_callback_channel = b"faux_channel_descriptor_bytes"
243230# callback_channel = mli.DragonCommChannel.find(expected_callback_channel)
244231
@@ -254,7 +241,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
254241# request = MessageHandler.build_request(
255242# reply_channel=callback_channel.descriptor,
256243# model=message_model_key,
257- # device=expected_device,
258244# inputs=[message_tensor_input],
259245# # outputs=[message_tensor_output_key],
260246# outputs=[],
@@ -264,7 +250,6 @@ def persist_torch_model(test_dir: str) -> pathlib.Path:
264250# msg_bytes = MessageHandler.serialize_request(request)
265251
266252# inference_request = worker.deserialize(msg_bytes)
267- # assert inference_request.device == expected_device
268253# assert inference_request.callback._descriptor == expected_callback_channel
269254
270255
0 commit comments