@@ -31,7 +31,7 @@ def get_unet(args):
3131 return get_shark_model (args , bucket , model_name , iree_flags )
3232 else :
3333 bucket = "gs://shark_tank/prashant_nod"
34- model_name = "unet_18nov_fp16 "
34+ model_name = "unet_22nov_fp16 "
3535 iree_flags += [
3636 "--iree-flow-enable-padding-linalg-ops" ,
3737 "--iree-flow-linalg-ops-padding-size=32" ,
@@ -44,7 +44,7 @@ def get_unet(args):
4444 # Tuned model is not present for `fp32` case.
4545 if args .precision == "fp32" :
4646 bucket = "gs://shark_tank/prashant_nod"
47- model_name = "unet_18nov_fp32 "
47+ model_name = "unet_22nov_fp32 "
4848 iree_flags += [
4949 "--iree-flow-enable-conv-nchw-to-nhwc-transform" ,
5050 "--iree-flow-enable-padding-linalg-ops" ,
@@ -77,7 +77,7 @@ def get_vae(args):
7777 )
7878 if args .precision in ["fp16" , "int8" ]:
7979 bucket = "gs://shark_tank/prashant_nod"
80- model_name = "vae_18nov_fp16 "
80+ model_name = "vae_22nov_fp16 "
8181 iree_flags += [
8282 "--iree-flow-enable-conv-nchw-to-nhwc-transform" ,
8383 "--iree-flow-enable-padding-linalg-ops" ,
@@ -89,7 +89,7 @@ def get_vae(args):
8989
9090 if args .precision == "fp32" :
9191 bucket = "gs://shark_tank/prashant_nod"
92- model_name = "vae_18nov_fp32 "
92+ model_name = "vae_22nov_fp32 "
9393 iree_flags += [
9494 "--iree-flow-enable-conv-nchw-to-nhwc-transform" ,
9595 "--iree-flow-enable-padding-linalg-ops" ,
0 commit comments