diff --git a/inference.py b/inference.py index 0ea7b65..a550c20 100644 --- a/inference.py +++ b/inference.py @@ -58,22 +58,12 @@ class BiMVFIModel: img0 = frame0.to(device) img1 = frame1.to(device) - _, _, h, w = img0.shape - if h >= 2160: - pyr_level = 7 - elif h >= 1080: - pyr_level = 6 - elif h >= 540: - pyr_level = 5 - else: - pyr_level = self.pyr_level - time_step_tensor = torch.tensor([time_step], device=device).view(1, 1, 1, 1) result_dict = self.model( img0=img0, img1=img1, time_step=time_step_tensor, - pyr_level=pyr_level, + pyr_level=self.pyr_level, ) interp = result_dict["imgt_pred"] diff --git a/nodes.py b/nodes.py index da5d233..4f325a9 100644 --- a/nodes.py +++ b/nodes.py @@ -59,7 +59,7 @@ class LoadBIMVFIModel: }), "pyr_level": ("INT", { "default": 3, "min": 3, "max": 7, "step": 1, - "tooltip": "Pyramid levels for coarse-to-fine processing. Only used for <540p. Higher res auto-selects: 540p=5, 1080p=6, 4K=7. More levels = captures larger motion but slower.", + "tooltip": "Pyramid levels for coarse-to-fine processing. More levels = captures larger motion but slower. Recommended: 3-5 for <540p, 5-6 for 1080p, 6-7 for 4K.", }), } }