|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import torch |
|
import math |
|
|
|
|
|
|
|
bs_search_table = [ |
|
|
|
{"res": 768, "total_vram": 79, "bs": 35, "dtype": torch.float32}, |
|
{"res": 1024, "total_vram": 79, "bs": 20, "dtype": torch.float32}, |
|
|
|
{"res": 768, "total_vram": 39, "bs": 15, "dtype": torch.float32}, |
|
{"res": 1024, "total_vram": 39, "bs": 8, "dtype": torch.float32}, |
|
{"res": 768, "total_vram": 39, "bs": 30, "dtype": torch.float16}, |
|
{"res": 1024, "total_vram": 39, "bs": 15, "dtype": torch.float16}, |
|
|
|
{"res": 512, "total_vram": 23, "bs": 20, "dtype": torch.float32}, |
|
{"res": 768, "total_vram": 23, "bs": 7, "dtype": torch.float32}, |
|
{"res": 1024, "total_vram": 23, "bs": 3, "dtype": torch.float32}, |
|
{"res": 512, "total_vram": 23, "bs": 40, "dtype": torch.float16}, |
|
{"res": 768, "total_vram": 23, "bs": 18, "dtype": torch.float16}, |
|
{"res": 1024, "total_vram": 23, "bs": 10, "dtype": torch.float16}, |
|
|
|
{"res": 512, "total_vram": 10, "bs": 5, "dtype": torch.float32}, |
|
{"res": 768, "total_vram": 10, "bs": 2, "dtype": torch.float32}, |
|
{"res": 512, "total_vram": 10, "bs": 10, "dtype": torch.float16}, |
|
{"res": 768, "total_vram": 10, "bs": 5, "dtype": torch.float16}, |
|
{"res": 1024, "total_vram": 10, "bs": 3, "dtype": torch.float16}, |
|
] |
|
|
|
|
|
def find_batch_size(ensemble_size: int, input_res: int, dtype: torch.dtype) -> int: |
|
""" |
|
Automatically search for suitable operating batch size. |
|
|
|
Args: |
|
ensemble_size (`int`): |
|
Number of predictions to be ensembled. |
|
input_res (`int`): |
|
Operating resolution of the input image. |
|
|
|
Returns: |
|
`int`: Operating batch size. |
|
""" |
|
if not torch.cuda.is_available(): |
|
return 1 |
|
|
|
total_vram = torch.cuda.mem_get_info()[1] / 1024.0**3 |
|
filtered_bs_search_table = [s for s in bs_search_table if s["dtype"] == dtype] |
|
for settings in sorted( |
|
filtered_bs_search_table, |
|
key=lambda k: (k["res"], -k["total_vram"]), |
|
): |
|
if input_res <= settings["res"] and total_vram >= settings["total_vram"]: |
|
bs = settings["bs"] |
|
if bs > ensemble_size: |
|
bs = ensemble_size |
|
elif bs > math.ceil(ensemble_size / 2) and bs < ensemble_size: |
|
bs = math.ceil(ensemble_size / 2) |
|
return bs |
|
|
|
return 1 |
|
|