Skip to content

Commit

Permalink
Update full_inference.py
Browse files Browse the repository at this point in the history
  • Loading branch information
ShiromiyaG committed Sep 6, 2024
1 parent dac763f commit ab866ea
Showing 1 changed file with 3 additions and 2 deletions.
5 changes: 3 additions & 2 deletions tabs/full_inference.py
Original file line number Diff line number Diff line change
Expand Up @@ -135,7 +135,7 @@ def max_vram_gpu(gpu):
if torch.cuda.is_available():
gpu_properties = torch.cuda.get_device_properties(gpu)
total_memory_gb = round(gpu_properties.total_memory / 1024 / 1024 / 1024)
return total_memory_gb - 2
return total_memory_gb / 2
else:
return "0"

Expand Down Expand Up @@ -388,9 +388,10 @@ def full_inference_tab():
batch_size = gr.Slider(
minimum=1,
maximum=24,
step=1,
label=i18n("Batch Size"),
info=i18n("Set the batch size for the separation."),
value=max_vram_gpu(0),
value=2,
interactive=True,
)
vocal_model = gr.Dropdown(
Expand Down

0 comments on commit ab866ea

Please sign in to comment.