79669925

Date: 2025-06-18 01:36:53
Score: 1
Natty:
Report link

Try use float16 not bfloat16, it seems that bfloat16 will use more VRAM.

torch_dtype=torch.float16

or try to use xformers to reduce VRAM usage.

self.pip.enable_xformers_memory_efficient_attention()
Reasons:
  • Low length (0.5):
  • Has code block (-0.5):
  • Low reputation (1):
Posted by: YE Chen