|
|
@@ -48,7 +48,7 @@ logger.add(sys.stderr, format=logger_format)
|
|
|
@lru_cache(maxsize=1)
|
|
|
def get_model(
|
|
|
config_name: str = "firefly_gan_vq",
|
|
|
- checkpoint_path: str = "checkpoints/fish-speech-1.4/firefly-gan-vq-fsq-8x1024-21hz-generator.pth",
|
|
|
+ checkpoint_path: str = "checkpoints/fish-speech-1.5/firefly-gan-vq-fsq-8x1024-21hz-generator.pth",
|
|
|
device: str | torch.device = "cuda",
|
|
|
):
|
|
|
with initialize(version_base="1.3", config_path="../../fish_speech/configs"):
|
|
|
@@ -139,7 +139,7 @@ def process_batch(files: list[Path], model) -> float:
|
|
|
@click.option("--config-name", default="firefly_gan_vq")
|
|
|
@click.option(
|
|
|
"--checkpoint-path",
|
|
|
- default="checkpoints/fish-speech-1.4/firefly-gan-vq-fsq-8x1024-21hz-generator.pth",
|
|
|
+ default="checkpoints/fish-speech-1.5/firefly-gan-vq-fsq-8x1024-21hz-generator.pth",
|
|
|
)
|
|
|
@click.option("--batch-size", default=64)
|
|
|
@click.option("--filelist", default=None, type=Path)
|