lm_inference_launch.py 3.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130
  1. #!/usr/bin/env python3
  2. import argparse
  3. import logging
  4. import os
  5. import sys
  6. from typing import Union, Dict, Any
  7. from funasr.utils import config_argparse
  8. from funasr.utils.cli_utils import get_commandline_args
  9. from funasr.utils.types import str2bool
  10. from funasr.utils.types import str2triple_str
  11. from funasr.utils.types import str_or_none
  12. from funasr.utils.types import float_or_none
  13. def get_parser():
  14. parser = config_argparse.ArgumentParser(
  15. description="Calc perplexity",
  16. formatter_class=argparse.ArgumentDefaultsHelpFormatter,
  17. )
  18. parser.add_argument(
  19. "--log_level",
  20. type=lambda x: x.upper(),
  21. default="INFO",
  22. choices=("CRITICAL", "ERROR", "WARNING", "INFO", "DEBUG", "NOTSET"),
  23. help="The verbose level of logging",
  24. )
  25. parser.add_argument("--output_dir", type=str, required=True)
  26. parser.add_argument("--gpuid_list", type=str, required=True)
  27. parser.add_argument(
  28. "--ngpu",
  29. type=int,
  30. default=0,
  31. help="The number of gpus. 0 indicates CPU mode",
  32. )
  33. parser.add_argument("--seed", type=int, default=0, help="Random seed")
  34. parser.add_argument("--njob", type=int, default=1, help="Random seed")
  35. parser.add_argument(
  36. "--dtype",
  37. default="float32",
  38. choices=["float16", "float32", "float64"],
  39. help="Data type",
  40. )
  41. parser.add_argument(
  42. "--num_workers",
  43. type=int,
  44. default=1,
  45. help="The number of workers used for DataLoader",
  46. )
  47. parser.add_argument(
  48. "--batch_size",
  49. type=int,
  50. default=1,
  51. help="The batch size for inference",
  52. )
  53. parser.add_argument(
  54. "--log_base",
  55. type=float_or_none,
  56. default=10,
  57. help="The base of logarithm for Perplexity. "
  58. "If None, napier's constant is used.",
  59. required=False
  60. )
  61. group = parser.add_argument_group("Input data related")
  62. group.add_argument(
  63. "--data_path_and_name_and_type",
  64. type=str2triple_str,
  65. action="append",
  66. required=False
  67. )
  68. group.add_argument(
  69. "--raw_inputs",
  70. type=str,
  71. required=False
  72. )
  73. group.add_argument("--key_file", type=str_or_none)
  74. group.add_argument("--allow_variable_data_keys", type=str2bool, default=False)
  75. group.add_argument("--split_with_space", type=str2bool, default=False)
  76. group.add_argument("--seg_dict_file", type=str_or_none)
  77. group = parser.add_argument_group("The model configuration related")
  78. group.add_argument("--train_config", type=str)
  79. group.add_argument("--model_file", type=str)
  80. group.add_argument("--mode", type=str, default="lm")
  81. return parser
  82. def inference_launch(mode, **kwargs):
  83. if mode == "transformer":
  84. from funasr.bin.lm_inference import inference_modelscope
  85. return inference_modelscope(**kwargs)
  86. else:
  87. logging.info("Unknown decoding mode: {}".format(mode))
  88. return None
  89. def main(cmd=None):
  90. print(get_commandline_args(), file=sys.stderr)
  91. parser = get_parser()
  92. args = parser.parse_args(cmd)
  93. kwargs = vars(args)
  94. kwargs.pop("config", None)
  95. # set logging messages
  96. logging.basicConfig(
  97. level=args.log_level,
  98. format="%(asctime)s (%(module)s:%(lineno)d) %(levelname)s: %(message)s",
  99. )
  100. logging.info("Decoding args: {}".format(kwargs))
  101. # gpu setting
  102. if args.ngpu > 0:
  103. jobid = int(args.output_dir.split(".")[-1])
  104. gpuid = args.gpuid_list.split(",")[(jobid - 1) // args.njob]
  105. os.environ["CUDA_DEVICE_ORDER"] = "PCI_BUS_ID"
  106. os.environ["CUDA_VISIBLE_DEVICES"] = gpuid
  107. kwargs.pop("gpuid_list", None)
  108. kwargs.pop("njob", None)
  109. results = inference_launch(**kwargs)
  110. if __name__ == "__main__":
  111. main()