LLaMA-Factory/scripts/cal_flops.py

34 lines
1.2 KiB
Python
Raw Normal View History

2023-09-16 15:40:41 +00:00
# coding=utf-8
# Calculates the flops of pre-trained models.
# Usage: python cal_flops.py --model_name_or_path path_to_model --batch_size 1 --seq_length 512
# Inspired by: https://www.deepspeed.ai/tutorials/flops-profiler/
2024-01-20 12:15:56 +00:00
from typing import Optional
2023-09-16 15:40:41 +00:00
import fire
import torch
2024-01-20 12:15:56 +00:00
from deepspeed.accelerator import get_accelerator # type: ignore
from deepspeed.profiling.flops_profiler import get_model_profile # type: ignore
2023-09-16 15:40:41 +00:00
from llmtuner import ChatModel
2023-11-14 12:58:37 +00:00
def calculate_flops(
2023-09-16 15:40:41 +00:00
model_name_or_path: str,
batch_size: Optional[int] = 1,
seq_length: Optional[int] = 256,
2024-01-20 12:15:56 +00:00
flash_attn: Optional[bool] = False,
2023-09-16 15:40:41 +00:00
):
with get_accelerator().device(0):
2024-01-20 12:15:56 +00:00
chat_model = ChatModel(dict(model_name_or_path=model_name_or_path, template="vanilla", flash_attn=flash_attn))
2023-09-16 15:40:41 +00:00
fake_input = torch.ones((batch_size, seq_length), dtype=torch.long, device=chat_model.model.device)
2024-01-20 12:15:56 +00:00
input_dict = {"input_ids": fake_input, "labels": fake_input.clone()}
flops, macs, params = get_model_profile(chat_model.model, kwargs=input_dict, print_profile=True, detailed=True)
2023-09-22 16:34:17 +00:00
print("FLOPs:", flops)
2023-09-16 15:40:41 +00:00
print("MACs:", macs)
print("Params:", params)
if __name__ == "__main__":
2023-11-14 12:58:37 +00:00
fire.Fire(calculate_flops)