Simplify conversion + more debug

This commit is contained in:
Casper Hansen
2024-03-17 20:21:46 +00:00
parent d43a79b7bf
commit 04168801c9
2 changed files with 27 additions and 21 deletions

View File

@@ -1,14 +1,26 @@
import torch
from tqdm import tqdm
from axolotl.monkeypatch.moe.moe import SparseMoeBlock
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
from transformers.models.mixtral.modeling_mixtral import MixtralSparseMoeBlock
def compute_memory_used_pct(device):
memory_used = torch.cuda.max_memory_allocated(device) / (1024**3)
memory_pct = (
memory_used
/ (torch.cuda.get_device_properties(device).total_memory / (1024**3))
* 100
)
return memory_pct
model_path = "mistralai/Mixtral-8x7B-Instruct-v0.1"
# Load model
model = AutoModelForCausalLM.from_pretrained(model_path, device_map="auto")
modules = {k:v for k,v in model.named_modules() if isinstance(v, MixtralSparseMoeBlock)}
for name, module in model.named_modules():
if isinstance(module, MixtralSparseMoeBlock):
with tqdm(modules.items(), desc="scatter moe") as pbar:
for name, module in pbar:
smoe = SparseMoeBlock(
experts=module.experts,
gate=module.gate,
@@ -18,6 +30,9 @@ for name, module in model.named_modules():
top_k=module.top_k,
)
setattr(model, name, smoe)
for device_index in range(torch.cuda.device_count()):
device_memory_pct = compute_memory_used_pct(device_index)
print(device_index, device_memory_pct)
tokenizer = AutoTokenizer.from_pretrained(model_path)
streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)