Hi,
Following the instructions in this post to load the same opt 13b. I have access to 8 Nvidea A100 80GB machines.
model = AutoConfig.from_pretrained("facebook/opt-13b")
runs successfully and the model card is available to use. Example of the model
OPTForCausalLM(
(model): OPTModel(
(decoder): OPTDecoder(
(embed_tokens): Embedding(50272, 5120, padding_idx=1)
(embed_positions): OPTLearnedPositionalEmbedding(2050, 5120)
(final_layer_norm): LayerNorm((5120,), eps=1e-05, elementwise_affine=True)
(layers): ModuleList(
...
...
However, device_map = infer_auto_device_map(model)
returns {'': 0}
despite 8 GPUs being available.
The code is exactly the same as the blog post
from accelerate import infer_auto_device_map, init_empty_weights
from transformers import AutoConfig, AutoModelForCausalLM
config = AutoConfig.from_pretrained("facebook/opt-13b")
with init_empty_weights():
model = AutoModelForCausalLM.from_config(config)
device_map = infer_auto_device_map(model)
Not sure why this is the case.
Any suggestions/help appreciated. Thanks!