You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
parameters, forward_fn, tokenizer, config=get_pretrained_model(
model_name="500M_1000G",
# Get embedding at layers 5 and 20embeddings_layers_to_save=(5, 20,),
# Get attention map number 4 at layer 1 and attention map number 14# at layer 12attention_maps_to_save=((1,4), (12, 14)),
max_positions=128,
)
Here it seems that different pretrained models have different configuration? I was wondering if you could further add detailed clarification about how to choose the embeddings_layers_to_save and max_positions. I also saw in some issues mentioning that we might need to set the max_positions to 1000? Just a bit confused and it would be the best if the authors could provide the suggested configures for each pretrained model somewhere in the tutorial or readme files.
The text was updated successfully, but these errors were encountered:
Hi,
When I looked at some examples of getting the pretrained models, I saw:
Here it seems that different pretrained models have different configuration? I was wondering if you could further add detailed clarification about how to choose the
embeddings_layers_to_save
andmax_positions
. I also saw in some issues mentioning that we might need to set themax_positions
to1000
? Just a bit confused and it would be the best if the authors could provide the suggested configures for each pretrained model somewhere in the tutorial or readme files.The text was updated successfully, but these errors were encountered: