flash-attention/training/configs/model/gpt2model/gpt2-medium.yaml
2022-11-28 17:34:40 -08:00

7 lines
85 B
YAML

# @package _global_
model:
config:
n_embd: 1024
n_head: 16
n_layer: 24