Skip to content

Commit

Permalink
Add further aligntment
Browse files Browse the repository at this point in the history
make dropout % and window size match
  • Loading branch information
gkielian committed Sep 17, 2024
1 parent de054e7 commit 57be408
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions gpt_conf.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ class GPTConfig:
vocab_size: int = 50304 # GPT-2 vocab_size of 50257, padded up to nearest multiple of 64 for efficiency
n_layer: int = 6
n_head: int = 6
n_kv_group: int = 6
n_kv_group: int = None
n_embd: int = 384

# Steering Vectors
Expand Down Expand Up @@ -38,8 +38,8 @@ class GPTConfig:
export_scale_matrices_npz: str = None
export_scale_matrices_each_eval: bool = False

dropout: float = 0.0
window_size: int = 128
dropout: float = 0.2
window_size: int = None
gate: bool = False
use_moe: bool = False
moe_layer_freq: int = 2
Expand Down

0 comments on commit 57be408

Please sign in to comment.