Skip to content

Commit

Permalink
Update Gemma2 attention scale (#694)
Browse files Browse the repository at this point in the history
* Update Gemma2 attention scale

* remove import
  • Loading branch information
mntss authored Aug 11, 2024
1 parent 464325a commit e30f96b
Showing 1 changed file with 0 additions and 3 deletions.
3 changes: 0 additions & 3 deletions transformer_lens/loading_from_pretrained.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,6 @@

import dataclasses
import logging
import math
import os
import re
from pathlib import Path
Expand Down Expand Up @@ -1246,7 +1245,6 @@ def convert_hf_model_config(model_name: str, **kwargs):
"rotary_base": 10000.0,
"positional_embedding_type": "rotary",
"use_attn_scale": True,
"attn_scale": math.sqrt(224),
"n_key_value_heads": 4,
"window_size": 4096,
"use_local_attn": True,
Expand Down Expand Up @@ -1274,7 +1272,6 @@ def convert_hf_model_config(model_name: str, **kwargs):
"rotary_base": 10000.0,
"positional_embedding_type": "rotary",
"use_attn_scale": True,
"attn_scale": math.sqrt(224),
"n_key_value_heads": 8,
"window_size": 4096,
"use_local_attn": True,
Expand Down

0 comments on commit e30f96b

Please sign in to comment.