Skip to content

Commit 5aa4c76

Browse files
authored
Fix mix-up with params
1 parent c92f265 commit 5aa4c76

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

ft-B-train-OpenAI-CLIP-ViT-L-14.py

+4-4
Original file line numberDiff line numberDiff line change
@@ -253,10 +253,10 @@ def forward(self, logits_per_image, logits_per_text):
253253

254254
# Potentially useful if you get gigantic gradient norms at the delicate layers near the input
255255
param_groups = [
256-
{'params': visual_parameters[:len(transformer_parameters)//2], 'lr': 1e-6}, # First half of the transformer
257-
{'params': visual_parameters[len(transformer_parameters)//2:], 'lr': 3e-6}, # Second half of the transformer
258-
{'params': transformer_parameters[:len(visual_parameters)//2], 'lr': 1e-6}, # First half of the vision transformer
259-
{'params': transformer_parameters[len(visual_parameters)//2:], 'lr': 3e-6}, # Second half of the vision transformer
256+
{'params': transformer_parameters[:len(transformer_parameters)//2], 'lr': 1e-6}, # First half of the transformer
257+
{'params': transformer_parameters[len(transformer_parameters)//2:], 'lr': 3e-6}, # Second half of the transformer
258+
{'params': visual_parameters[:len(visual_parameters)//2], 'lr': 1e-6}, # First half of the vision transformer
259+
{'params': visual_parameters[len(visual_parameters)//2:], 'lr': 3e-6}, # Second half of the vision transformer
260260
]
261261

262262
# Default optimizer AdamW (not recommended). Set to "AdamW(param_groups, ...)" to use above differential learning rates

0 commit comments

Comments
 (0)