Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
unify how to freeze some parameters for coca pre-training (#526)
Summary: 1. we already have support of freezing vision encoder; as experiment goes, we want to experiment to freeze other part of coca, e.g., text decoder. This diff provides unified way of freezing/unfreezing modules, the same way as we are doing for linear probe or finetune. 2. add configuration of using MLP instead of attention pooler for vision adapter; 3. for output projection in text decoder, change bias=False to True. In many other places, e.g., LP head, ember's output module and LLAVA, they are using bias=True (which is default value in Linear). Differential Revision: D54559503 Privacy Context Container: 303860477774201
- Loading branch information