Pruning Adapters with Lottery Ticket
Massively pre-trained transformer models such as BERT have gained great success Hair Repair Mask in many downstream NLP tasks.However, they are computationally expensive to fine-tune, slow for inference, and have large storage requirements.So, transfer learning with adapter modules has been introduced and has become a remarkable solution for those