Is there some Parameter Efficient Fine-Tuning technique like LORA(low rank adaptation) is available for CNN to reduce the GPU memory usage while training/fine-tuning the network?
Is it possible to apply LoRA for CNN kernels which will be of (3X3, 5X5, 7X7)?
Any article recommendations will be highly helpful. Thanks
Thank you for your response! I read in your paper that youâve added SSF to a ResNet50, did you add it to only the ResBlocks? And throughout the entire model? I tried to add it to the resblocks of a segresnet like
` scale, shift = init_ssf_scale_shift(in_channels)
self.ssf_scale = scale
self.ssf_shift = shift
def forward(self, x):
if self.tuning_mode == "ssf":
identity = x
x = self.norm1(x)
x = self.act(x)
x = ssf_ada(self.conv1(x), self.ssf_scale, self.ssf_shift)
x = self.norm2(x)
x = self.act(x)
x = ssf_ada(self.conv2(x), self.ssf_scale, self.ssf_shift)
x += identity`.
Then I only train the params with âssfâ in their name, and this gives very poor results.
Yes, LoRA can be adapted for CNNs to reduce GPU memory usage. Check âLoRA: Low-Rank Adaptation of Large Language Modelsâ for foundational concepts.