Complexity of attention in dynamic convolution

I’m reading this paper , Dynamic convolution-Attention over Convolution kernels.

    I couldn't understand the complexity of attention i.e., 

Pic from above paper

How to calculate O(π(x)) ? Explain.