Attentional Transformers and BERT for every sequential data

Recently papers like Attentional Transformers and Bert have been gaining attraction with NLP techniques (https://arxiv.org/pdf/1810.04805.pdf). I was wondering if these techniques can be applied for other sequential data like genomic amino acid and stock predictions? #yolo