I am trying to implement Skip-gram following Mikolov’s paper. http://papers.nips.cc/paper/5021-distributed-representations-of-words-and-phrases-and-their-compositionality.pdf
I am wondering if there is a method to do multiple negative sampling easily with PyTorch. Is there a way to define a distribution by word frequency, and sample n (say, n = 5) word together, for the purpose of computing negative sampling.