Ahmadzei's picture
added 3 more tables for large emb model
5fa1a76
LSH self attention uses the locality sensitive hashing mechanism proposed in
Practical and Optimal LSH for Angular Distance to assign each of the tied key
query embedding vectors to one of config.num_buckets possible buckets.