To move the data on the relative dependencies of different tokens showing at unique locations within the sequence, a relative positional encoding is calculated by some type of Discovering. Two famed kinds of relative encodings are:LLMs need considerable computing and memory for inference. Deploying the GPT-3 175B model demands no less than 5x80GB A