问题 I have checked the PyTorch tutorial and questions similar to this one on Stackoverflow. I get confused; does the embedding in pytorch (Embedding) make the similar words closer to each other? And do I just need to give to it all the sentences? Or it is just a lookup table and I need to code the model? 回答1: nn.Embedding holds a Tensor of dimension (vocab_size, vector_size) , i.e. of the size of the vocabulary x the dimension of each vector embedding, and a method that does the lookup. When you