This paper presents a novel graph representation learning approach utilizing graph multiset pooling to effectively encode graphs by considering structural constraints and node interactions. The proposed graph multiset transformer (GMT) model demonstrates substantial performance improvements in graph classification, reconstruction, and generation tasks compared to existing methods, while also showcasing efficiency in memory and time usage. The findings suggest that the GMT can reach the expressive power of the Weisfeiler-Lehman test, paving the way for enhanced graph analysis techniques.