Overall it's worse than the other frontier models, but it's decent for queries about breaking news, due to being trained on twitter data.
It's also better for queries about controversial topics, and topics that the other labs have deemed to be "unsafe".
Politically, it differs quite a bit from other models.[0] It's right leaning, although it's closer neutral than other models, defining what neutral is a challenge though.
The study you link to doesn't take into consideration the Overton window of opinions. Perhaps there's some dimension along which you could say that one ideology lies 'opposite' to another political persuasion, but that doesn't necessarily mean that the two ideologies are equally acceptable to support in a given society.
I don't think calling defining neutral a 'challenge' does the question justice - neutral will always be context-dependent, and what may be in the center of the Overton window of one society may be unpopular or even highly illegal in a different society.
Grok and Elon's ventures in general should really get the Purpose of a System treatement in public discourse. For all we know the purpose of Grok is to make nude edits of people. You can assign this to left or right leaning as you please.
Politically, it differs quite a bit from other models.[0] It's right leaning, although it's closer neutral than other models, defining what neutral is a challenge though.
[0]: https://arxiv.org/abs/2603.23841