Long ago in a planet far, far away...
For a long time in the US, there was no official "standard" for utility voltages, each utility made their own standards. That was a nightmare for equipment mfrs, but that's the way it was. In the Depression, a government program called the REA was implemented to get electricity out to farmers in rural areas in order to boost productivity. In that effort, the lack of standards loomed big for the REA installers, so national standards were established. Voltages available from utilities ranged, in the "440V" realm for example, between 416V and 500V, so motor mfrs got together in what eventually became NEMA and designated a design standard calling for 460V, a compromise, but included a range of +-10% so that it covered all of the existing utility standards; 460-10% = 414V, 460 +10% = 506V. But motor mfrs who already were making "440V" motors were slow to change, it meant new winding designs, and that was expensive. So the practice continued for years. Then when we started buying stuff from Europeans and Asians, who had a completely different set of design standards, many of them either misread our requirements and started making "440V" stuff because they didn't know better (not being members of NEMA), or they PURPOSELY took advantage of the situation because they REALLY made 400V stuff, and 400V + 10% = 440V.
Still though, as the others said, should be fine. I was just giving you the "how did they get there" story.