92.9k views
0 votes
Ow did the US change the government of Japan after World War II?

User CPI
by
3.8k points

2 Answers

1 vote

Answer:

After Japan surrendered in 1945, ending World War II, Allied forces led by the United States occupied the nation, bringing drastic changes. Japan was disarmed, its empire dissolved, its form of government changed to a democracy, and its economy and education system reorganized and rebuilt.

User Justin Lange
by
3.5k points
3 votes

It created a democratic government.

User Yuan He
by
4.1k points