64.1k views
2 votes
How did japan change after world war II

User Iled
by
3.2k points

2 Answers

5 votes
Japan has to deal with a lot of charges and issues after the world war. Just recently, they finally finished paying us off the money they owed us.
User Joe Johnston
by
4.1k points
5 votes

Answer: After Japan surrendered in 1945, ending World War II, Allied forces led by the United States occupied the nation, bringing drastic changes. Japan was disarmed, its empire dissolved, its form of government changed to a democracy, and its economy and education system reorganized and rebuilt.

Step-by-step explanation:

User Scubadivingfool
by
3.7k points