The early 20th century was an era of business expansion and progressive reform in the United States. The progressives, as they called themselves, worked to make American society a better and safer place in which to live. ... This generation of Americans also hoped to make the world a more democratic place.