Answer:
How did the United States become more democratic between the 1790's and the 1830's? It changed from being an adult white male that must own property to be able to vote and hold office to just being a adult white male to be able to vote and hold office. ... His presidency marked the opening of a new and more democratic era.
Step-by-step explanation: