Answer:
The political doctrine or belief held by the United States, particularly during its expansion, that the nation had a special role and divine right to expand westward and gain control over the continent.
Step-by-step explanation:
How did manifest destiny lead to westward expansion?
Manifest Destiny brought money, land, resources, and a strengthened economy to the Americans. This idea that it was their destiny to expand caused Americans to disregard the territorial rights of Native Americans, wiping out many tribes and causing a cultural divide, tension and wars.