From Wikipedia, the free encyclopedia.
The Western United States of America has played a significant role in both history and fiction. The term Wild West relates to an ever-shifting nebulous land populated by cowboys and Indianss, trappers and explorers; and located to the west of the settled, law-abiding United States and beyond The Frontier.
See also: The West (U.S.)
This article is a stub article. You can help Wikipedia by improving it.\n