For decades, Hollywood has shaped the myth of the American West through its fixation on cowboys and Indians — a genre that has dominated cinema and television with romanticized portrayals of frontier ...