Western Film

The American Film Institute defines Western films as those "set in the American West that [embody] the spirit, the struggle and the demise of the new frontier."

Film edit

Western drama edit

Television series edit