- Jul 10, 2006
- 29,873
- 463
- 126
Some progressive posted something that got me to thinking. He/she basically said that states were only needed because existing communications prevented the federal government from effectively exercising its will on all Americans. It's apparent that the left in general sees no role for states in health care. So I'm curious as to how others, particularly self-identified liberals and/or progressives, see the proper role of the fifty states in twenty-first century America. Is the Tenth Amendment no more than flotsam upon the sad wreckage of our once-proud Constitution?
Should the states be abolished and replaced with federal administration and enforcement offices?
Should states be in control of any particular area, perhaps education?
Or charged only with raising funds to operate federally mandated programs such as Medicare?
Should they be in charge of business within their borders?
Relegated to social issues? Barred from social issues?
Are states anachronistic entities better put to sleep, or vital sources of leadership and self-governance?
Should the states be abolished and replaced with federal administration and enforcement offices?
Should states be in control of any particular area, perhaps education?
Or charged only with raising funds to operate federally mandated programs such as Medicare?
Should they be in charge of business within their borders?
Relegated to social issues? Barred from social issues?
Are states anachronistic entities better put to sleep, or vital sources of leadership and self-governance?