As it stands, Rogers explained, we’re losing somewhere between $100 billion and $400 billion worth of intellectual property to theft each year. This, he said, is of particular concern to the Department of Defense, which watches as its contractors networks are regularly compromised by adversaries.
To the extent that his statement reflects how US tax dollars are spent, the situation is to a considerable extent the result of military downsizing under President Clinton, which wasn’t really downsizing, but outsourcing. Issued on May 21, 1996, Executive Order 13005 – Empowerment Contracting states its objectives as follows:
In order to promote economy and efficiency in Federal procurement, it is necessary to secure broad-based competition for Federal contracts. This broad competition is best achieved where there is an expansive pool of potential contractors capable of producing quality goods and services at competitive prices. A great and largely untapped opportunity for expanding the pool of such contractors can be found in this Nation’s economically distressed communities.
The problem is, this way of approaching “efficiency” leads directly to reduced accountability. Those who rail against government inefficiency don’t understand that accountability is not efficient: it is not efficient to justify your actions at every step. So the push to make government more lean and “efficient” by outsourcing government functions to the private sector leads directly to an erosion of accountability. You can’t have both accountability and efficiency as policy goals.
Pulitzer prize winning historian Gary Wills suggested that, for example, part of why the Manhattan Project was conducted with such extraordinary secrecy was specifically to evade accountability. The Russians knew what we were up to, the Germans probably knew too, it was the American people kept in the dark. Wills argues this was probably to avoid potential opposition to the development of nuclear weapons in light of the 1925 Geneva Protocols against chemical and biological weapons. Around the globe, people were shocked by the destructiveness of mechanized warfare during World War I and by the use of chemical weapons. The First World War and the technological horrors is brought were still very much in public memory by the time World War II came around.
Wills also points out that this use of secrecy to evade accountability was no isolated instance. When the US bombed Cambodia, the Cambodians knew it, it was US citizens kept in the dark. When the US invaded Cuba, the Cubans knew what was happening and the Soviets knew, it was US citizens kept in the dark.
Today we have active drone campaigns in at least eight foreign countries responsible for the deaths of thousands in what is essentially an undeclared global war. Insofar as the targets are terrorists, the terrorists know they’re being targeted. Again, it’s US citizens kept in the dark.
The origin of the “state secrets” doctrine derives not from any law that Congress passed, but from efforts by the US military to evade accountability over flaws in the engine design of a new aircraft, which led to the deaths of several citizens.
Accountability is not efficient. To increase accountability with surveillance matters, there needs to be a reduction in contracting, which means, the government needs to get bigger. Edward Snowden — a contractor himself — would seem to be a clear cut example in support of this view.
There’s no Internet without surveillance. The Internet was built by the US military to be robust, not for privacy or security. Privacy was not part of the Internet’s design goals.
The Internet became a commonplace household word in part because of the hype surrounding an economic bubble created during the presidency of Bill Clinton. Under Bill Clinton, the US Congress also enacted the Communications Assistance for Law Enforcement Act at the same time that Windows 95 introduced Americans to personal computers and the phrase “information superhighway” introduced Americans to networking. Surveillance was an integral part of handing the Internet over to commerce.
The relationship between commerce and the surveillance state is now well-established: Apple and Microsoft are suspect, and Yahoo has made surveillance a business proposition — as per 18 U.S.C. § 2706, Yahoo’s 2009 rates ran as follows:
Basic subscriber records cost $20 for the first ID, $10 per ID thereafter; basic group Information (including information about moderators) cost $20 for a group with a single moderator; contents of subscriber accounts — including email — cost $30-$40 per user; contents of groups cost $40 – $80 per group.
Given that typical internet advertising revenue brings in only pennies per click, the current scale of Internet surveillance clearly implies that spying on customers is big business for online firms.
Other telecommunications carriers have made similar overtures, some companies have faced legal and economic reprisal for refusing to cooperate, and yet others have availed themselves of their free speech rights as corporate persons to engage in this dubious commerce.
It should be reason enough to be disturbed by NSA surveillance that the Founders prohibited this type of information gathering in the 4th Amendment to the US Constitution. The excuse “I’ve got nothing to hide” misses the point. The government should obey the law, that’s a core feature of what “rule of law” means. And the example of non-violent resistance through non-participation set by Ghandi and the Southern Christian Leadership Conference and vegetarians and vegans offers a clear a lesson for how to resist the surveillance society: stop participating in an abusive system. The Internet is cruelty to human animals and it’s bad for the social environment.
If it weren’t for so many Americans purchasing data plans on “smart” phones, purchasing home Internet access, and dutifully reporting their daily thoughts and habits psychological makeup on FaceBook accounts, the costs to Uncle Sam for maintaining the current surveillance state would very rapidly prove prohibitive. That is, if the government had to pay your phone bill and your internet costs and pay a spy to follow you around to listen in on your conversations, it could no longer afford to spy on everybody. Through consumer habits and the cultural value placed on convenience, Americans effectively subsidize the surveillance state on behalf of the government. Dan Geer stated the matter succinctly: our online choices are between freedom, security, and convenience, but we can only pick two.
From a cost perspective, a “vegetarian” approach to resisting the surveillance state (that is, by simply opting out) is an inexpensive solution that aims at increasing the cost of surveillance to the state. This approach requires little social coordination other than a shared will to change prevailing circumstances — and a little personal initiative. Such a “vegetarian” approach also serves to inject additional uncertainty into what data is gathered (thereby diminishing the value of what data Uncle Sam does collect). This doesn’t mean life without the internet any more than vegetarianism means life without food, it just means being more selective about where your internet comes from, where you take it, and what you do with it.
You don’t need to be online all day. A good starting point would be to make a habit of leaving your cellphone tracking device at home once in a while. Just because your cellphone is wireless, that doesn’t mean you need to take it with you everywhere you go. If you take it with you everywhere you go, it’s more of a tracking device than a phone. When Uncle Sam looks through your cell tower data, changing your cellphone habits will increase the uncertainty as to your location at any given time during the day.
If you care to preserve “democracy,” all that’s really needed is a little social coordination and a willingness to put up with a little less “convenience.” This may sound incompatible with the modern world, but there’s good reason to get motivated: the modern world is incompatible with the perpetuation of the human race. There’s more at stake than a little privacy, though the more fundamental problem is bound up with the psychology of consumer society: in a growth economy based on persuasion though advertising — where consumers must make choices about the allocation of their scarce resources — every new product requiring new investment must be presented as needful and fundamental to the modern way of life.
Many people know things have gone awry with the modern world: between the threats posed by persistent national militarism, thermonuclear war, war over resources, mass hunger, environmental degradation, climate change, shortening attention spans, new communicable diseases — something is clearly wrong. And yet, somehow, everyone looks to another for the solution. Nobody is willing to see their complicity and change their behavior. So: if you don’t like internet surveillance, stop surveilling yourself. The problem isn’t some nebulous “big brother,” it’s you. The government isn’t going to change its behavior, so stop waiting for the government to save you from the government. You have to save yourself from yourself.
In 1988, Guy Debord observed:
“We believe we know that in Greece, history and democracy appeared at the same time. We can prove that their disappearances have also been simultaneous.”
“To this list of the triumphs of power we should, however, add one result which has proved negative for it: a State, in which one has durably installed a great deficit of historical knowledge so as to manage it, can no longer be governed strategically.”
This absence of historical knowledge manifests itself today. In Fall of 2013, after weeks of partisan gridlock, the US Congress managed to re-opened the Federal government with a last-minute deal. This partisanship, however, is not ideological: it is emotional, it is irrational, and its results are unpredictable.
In response to this minor “accomplishment” President Obama remarked:
“Let’s work together to make the government work better, instead of treating it like an enemy, or making it worse. That’s not what the founders of this nation envisioned when they gave us the gift of self-government.”
The President’s statement seems, on its face, uncontroversial — and that’s a big part of the problem. His statement is profoundly anti-historical, and in the most problematic manner possible, reads present values into the past.
Until 1850 or so, only white men with substantial wealth — such as bankers, factory owners, or plantation owners — were allowed the vote. Renters, subsistence farmers, and the laboring majority — whites and blacks — lacked political representation at the time of the US’s founding. The Constitution made no mention of suffrage until the 14th and 15th Amendments in 1868 and 1870 — almost 100 years after the original Constitution was ratified. In 1875, the US Supreme Court explicitly ruled that the 14th Amendment — which defined citizenship for the first time — did not give women the right to vote. Women didn’t get the vote until 1920. Blacks didn’t get full rights until the Civil Rights Act of 1964 and the Voting Rights Act of 1965.
That’s an odd “gift of self-government.” Truly, it sounds more like a long, hard struggle to obtain self-government — IN SPITE OF the Founding Fathers. Indeed, despite being subjects to the Crown, the British managed to outlaw the slave trade, abolish slavery, and enfranchise women before the United States. And despite being subject to the Crown, the British today enjoy many of the same rights — like freedom of religion, freedom of the press, freedom of speech — that Americans consider distinctly American innovations under the Bill of Rights.
Though the myth of Democracy endures, it is plain to see that the Founders feared Democracy, and made provision to prevent its emergence in the New World. That so much time passed before American citizens obtained universal suffrage is testament to the effectiveness of the Founding Fathers’ plans.
In arguing for a new Constitution to replace the Articles of Confederation, Founding Father Elbridge Gerry complained to the Constitutional Convention on May 31, 1787: “The evils we experience flow from the excess of democracy.” Founding Father Edmund Randolf also complained about the “turbulence and follies of democracy.” In Convention, Founding Father John Dickinson argued against expanding political enfranchisement: “The danger to free governments has not been from freeholders, but those who are not freeholders.” Dickinson went so far as to claim that a constitutional monarchy was “one of the best Governments in the world.”
Founding Father Alexander Hamilton, in Convention on June 18, 1787, expressed his belief that “nothing but a permanent body can check the imprudence of democracy … you cannot have a sound executive upon a democratic plan.” In The Federalist #10, James Madison wrote: “democracies have ever been spectacles of turbulence and contention, have ever been found incompatible with personal security and the rights of property.”
Thomas Jefferson — who did favor democracy — was not particularly effective in pushing his views. He was not even in the country when the Constitution was drafted and ratified.
The attainment of self-government in the United States was no “gift.” President Obama is either unaware of this nation’s history, or perhaps he has no problem brushing it aside in his public statements. Perhaps he is content that the myth of America’s long history of revolutionary self-government is a suitable expedient in the short-sighted calculus of contemporary electoral politics.
Assuming a political narrative where “liberals” oppose “conservatives,” one might think the first black president in the United States would show some slight interest in calling attention to the heritage of liberal reformers and progressives who paved the way for him to attain high office. One might suppose it would be in his political advantage to make it clear for all to see that the rights most “conservatives” today enjoy are the result of the efforts of their political adversaries.
Instead, the President’s choice to ignore this history — and to implicitly endorse the a-historical nationalist myth favored by self-described “conservatives” — obscures real threats to what measure of democracy Americans have gained by long struggle.
“Conservatives” who think they are defending the “gift” of democracy vehemently favor so-called “voter ID” laws — which have the effect of disenfranchising students and the elderly. The ostensible rationale for these laws — that they prevent voter fraud — not only points to a problem that does not appear to exist, but affects a sort of bait-and-switch. These laws brush aside legitimate concerns about election fraud — they ignore systemic flaws with electronic voting machines, irregularities in the 2004 elections which may have been covered up, voting irregularities and questionable legal activities surrounding the contested 2000 election, and ongoing voting irregularities in parts of small town America like Waukesha, Wisconsin.
In addition to enacting “voter ID” laws, “conservative” governors have been purging voter rolls, and continue to push for national policies — such as the failed War on Drugs and mandatory minimum sentencing laws — which have had the net effect of leaving one in five black men disenfranchised.
The “conservatives” perhaps don’t know to what condition their “traditionalist” views are returning this country — and our “liberal” President does not seem particularly concerned with remedying the matter. And for all the lofty speech of “history” surrounding the 2008 elections, the word was routinely used not by way of elucidating the past, but by way of branding what was then the present moment.
All this is perhaps fitting. President Obama’s policies are by and large center-right. With the exception of gay marriage, he has a dismal civil rights record, that includes granting amnesty to CIA torturers, failing to close the Guantanamo Bay detention facility, extrajudicial assassinations of US citizens, requesting indefinite detention provisions in the 2012 National Defence Authorization Act, negotiating secret treaties, expanding and legitimating Bush-era surveillance programs — despite well-documented evidence detailing what these programs can lead to, violating the sovereignty of foreign nations with drone strikes, prosecuting whistleblowers with a vengeance … and the list goes on.
Mr. Obama’s calls for “unity” are perhaps well-intentioned, but cannot be strategically effective in the absence of historical knowledge among the population. Something closer to a one-party system is unlikely to mend the damage caused by a dysfunctional two-party system. What America needs is not “unity” but a real opposition party — a role that, in the face of “conservative” efforts in the Tea Party Caucus — the Democratic party seems unwilling or unable to fulfill.
Although many states in the US have been passing laws allowing for medical and even recreational use of cannabis, the substance remains illegal at the federal level. Specifically, cannabis is classified as a Schedule I substance under the Controlled Substances Act. This classification is on the basis of three main criteria:
1. The drug or other substance has a high potential for abuse.
2. The drug or other substance has no currently accepted medical use in treatment in the United States.
3. There is a lack of accepted safety for use of the drug or other substance under medical supervision.
Cannabis is classed with Heroin and LSD. Cocaine, Morphine, and Oxycodone are less tightly regulated Schedule II substances.
As cannabis use persists among teenagers and adults, both legally and illegally, more and more people — especially young people — are seeing first hand that the risks associated with the use of cannabis don’t square properly with the federal government’s treatment of cannabis. If the government has an interest in protecting young people from the risks of substance abuse, they also have an interest in providing accurate information and formulating sensible policies that don’t simultaneously undermine their own credibility.
Problems with the Current Classification of Cannabis
The Schedule I classification of cannabis has a number of problems. First, abuse is hard to quantify, and just what patterns of cannabis usage fall under this rubric are not well defined. Second, there are few quantifiable safety concerns with cannabis: the substance is profoundly non-toxic, and it is, for all practical purposes, impossible to overdose on cannabis. This distinguishes cannabis from other Schedule I substances like heroin, and even from legal recreational drugs like alcohol, which is a contributing factor in the death of some 80,000 Americans each year. Third, the current federal scheduling of cannabis does not take into consideration the accepted medical use of cannabis in a number of states. The Department of Veterans Affairs has issued a formal directive permitting the clinical use of cannabis in those states where medical uses are approved. Researchers studying the relative risks and merits of the substance encounter great difficulties acquiring suitable samples to study, and their findings are of limited applicability to the way the substance is routinely consumed in a non-standardized, non-regulated black market.
Perhaps the most dramatic difficulty with the federal government’s position on cannabis is that the US Department of Health and Human Services holds a patent on medical uses of cannabis. Issued in 2003, US Patent #6630507 is titled “Cannabinoids as antioxidants and neuroprotectants.” The patent examines a molecule found in cannabis, CBD, though the chemical mechanism the patent identifies should be present in all cannabinoids, including THC. The patent notes that “cannabinoids are found to have particular application as neuroprotectants, for example in limiting neurological damage following ischemic insults, such as stroke and trauma, or in the treatment of neurodegenerative diseases, such as Alzheimer’s disease, Parkinson’s disease and HIV dementia,” and also indicates that cannabinoids offer a unique delivery mechanism due to the facility with which these molecules can cross the blood-brain barrier.
When cannabis was originally listed as a Schedule I substance in 1970, the classification was intended to be provisional, pending the results of an ongoing study. The National Commission on Marijuana and Drug Abuse issued the study findings in 1972, finding that there “is little proven danger of physical or psychological harm from the experimental or intermittent use of the natural preparations of cannabis.” Although the study recommended de-criminalizing cannabis and treating use or possession akin to alcohol, President Nixon chose not to implement the Commission’s recommendations, and marijuana has remained a Schedule I substance since. Although whole-plant marijuana remains a Schedule I substance, the synthetic THC called dronabinol — sold under the brand name Marinol — is classified as a less-restricted Schedule III substance.
Social Attitudes Affecting Cannabis as Medicine
In the United States, a lot of opposition to medical cannabis laws have presumed that such laws are just a “first step” towards outright legalization. While there is little to suggest such an outcome would be inherently detrimental, there is also ample evidence that supports medical uses of cannabis on the substance’s own merits.
What presents a more profound problem to the public is in part a tacit sociology of medicine that limits and proscribes how individuals view treatment. Politicians have adopted these cultural attitudes unquestioningly — indeed, the authoritarian personalities of these politicians wouldn’t allow them to ask such questions. Those who are open to such questions don’t dare assert themselves, despite polling results that show 70-85% of Americans favor significant changes in current federal policy.
Of particular note in this regard is the unexamined notion that medicine has to come in the form of an expensive bitter pill. The notion that medicine might also be pleasurable is anathema, and that healing might be enjoyable is equally heretical. Medicine is still penance, disease is sin, the new medical complexes are cathedrals, and doctors are the high priesthood, mediating between this world and the next, serving as both the front line and the last defense against the forces of corruption, decay, and disorder.
We apologize when we call in sick to work, and are stigmatized by our ailments. Just as the medieval church was one of the largest landlords in Europe, today’s medical industry claims vast swaths of the GDP. In the US, healthcare spending exceeds the 10% tithe commanded by the medieval church. The religion analogy is quite complete, and includes the irreligiousness of the most ardent devotees. Hospitals, gathering together the diseased, are diseased. They are morally perverse and rotten with wealth.
Along with unexamined notions of how medicine fits into our culture, there is another factor promoted by our culture, related to the ideology of Progress. Progress holds that the future will always bring improvements, that all new technology is better technology, and that what is new must replace what is old. From within the confines of this ideology of Progress, it seems on the face of things obvious that any new pill is inherently superior to “natural preparations.” This is, unfortunately, quite difficult to establish with any certainty.
There are easy-to-identify counter-examples where modern medicine has delivered a harmful product: the recall of pills like VIOXX make a big splash in the media, and create the impression that these are exceptions to the general rule that modern medicine routinely delivers improvements. But these issues have been with medicine for a long time: heroin, for example, was originally invented by the pharmaceutical company Bayer, and marketed as a non-addictive alternative to morphine.
The litany of prescription painkillers marketed since Bayer invented heroin have now surpassed car crashes in the number of annual deaths they cause, accounting for some 90% of all poisonings. The number killed by these drugs amounts to about ten 9-11’s each year — every year. Instead of figuring out how to deal with this plague, however, the US throws more and more money at the medical industry, which keeps developing new drugs with serious side effects and abuse potential. The broader, social implications of this are even more troubling.
The Decline of Western Medicine
Most of modern medicine is unnecessary. After sanitation and hygiene, antibiotics, analgesics and anesthetics, and vaccines, most of modern medicine is devoted to coping with the side effects of industrialization. This effect can be seen in diet particularly, but also with respect to such vectors as environmental pollution. Environmental pollution may take the form of contaminants in the air and water, particulate matter in the air (which causes diseases like asthma), or increased radiation in the environment (due to industrial processes, residues from atmospheric nuclear testing, or because of solar radiation that is increased by a depleted ozone layer in the upper atmosphere).
If he or she lives past the age of 15, the typical hunter-gatherer stands a reasonable chance of remaining healthy and active into their 70’s, with a strong social support network to care for them as they age. The modern US health care industry really doesn’t do all that much better. A sizable portion of the modern improvements in life expectancy over what is offered by a hunter-gatherer society come from improved infant mortality, a hygiene problem identified by Ignaz Semmelweis in 1847. Hand washing is a an extraordinarily cheap and effective medical technology. Antibiotics, which were developed for around $20,000 of basic research, have saved many more individuals from childhood disease, and increased the range of surgeries that are possible.
As modern medicine grows more expensive, its productivity declines precipitously. This decline in productivity can be measured in terms of substantive outcomes or in terms of the cost per patent. Either way, the role of diminishing returns in this field is not adequately addressed in the contemporary discourse.
Most of the big medical breakthroughs of the last 300-500 years were inexpensive. Everything recent is increasingly expensive and of rapidly declining effectiveness compared to basic innovations like sanitation or antibiotics. Most modern medicines and medical procedures could be avoided through less expensive means, specifically, through dietary and behavior modification.
The cost of medicine detracts from other public welfare programs, such as nutrition, food security, education, and mass transit, all of which yield a far greater return on investment than modern medicine.
At some point, the moral aspects of modern medicine need to be evaluated in terms of the social cost. For example: as a percentage of GDP, the US spends three times more money on the healthcare industry than on education. We know that basic education makes us smarter, better socialized, and better equipped for employment; but most medicine isn’t really making us all that much healthier.
Marijuana and Medicine
Progress makes a raw agricultural commodity like cannabis seem suspect as a medicine, through really, it is modern medicine that should be suspect. Whereas a typical television commercial for a new pharmaceutical product will often devote more than half its airtime to potential side effects, no similarly funded social initiative exists to teach Americans how to eat properly, or how to prepare nutritious foods. Nutritious foods or vegetarian diets are routinely mocked by Americans.
Somehow, none of this is a medical problem. Rather, in the discourse, these are treated as political problem. So doctors, not being politicians, stay out of politics; and, somehow, proper diet is only of tangential concern to the medical industry, while new drugs of dubious effectiveness are promoted as indispensable innovations.
Somehow health is not a medical issue, only disease warrants attention. And where medicine and politics do intersect on this issue of cannabis, instead of informed discussion, the public is treated to a wall of silence, or else jokes about hapless stoners.
In the wake of this most recent school shooting in Connecticut, it is important to remember that these are still isolated incidents. That these cases become public spectacles represents an ideology of indoctrination more than they represent symptoms of a peaceful society changing into something frightening and more violent.
Our society is already violent. 12,000 gun homicides annually ranks our “peaceful” Homeland among many war zones. This is a 10-year low. For every gun homicide, there is roughly one accidental gun death. But very few of these deaths take place in mass shootings: they are overwhelmingly the results of domestic disturbances, armed robberies, and gang violence. Yet, when these more common types of shootings occur, the media is largely silent. If it were otherwise, the news would be about nothing but shootings.
When the media picks up gun violence as a topic — typically when a rare mass shooting occurs — the narrative the media propagates tells audiences that these acts are result of some “disturbed” individual. Now, clearly, shooters in domestic disturbances are “disturbed” at the time, but the media means to say something about psychological pathology: that mass shooters are somehow unhinged and acting irrationally. Relatively little attention is paid to what sorts of social pressures these mass shooters may have been under. And almost no attention is paid to the significance of the fact that these shooting sprees often end with a suicide.
The Suicide Shooter
What should be deeply disturbing to media consumers is that the sensationalism of the media coverage is oriented more towards inflicting emotional trauma on audiences than providing a useful description of what gun violence in the country actually looks like.
The media seems to be operating under the premise that when American audiences see coverage of events like this, audiences implicitly understand that these events are stand-ins for countless other events that vary widely in their specifics. All the Tweets and FaceBook posts to the effect that, “if only more teachers carried weapons in the classroom” seem to contradict such an assumption. I would wager that if more teachers were armed in class, there may be more school shootings: it is not safe to assume that all teachers love children unconditionally, or are a paragon of infinite patience.
Perhaps most disturbing about the media coverage of these events — in a sociological sense — is the uniform lack of discussion of the frequent suicidal climax of the killing. Perhaps the media feels it is enough to state that the shooters are “disturbed,” and that suicide therefore seems natural enough as a conclusion to such a “disturbed” episode. But we do have another paradigm for suicide attacks like these: suicide bombers in the Middle East.
It is not so easy to write off Middle Eastern suicide bombers as uniformly “disturbed.” They are trained in their attack strategies — much like our suicide shooters plan assaults. Suicide bombers are motivated by ideology or other political goals — which often enough appears to be the case with our suicide shooters. While suicide bombers are promised great rewards in the afterlife, our suicide shooters live in a culture saturated with aesthetic treatments of violence, such that the rush of a real-life shoot out and the subsequent notoriety may be reward enough.
Often enough, the perpetrators of these acts of mass violence do have reasons, though, due to the media’s sensationalism and lack of analysis, we never get any closer to understanding those reasons or, consequently, understanding what we can do address them. All we are left with, as a nation, is televised grieving, emotional trauma inflicted by the media over distant events, and nebulous debates about “gun control.”
The Calculating Killer
Joe Stack, who after several failed attempts at pursuing the American Dream through entrepreneurism, eventually committed suicide by flying a private plane into an IRS tax office in Texas. Joe Stack wrote a suicide note, titled, “Well Mr. Big Brother IRS man… take my pound of flesh and sleep well.” In it, he stated:
“If you’re reading this, you’re no doubt asking yourself, ‘Why did this have to happen?’ The simple truth is that it is complicated and has been coming for a long time. The writing process, started many months ago, was intended to be therapy in the face of the looming realization that there isn’t enough therapy in the world that can fix what is really broken. Needless to say, this rant could fill volumes with example after example if I would let it. I find the process of writing it frustrating, tedious, and probably pointless… especially given my gross inability to gracefully articulate my thoughts in light of the storm raging in my head. Exactly what is therapeutic about that I’m not sure, but desperate times call for desperate measures.
“We are all taught as children that without laws there would be no society, only anarchy. Sadly, starting at early ages we in this country have been brainwashed to believe that, in return for our dedication and service, our government stands for justice for all. We are further brainwashed to believe that there is freedom in this place, and that we should be ready to lay our lives down for the noble principals represented by its founding fathers. Remember? One of these was ‘no taxation without representation’. I have spent the total years of my adulthood unlearning that crap from only a few years of my childhood. These days anyone who really stands up for that principal is promptly labeled a ‘crackpot’, traitor and worse.”
Joe Stack doesn’t sound like his reason has become unhinged, he sounds like society pushed him to the breaking point.
Some time earlier, the Unabomber, Ted Kaczynski, wrote in in his Manifesto:
“96. As for our constitutional rights, consider for example that of freedom of the press. We certainly don’t mean to knock that right: it is very important tool for limiting concentration of political power and for keeping those who do have political power in line by publicly exposing any misbehavior on their part. But freedom of the press is of very little use to the average citizen as an individual. The mass media are mostly under the control of large organizations that are integrated into the system. Anyone who has a little money can have something printed, or can distribute it on the Internet or in some such way, but what he has to say will be swamped by the vast volume of material put out by the media, hence it will have no practical effect. To make an impression on society with words is therefore almost impossible for most individuals and small groups. Take us (FC) for example. If we had never done anything violent and had submitted the present writings to a publisher, they probably would not have been accepted. If they had been accepted and published, they probably would not have attracted many readers, because it’s more fun to watch the entertainment put out by the media than to read a sober essay. Even if these writings had had many readers, most of these readers would soon have forgotten what they had read as their minds were flooded by the mass of material to which the media expose them. In order to get our message before the public with some chance of making a lasting impression, we’ve had to kill people.”
Say what you will about his methods, but his thinking was not delusional or unhinged from reason: this was a man pushed by society to his breaking point. His actions were not the result of an irrational outburst. He sought social change.
For a media apparatus that literally sold us the lies that lead up to our 2003 invasion of Iraq; that has been largely silent on the erosion of the 4th Amendment under the FISA amendments and the free reign given to the National Security Agency; that has been largely silent on the erosion of the 5th Amendment under the 2012 NDAA; that is silent about illegal drone campaigns in foreign countries, extrajudicial assassinations of US citizens without trial, that still hasn’t gotten to the bottom of our war crimes in Fallujah, or what actually happened in the 2000 elections; that allows the sprawling security state unchecked growth, and that even cheerleads for the technologies that make it all possible, subsidized by mass production yields in consumer goods … all the media’s moralizing on issues like this rings hollow.
Anarchist thinker Emma Goldman had a different take on events like these:
“To Analyze the psychology of political violence is not only extremely difficult, but also very dangerous. If such acts are treated with understanding, one is immediately accused of eulogizing them. If, on the other hand, human sympathy is expressed with the Attentäter, one risks being considered a possible accomplice. Yet it is only intelligence and sympathy that can bring us closer to the source of human suffering, and teach us the ultimate way out of it.”
Emma Goldman suggests that such violent outbursts are the product of an unusually sensitive soul, not an irrational madman or an unfeeling sociopath. She does not condone these events, but seeks to understand them both emotionally and rationally, as part of a larger program towards making the world a more just and equitable place. The media could learn a few things from her approach.
When the media or politicians raise the spectre of “gun control” following such incidents — which is logical enough, given the total lack of any more substantive debate about the causes and conditions underlying such violent outbursts — a certain highly vocal group of Americans starts clamoring about their rights.
Recent Supreme Court rulings depart from precedent dating back to the 1930’s, mainly, the Miller Case, which affirmed that a sawed-off shotgun is not a militia weapon, and that individual ownership of such weapons can therefore be curtailed.
The decision in the 1939 Miller case reads, in part:
“The Court cannot take judicial notice that a shotgun having a barrel less than 18 inches long has today any reasonable relation to the preservation or efficiency of a well regulated militia, and therefore cannot say that the Second Amendment guarantees to the citizen the right to keep and bear such a weapon.”
In their close-reading of the text of the Constitution, the justices in Miller asserted:
“The Constitution, as originally adopted, granted to the Congress power —
“To provide for calling forth the Militia to execute the Laws of the Union, suppress Insurrections and repel Invasions; To provide for organizing, arming, and disciplining, the Militia, and for governing such Part of them as may be employed in the Service of the United States, reserving to the States respectively, the Appointment of the Officers, and the Authority of training the Militia according to the discipline prescribed by Congress.
“With obvious purpose to assure the continuation and render possible the effectiveness of such forces, the declaration and guarantee of the Second Amendment were made. It must be interpreted and applied with that end in view.”
The Second Amendment, in the strictest sense, is not the “gun” amendment, but the “militia” amendment. The introductory clause, “A well regulated militia,” serves the legal function of a “whereas” clause, delineating the scope of the provision. In the Second Amendment’s equivalent clause under the Articles of Confederation, the text is explicit in that arms are to be kept “in public stores.”
This is not to say there is not a Constitutional right for individuals to keep and bear privately-owned weaponry; but if the individual right to own a private firearm derives from the Second Amendment, it is the result of activist judges rather than the original intent of the Founding Fathers. An individual right to own a firearm is more plausibly found in the 9th or the 10th Amendment.
Gun advocates, however, do not make this argument, because although they frequently want smaller government and despise activist judges, there is no judicial case history testing gun rights under these more plausible Amendments.
Moreover, if individual gun ownership derives from the 9th or 10th Amendment, this also opens the door to State regulation of gun ownership — something these anti-Federal government, States-rights gun advocates paradoxically want to avoid (opposition to the Chicago and DC handgun bans illustrate this point). They want to have their cake and eat it too, where the Federal government and gun rights intersect.
Now, we are all accustomed to restrictions put on legal products to which we otherwise feel we have a right to purchase: there is a minimum drinking age for alcohol, all cars must drive on the same side of the street and stop at red lights, all electronic devices must limit their electromagnetic interference. There are countless other instances which, under certain conditions, may represent an inconvenience, but which, on the whole, represent the needs of a greater social good.
But, by and large, the largest source of confusion and agitation on this point comes from the National Rifle Association. Started as a support group for Marines, the powerful lobby completely ignores that gun ownership today is a completely different proposition that it was in the early days of the Republic. The group’s position is completely a-historical.
In the Revolutionary era, guns were hand-made artisan items, not mass-produced commodities. Maybe one out of eight men owned a gun. As many as half these guns didn’t work, but were viewed as property nevertheless and passed on from generation to generation. They weren’t essential for hunting, as many Americans lived in cities or raised crops and livestock as farmers. There weren’t organized municipal police forces at the time. The original “individual mandate” from the Federal government required that able-bodied me purchase firearms due to a massive shortage.
Today, there is one gun per man, woman, and child in the country. It is a very different situation. And here, perhaps, lies the most atrocious facet of the NRA: it is little more than an industry trade group masquerading as a civil rights organization.