Search This Blog

Showing posts with label computing. Show all posts
Showing posts with label computing. Show all posts

Social Media… Amplifying The Inner Sheep

In the early days of computing, those involved were pioneers, innovators, original thinkers, the super-passionate, driven by curiosity and the purity of adventure. These were a strong people. Proud. And though it sounds precious to say so, the early computing pioneers were not unlike other gritty pioneers in that they enjoyed being out well beyond the known, striding forward where there was no trail, no strength in numbers, no peer support, no mentors, no accepted pattern.

Now? Well now that computers have become ubiquitous, now that everyone has one, uses one like they use a knife and fork, the pioneers have long since been replaced by Plumber Joe, by the middle of the bell curve, by everyman and everywoman and their kids.

The computer industry is market driven! I hate that it has taken me this long to recognize the importance and implications of this now overwhelmingly obvious and simple fact. The diffusion of computers into the daily routine of the entire population has resulted in a dramatic shift in the demand landscape that informs what computing becomes. The market for computing is its users. The user today, the average user, is a whole different animal than the user/creator that defined its early history. Those of us that jumped in early probably resist the idea that what we care about really doesn't matter anymore. Though it might be true that knowledge and a deeply theoretical bases for that knowledge still matters, from a consumer market demand perspective, we grey hairs are simply invisible. The fact that we nerds ever defined a market for anything at all is the more surprising historical footnote. It is a bittersweet realization, that success of our industry would of mean the marginalization of its founders.

In every grade school class I attended there were a few kids (one or two) who were driven by a passion to know, to understand, to create. The rest, well the rest excelled at a completely different set of skills, getting through the day, unnoticed, blending in. The two groups couldn't be more different. The inquisitive few were internally driven. The rest were driven by the outward demands of success as defined by the curve. The inventive minds competed against their own ability to invent. The rest competed amongst themselves over the coveted 60 percentile that would define passing the class.

The computing market is now dominated by that larger part of the human population that defines success as climbing (which ever way possible) on top of the 60 percent of the population (of other social climbers) that makes up the bulk and center of any bell curve. As kids, these were the people who spent most of their time comparing themselves to the kids next to them. Looking over their shoulder at the other kid's test answers. Studying together so that they knew the base line before they got to the actual test. I say "climb to the top" but the word "top" when describing a bell curve does a disservice to the real meaning of averages. What we call the top of a bell curve is really the center of a frequency distribution. Climbing to the top is really a struggle to get into the center. Like fish trying to avoid a shark, there is a natural human tendency away from being alone, away from the vulnerability that is the open water that is original ideas and behavior. As a result, we constantly seek the protection of others. Each of us, as humans, spend a good deal of our energy trying to determine and then contort our behavior to that which best describes the center of normative behavior and attitude.

The similarities between schooling fish and human socialization pressures are profound. But there is one important difference. Where fish seek the center to avoid the teeth and gut of another species, the predator we humans work so hard to avoid is us, is public ridicule, being seen as different, standing out! We are in a very real sense both sheep and the sheep dogs nipping at the the sheep's legs. It is obvious that evolutionary pressures have conspired to build into our brains at least two modes of processing and that they are, at least at times, antagonistic. One is a great big "what if?" simulator, a novelty machine… the other, a social manors restriction system that cautions at the point of pain, behavior the least bit novel or different.

I have traversed the usual nature/nurture, cultural/evolutionary minefields. What I come to is this; traits that exist universally across most cultures and experienced many times within each individual's life, are most probably behaviors that have a significant genetic/physical component... are common regardless of our developmental environment and experience. Humans are obviously capable of profound novelty and abstract pattern analysis. But there is also a pervasive behavioral overlay of social control of which we are simultaneously, willing participant, and victim. What is confounding is the extent to which each system interferes with the function and success of the other… and that they are so diametrically opposed.

With regard to schooling (and herding) behavior, that which we share in common with fish (and sheep), is an indifference to where the school is, in which direction it is moving, and how fast. Under the social threat that triggers schooling, all that matters is that each of us as individuals finds our way as close as possible to the center. Humans will go along with almost any plan so long as social grouping allows us to avoid being seen as different. Obvious examples: Nazi Germany, Slavery in the southern U.S., the Nanking Sex Trade, etc.

As computers have been made "user friendly" and as the cost of ownership has dropped, this center of the bell curve, this mad fight for self-similarity that defines who we are as a species, this reflection and homogenization of the greater us, has become the market for computing. Which makes sense. Diffusion and all. But the whole history of modern computing is so short, just 40 years now, that it is surprising and a bit of a shock to realize finally that from a market perspective, computing is a mature industry. How could an industry just a generation old have transited already from go-it-alone pioneer to "I'm Love'n It" average?

The implications are huge. In particular, this insight brings social media into sharp ironic focus. Social media brings to computing the same access to community monitoring and control that gossip and storytelling brought to the camp fire. It is to computing what cheating off of your neighbor's test is to being a kid. As a person who likes to think of himself as a pioneer, I have reacted in horror, disbelief, and frustration to what has looked like fifteen years of computer industry regression.

If you accept that computers, like minds, are abstraction (language) machines, then it makes sense to wonder to what extent the human brain has suffered the same evolutionary pressure towards social assimilation and at least plausibly, away from innovation and novelty. To what extent is the rarity of the product of profound human creativity a reflection of actual physical and logical limits on and causal costs to creativity itself, and to what extent is the same rarity a product of evolutionary molded physical properties of the brain that conspire to restrict the production of novelty as a result of even greater survival pressure to promote behaviors that honor social cohesion?

If the current overwhelming trend that sees the computer as more of a social monitoring mechanism, and less a creative tool, is a trend that reflects market demand, then the same questions I am asking of the market pressures that shape the machinery of social media must be asked of the cultural pressures that have through evolutionary time shaped the mind. So long as computation is primarily consumed by human beings, both computer and mind will be shaped by the same evolutionary pressures. As technical barriers are overcome, the industry can and does react more fluidly and with higher fidelity to the demands of its consumers.

At which point, the question becomes; Which heaps greater selective pressure on the evolution of computing, the need for tools that stand in for skills the human brain lacks, or the need for tools that amplify our most attention demanding desirers? Can the two co-exist and co-evolve productively? Again, the question is asked practically of computation and at the same time, philosophically or anthropologically of the human brain and the cultural stew in which each is both product and survival pressure.

Where computing used to take its shape from the fecund imagination of computational luminaries, it has of late been lead instead in the pursuit of the lizard brain in all of us, the sub-conscious urges and fears that inform social mediation behavior. The result is all of this social media dribble, the likes of "Twitter", "13 Seconds", "myface and spacebook" [sic], and numerology based pizza ordering "iPhone Apps." What advantages do such time wasters render? Some argue that social media was the logical communication oriented extension of email and personal web sites, that social media greases mechanisms deeply human and "natural". I remain dubious to these claims. I tend to group the brand of communication that social media seems to breed with more negative forms of group behavior like cults, mass hysteria, fundamentalism, and other behaviors unique to group-think.

And what of pop-culture notions like "collective intelligence", "global awakening, and "cultural consciousness" which seem to be born of transcendent utopian notions (not dissimilar to those that feed religion and spirituality). The adherents of these optimisms appear to be blissfully unhindered by the need for causally logical argument or empirical evidence. If our computers have become social monitoring devices (at the expense of facilities that enable creativity), is there a danger that they will further distort our already distorted sense of truth? If a computer "wants" to agree with us more than it wants to accurately calculate a value, then we might already have crossed the threshold into a world where 2 plus 2 really does equal 5 (if the computer says so, it must be true!).

It would be irresponsible for me not to at this point remind myself to question my own rhetorically close topics.

These questions and trends have profound implications to populist concepts we tend to romanticize but rarely examine in detail. Democracy, plural-icy, consensus, society, culture, community, equal rights, individuality, etc. As the computer industry becomes more and more sensitive to consumer demand, its product WILL become a device that does a better and better job at the automation and magnification of human idiosyncratic behavior, of superstition, mythos, hubris, rhetoric, ego, at any and all of the emotional side effects of evolutionary pressures.  Forget about the cold indifference of causal truth that has motivated so many sci-fi stories.

The real villain to be feared in any inevitable future is the computer as hubris amplifier.

Computing's new "social media" face might disturb my pioneer sensibilities, but it reflects the satisfaction of common demand. As any market matures it learns to overcome the physical and conceptual obstacles that so plagued it in its earlier years. Unburdened by things like processor speed and storage density, the computer industry was able to pursue directions more in line with human consumptive desire than with the technical or theoretical goals of computer "scientists". Marketeers trump scientists when the saturation of a product becomes universal.

It all makes sense. I am still depressed by the anti-innovation implications of the mass market dumbing down of computing, but at least I understand why it happened, what it means. Knowledge, even depressing knowledge, should open doors, should allow more efficient planning and prediction. But what exactly are the implications when a creative tool is hijacked by a larger urge to avoid at all costs, change and novelty. What happens when the same mass-market demand pressures that cause fads and trends focus their hysterical drive towards homogeneity onto the evolution of a tool originally intended for and idealized for creative exploration? What exactly do you get when you neuter the rebellion right out from underneath Picasso's brush?, When you force Darwin to teach sunday school?

Just what does it mean when our creative medium becomes sensitive to social interaction? Pen and paper never knew anything about the person wielding them, certainly didn't know how the greater society was reacting to what was being written or drawn.

If the average human feels more comfortable doing exactly what everyone else is doing, seeking the center, would much rather copy the answers off of their desk-mate's test than understand the the course content, well then it only makes sense that, we, the royal "we", would use this computing tool in the same way that we use the rest of the stuff in our lives, to help us find the social center and stay there.

It's not just the marketplace that has shifted towards the demographic center. The schooling mentality has crept into and now dominates computing as an industry. Personnel and management which in the early days of computing was awkwardly staffed by engineers and scientists and groupie hobbyists is now as diverse (homogenous?) a mix of humans as you could find in any industry. Even the scientists are cut from a different cloth. It takes a special and rare (crazy) human being to invent an industry from nothing. When avocations become well funded departments at major universities, the graduates are not likely to be as intellectually adventurous (understatement). As any MBA knows, the success of an industry is most sensitive to its ability to understand and predict the demand of its market. Who better to know the center of the consumer bell curve, the average Joe and Jane, than that same Joe and Jane? Joe and Jane Average now dominate the rank and file workers that make up the computer industry. This means administration, it also means sales and marketing, both of which make sense. Less intuitive, is but equally understandable, Joe and Jane Average have taken over the research and design and long range planning arms of the computer industry. Even where it isn't the actual Joe and Jane, it is people who do a kick-ass job of channeling them.

Does market saturation mean the evolution of computing has reached its zenith? I don't think it does. But, once an industry has reached this level of common adoption, the appearance of maturity and stability are hard to shake. Momentum and lock-in take hold. I have tried repeatedly to sell paradigm-busting and architectural re-imaginings of the entire computing paradigm to valley angles and big ticket VC firms only to realize that I wasn't selling to the current market. Try opening a VC pitch with "The problem with the computer industry is…" to a group of thirty-five year old billionaires who each drove to the meeting in custom ordered european super cars. Needless to say, their own rarefied experience makes it hard for them to connect with anything that follows. This is a classic conundrum in the study of evolution. I presume it is a classic conundrum facing evolution itself. Why should a system that is successful in the current environment ever spend any energy on alternative schemes? How could it? It is hard to even come to the question "What could we be doing better?" when surrounded by the luxury of success.

At the same time, it is unlikely (impossible?) that any system, no matter how successful in the present, will long define success in the future. It might even be true that the more successful a scheme, the more likely that scheme will hasten an end to the environment that supported it (through faster and more complete exploitation of available resources). The irony of success!

But we a are smart species. We are capable of having this discussion aren't we? So we might be prepared as well to discuss the inevitability of the end of the current computational scheme. No? To prepare as a result, for the next most likely scheme (or schemes)? Especially those of us who study language, information, computation, complex systems, evolution. Especially an industry that has so advanced the tools and infrastructure of complexity handling. No? Surely we in the computer industry are ideally situated to have a rational vantage from which to see beyond the success of the current scheme? Yet, for the reasons I have already postulated (market homogeneity and success blindness), and others, we seem to be directionless in the larger sense, incapable of the long range and big picture planning that might help us climb out of our little eden and into the larger future that is inevitable and unavoidable. Innovations springing forth from the current industry pale in comparison to those offered up 20, 30, even 40 years ago.

Speaking of which; I just found the resource list for Pattie Maes' "New Paradigms In Human Computer Interaction" class at MIT's Media Lab. These are video clips of speeches and demos of early and not so early computing pioneers showing off their work prescient work. Mind blowing. The future these folks (from places like MIT, Brown, Stanford Research, the Rand Corporation, Xerox PARC, and other institutions), well, it is sooooo much more forward looking than what has become of computing (or how the average computer is used). Everyone would do well to sit down and view or re-view these seminal projects in the context of their surprisingly early introduction.

I have written quite a few essays lambasting what I see as the computing industry's general loosing of its collective way… at the very least, a slowing down of the deep innovation that drove computing's early doers. Even when potentially powerful concepts ("Semantic Computing", "Geo-Tagging", "Augmented Reality", "User-Centered Cloud Computing") are (re-)introduced, their modern implementations are often so flawed and tainted by an obsession to kowtow to market pressures (or just plain lie or fake it) that the result is an insult, a blaspheme of the original concept being hijacked.

Over ten years ago, I gave a talk at Apple titled: "Apple Got Everything It Has For Free, And Sat On Its Ass For Eight Years While The Rest Of The World Caught Up".

Which is true, at least with regard to the Lisa/Macintosh which Xerox PARC handed them (check out the Star System) and the way they just kind of sat on the WISIWG mode-less graphical interaction scheme while other computer companies (Borland and then, reluctantly, Microsoft) eventually aped the same. At the time of my presentation, Apple had wasted its eight year lead extrapolating along the same scheme… a bigger badder WIMP interface, when they could have been introducing paradigm vaulting computational systems (that would put Seattle on another eight year chase).

But from a marketing perspective I couldn't have been more wrong. I have got to keep reminding myself that I no longer represent the market for computers! I wish I did, but I don't. I am an outlier, a small dot on a very long tail, I am pluto or maybe even just some wayward ice and dust comet to the big ordinary inner planets that trace out wonderfully near-circular orbits around the sun. In later presentations, I explained that Apple's critically acclaimed "Think Different" campaign and the elitist mindset from which it was derived, was the reason they had never garnered more than 2 or 3 percent of the computer market. I explained that Bill Gates' "genius" lie not in his profound insight, but in his ability to understand the motivations that drive the average person… namely to never be caught doing something that someone else could question. That means acting the same as everyone else. That means knowing how everyone else is acting. That means social media!

Nobody (other than wacky outliers like me) wants to be compared to iconoclasts like Einstein or Picasso or Gandhi or Gershwin. Very few people really want to "think different". Most people wouldn't be caught dead risking that type of public audacity. You have got to be pretty confident that you have an answer to the dreaded question "Why are you doing that?" to ever DO THAT (individually creative thing) in the first place.

Pioneers know exactly why they do what they do. They are driven by knowing more than others and by the excitement of being somewhere others haven't been… by being very much outside of the ball of fish that others seek as protection.

But if you want to sell a billion computers instead of just a few thousand, then you want to pay attention to the fish in all of us and not to the smiling and sock-less Einstein's on a bike.

But the larger and longer implications of mass market sensitivity are profound. While it is entirely true that paying attention to the center of the cultural bell curve will allow any industry to exploit more of the total available consumption potential, such behavior does not necessarily produce the paradigm jumping disruption upon which long term progress depends. If your Twinkies are selling really well, you might not notice that your customers are all reaching morbid levels of obesity and malnutrition or that the world is crumbling around the climate changing policies upon which your fast food empire is based.

The satisfaction of human center-of-the-fish-school urge is not necessarily the best recipe for success if success means more than short term market exploitation. In the long run, potential and satisfaction are decidedly not the same thing; they are, as a matter of fact, very often mutually antagonistic. Rome comes to mind. What comes to mind when I mention the phrase "dot com" or "mortgage backed securities" or "energy deregulation".

The mass market topology that has driven computing and communication towards a better and better fit with the demands of the largest and most homogenized consumer base the earth has ever witnessed could very likely work against the types of creative motivations that might be necessary to rescue us from the bland and the average, from the inward facing spiral of self-sameness that avarice alone yields. I am increasingly worried that the computer's seemingly endless potential to be warped and molded chameleon-like to perfectly satisfy our most basic evolutionary proclivities, to amplify unhindered, urges made strong against real scarcity in the natural environment, has already so distracted us within our egos and desires that we might not be able to pull our heads out before we get sucked completely down and into our very own John Malkovich-ian portals of endless identity-amplication.

And why does it matter? Because, like it or not, progress in every single field of human endeavor is now predicated on advancements in computation. More profound than anything else science has discovered about nature is that information is information. Information is domain agnostic. Which means that advances in information processing machinery benefit all information dependent fields of inquiry. Which is every field of inquiry! Which also implies that all fields of human inquiry are increasingly sensitive to the subtle ways that any one computational scheme effects information and the scope of information process-able within that scheme.

At any rate, it is alarming (if understandable) to me that the trending in this industry is towards a re-molding of the computer into ego amplifier, pleasure delivery system, truth avoidance device, distraction machine, Tribbles (as in "The Trouble With…"). The more insightful among us may want to place a side bet or two (if only as evolutionary insurance) on more expansive futures. Some of us are not so distracted by the shininess of these machines or by how they are getting better and better at focusing our attention at our own navels, to see futures for computing that are more expansive than the perfect amplification of the very human traits least likely to result in innovation or progress. There is still time (I hope).

Randall Reetz

Viewing Our Open Economy Throgh A Closed Economy Lookingglass

And have an idea. Plot the ratio of the total number of dollars in the stock market vs. the total number of dollars in the US economy sans the market (all over the same 100 year history).

My guess is that there have been spikes in this stock vs. M. ratio that correspond with new money entering the market, either shifted from other domestic segments (real-estate, retirement accounts, etc.) or from international influx of investments (rapidly rising wealth of Asia and rest of world, sudden collapse of a large industry, commodity, or governmental or regional stability). Very few of standard economic metrics measure true macro or global interaction between geo-scale segments.

Comparing the Dow Jones against itself over time, or rarely, against other metrics like the GDP is different than reifying this and other comparisons as named metrics in and of themselves. But completely missing is geo-scale metrics that track the shifting values across regions and segments, in effect treating economic entities as markets competing for the maximum percentage of the total global value or geo-M.

What I am getting to is some way to accurately read the total global M and total labor L and total energy use E and to track the motion in real time of the density of these values geographically, or geopolitically, or by production or consumption segment. Once a true global economic sandbox tracker/simulator has been built, we will have the ability to read the economy as it is, in real time, and find the factors that sit at the base of the actual influence hierarchy that drives and causes economic flux.

My suspicion is that actual economic growth is equivalent to, always and only the result of increases in the means and use of tools and infrastructure that can build more for less labor... productivity, and that, in the absence of true growth in productivity, a market responds through acts of trickery which are ultimately not supportable and always culminate in crashes or "adjustments" which tend to pull market values into closer alignment with actual M and productivity values.

When new money comes into a market, standard supply and demand metrics are no longer accurate predictors or models. When new money comes into a market, standard supply and demand metrics will tend to say that the value of a product has risen. In a closed system this evaluation would most often be accurate In an open system, where money can stream into a market not to chase a product for consumption or industry, but just because that market seems a more attractive place in which to speculate than others, it throws the whole system into instability. Producers under such conditions are want to make more shoes, even though people are not growing more feet or walking more holes into their souls. End-consumers begin at inflationary times like these, to speculate with their purchases, buying products and commodities not because they need them, but because it seems foolish not to. Products and commodities take on a currency-like property, and through over-valuation supersaturate the market... leading to an inevitable value crash. If a large enough percentage of an economy's value has been suckered into such a bubble, the crash will bleed over into the economy as a whole... hitting the financial and banking markets first and hardest.

Over the past two decades, three fundamental factors have made large markets in the west especially sensitive and vulnerable to these new-money boom/crash cycles.

First, the undeveloped economies of the world have experienced exponential growth as they adopt tools and infrastructures borrowed from the first world. Importantly, because the third and second world represented the vast majority of the world's population and geography, this explosion in wealth (though still on average only bringing the third world slightly out of poverty) began to represent (by shear size) a larger and larger segment of the global economy. Remember that this "rest of world" economy represents roughly six times the population of the first world. Even smallish changes to a segment of this relative multiple have huge effect on the total global economy. And the actual changes have not been small. Second and third world economies have absolutely exploded! Much of this new money has of course been reinvested into the local economies from which it sprung. But, increasingly, larger and larger chunks of this new money have gone searching for boutique markets like the New York Stock Exchange and its equivalent in Japan, Germany, England, France, and the EU.

The second factor has to do with the paucity of true growth in productivity experienced in western and first world economies during this same twenty or thirty year period. In post-industrial economies, regions that have secure and constant access to reliable transportation of goods and services (shipping, highway, rail, and air transport), ready and secure capital (through investment banking and business and consumer credit), private property ownership (as a means to secure capitalization), education (to steadily feed highly skilled workers into labor markets), and governments that protect and promote the well being and promote the success of their citizens en-mass, and who have built a dependable infrastructure to create, extract, and distribute energy, and the means to grow and process foods cheaply on industrial scales... these rare segments of global marketplace... have had these capabilities for some forty years. Excepting of course for incremental gains made in efficiency of the above systems as a result of new knowledge and tools that result from better understanding of nature through advances in the sciences productivity has largely leveled off and remained level for the better part of a quarter century. What of the computer? you say. Surely the computer and the World Wide Web have had a huge positive impact on first world economies. But interestingly, the net net economic effect of computation and the digital networks it creates, has been surpassingly neutral. We do pump a larger and larger percentage of first world moneys into computation and its infrastructure that consume and use computers.

Real productivity metrics have yet to precipitate outward from the large success of computing as a market and into the larger first world economy. Ironically, the computer industry's success in the west may have impacted second and third world economies the most. It may be that money made in the computing industry flowed more deeply and directly into the rest of world economies where much of the computer industry does its manufacturing, assembly, and customer support. That computing has not resulted in measurable increases in first world productivity has computer industry insiders scratching their heads. During the Dot Com boom, pro-industry annalists creatively sidestepped this uncomfortable truth by inventing the idea of a "new economy" or "cyber economy", famously proclaiming, "The old rules and metrics don't apply". They were wrong, in the short term, but maybe, just maybe, in the long term they will be correct.

I suspect that the true economic benefits or potential benefits caused by the computer and computing upon global productivity have yet to be realized. The computer industry has spent the last 30 years largely learning how to get computers to do what we did (though slower and more awkwardly) before we had computers (writing, printing, telephony, accounting, payroll, data processing, advertising, point of purchase, audio and video broadcast, mathematics, graphing, market tracking and trading, banking, news and reporting, information sharing, post and mail, libraries, process control, etc.). This conversion has been expensive, and time consuming. Much of the time, we have proceeded as an industry (and a society) without a clear goal. Let me restate; neither the computing industry or the consuming public has had a clear idea where computing has been or ought to be going. Much of the time, both industry and market have been happy just to see what new (old) thing the computer can be taught to do... blindly building and consuming our way into the future just because it is "cool" or "fun" or "neat" or adds some strange and intoxicating "immediacy" to our daily lives (even when that immediacy does not equate effectiveness or lead us to deeper and more efficient infrastructure's necessary to cause the kinds of profound increases in productivity we expect from new technology paradigms).

I am a big believer in the future of computing, or the future that computing could build towards, but this belief is contingent upon society getting to a deep clarity of understanding about what computing is and why it matters. We have got to work hard at determining the difference between that which is cool and that which is profound. That which we want and that which will change the world. Until then, we are simply designing and producing towards consumption which will make segments of the industry rich and will bring money from the consuming west into emerging economies, but it will not ultimately support real growth, the kind of growth that is supported by knowledge, tools, and infrastructure that have the capacity to catapult productivity to the next level (the way the tractor pulled plow, germ theory, general education, the steam engine, and electricity have done in the past).

[to be continued...]