Author Archives: Greg Satell

About Greg Satell

Greg Satell is a popular speaker and consultant. His latest book, Cascades: How to Create a Movement That Drives Transformational Change, is available now. Follow his blog at Digital Tonto or on Twitter @Digital Tonto.

You Must Accept That People Are Irrational

You Must Accept That People Are Irrational

GUEST POST from Greg Satell

For decades, economists have been obsessed with the idea of “enlightened self-interest,” building elaborate models based on the assumption that people make rational choices. Business and political leaders have used these models to shape competitive strategies, compensation, tax policies and social services among other things.

It’s clear that the real world is far more complex than that. Consider the prisoner’s dilemma, a famous thought experiment in which individuals acting in their self-interest make everyone worse off. In a wide array of real world and experimental contexts, people will cooperate for the greater good rather than pursue pure self-interest.

We are wired to cooperate as well as to compete. Identity and dignity will guide our actions even more than the prospect for loss or gain. While business schools have trained generations of managers to assume that they can optimize results by designing incentives, the truth is that leaders that can forge a sense of shared identity and purpose have the advantage.

Overcoming The Prisoner’s Dilemma

John von Neumann was a frustrated poker player. Despite having one of the best mathematical minds in history that could probably calculate the odds better than anyone on earth, he couldn’t tell whether other players were bluffing or not. It was his failure at poker that led him to create game theory, which calculates the strategies of other players.

As the field developed, it was expanded to include cooperative games in which players could choose to collaborate and even form coalitions with each other. That led researchers at RAND to create the prisoner’s dilemma, in which two suspects are being interrogated separately and each offered a reduced sentence to confess.

Prisoner's Dilemma

Here’s how it works: If both prisoners cooperate with each other and neither confesses, they each get one year in prison on a lesser charge. If one confesses, he gets off scot-free, while his partner gets 5 years. If they both rat each other out, then they get three years each—collectively the worst outcome of all.

Notice how from a rational viewpoint, the best strategy is to defect. No matter what one guy does, the other one is better off ratting him out. If both pursue self-interest, they are made worse off. It’s a frustrating problem. Game theorists call it a Nash equilibrium—one in which nobody can improve their position by unilateral move. In theory, you’re basically stuck.

Yet in a wide variety of real-world contexts, ranging from the survival strategies of guppies to military alliances, cooperation is credibly maintained. In fact, there are a number of strategies that have proved successful in overcoming the prisoner’s dilemma. One, called tit-for-tat, relies on credible punishments for defections. Even more effective, however, is building a culture of shared purpose and trust.

Kin Selection And Identity

Evolutionary psychology is a field very similar to game theory. It employs mathematical models to explain what types of behaviors provide the best evolutionary outcomes. At first, this may seem like the utilitarian approach that economists have long-employed, but when you combine genetics with natural selection, you get some surprising answers.

Consider the concept of kin selection. From a purely selfish point of view, there is no reason for a mother to sacrifice herself for her child. However, from an evolutionary point of view, it makes perfect sense for parents to put their kids first. Groups who favor children are more likely to grow and outperform groups who don’t.

This is what Richard Dawkins meant when he called genes selfish. If we look at things from our genes’ point of view, it makes perfect sense for them to want us to sacrifice ourselves for children, who are more likely to be able to propagate our genes than we are. The effect would logically also apply to others, such as cousins, that likely carry our genes.

Researchers have also applied the concept of kin selection to other forms of identity that don’t involve genes, but ideas (also known as memes) in examples such as patriotism. When it comes to people or ideas we see as an important part of our identity, we tend to take a much more expansive view of our interests than traditional economic models would predict.

Cultures of Dignity

It’s not just identity that figures into our decisions, but dignity as well. Consider the ultimatum game. One player is given a dollar and needs to propose how to split it with another player. If the offer is accepted, both players get the agreed upon shares. If it is not accepted, neither player gets anything.

If people acted purely rationally, offers as low as a penny would be routinely accepted. After all, a penny is better than nothing. Yet decades of experiments across different cultures show that most people do not accept a penny. In fact, offers of less than 30 cents are routinely rejected as unfair because they offend people’s dignity and sense of self.

Results from ultimatum game are not uniform, but vary in different cultures and more recent research suggests why. In a study in which a similar public goods game was played it was found that cooperative—as well as punitive—behavior is contagious, spreading through three degrees of interactions, even between people who haven’t had any direct contact.

Whether we know it or not, we are constantly building ecosystems of norms that reward and punish behavior according to expectations. If we see the culture we are operating in as trusting and generous, we are much more likely to act collaboratively. However, if we see our environment as cutthroat and greedy, we’ll tend to model that behavior in the same way.

Forging Shared Identity And Shared Purpose

In an earlier age, organizations were far more hierarchical. Power rested at the top. Information flowed up, orders went down, work got done and people got paid. Incentives seemed to work. You could pay more and get more. Yet in today’s marketplace, that’s no longer tenable because the work we need done is increasingly non-routine.

That means we need people to do more than merely carry out tasks, they need to put all of their passion and creativity into their work to perform at a high-level. They need to collaborate effectively in teams and take pride in the impact their efforts produce. To achieve that at an organizational level, leaders need to shift their mindsets.

As David Burkus explained in his TED Talk, humans are prosocial. They are vastly more likely to perform when they understand and identify with who their work benefits than when they are given financial incentives or fed some grandiose vision. Evolutionary psychologists have long established that altruism is deeply embedded in our sense of tribe.

The simple truth is that we can no longer coerce people to do what we want with Rube Goldberg-like structures of carrots and sticks, but must inspire people to want what we want. Humans are not purely rational beings, responding to stimuli as if they were vending machines that spit out desired behaviors when the right buttons are pushed, but are motivated by identity and dignity more than anything else.

Leadership is not an algorithm, but a practice of creating meaning through relationships of trust in the context of a shared purpose.

— Article courtesy of the Digital Tonto blog
— Image credit: Pixabay

Subscribe to Human-Centered Change & Innovation WeeklySign up here to join 17,000+ leaders getting Human-Centered Change & Innovation Weekly delivered to their inbox every week.

Learning Business and Life Lessons from Monkeys

Learning Business and Life Lessons from Monkeys

GUEST POST from Greg Satell

Franz Kafka was especially skeptical about parables. “Many complain that the words of the wise are always merely parables and of no use in daily life,” he wrote. “When the sage says: ‘Go over,’ he does not mean that we should cross to some actual place… he means some fabulous yonder…that he cannot designate more precisely, and therefore cannot help us here in the very least.

Business pundits, on the other hand, tend to favor parables, probably because telling simple stories allows for the opportunity to seem both folksy and wise at the same time. When Warren Buffet says “Only when the tide goes out do you discover who’s been swimming naked,” it doesn’t sound so much like an admonishment.

Over the years I’ve noticed that some of the best business parables involve monkeys. I’m not sure why that is, but I think it has something to do with taking intelligence out of the equation. We’re often prone to imagining ourselves as the clever hero of our own story and we neglect simple truths. That may be why monkey parables have so much to teach us.

1. Build The #MonkeyFirst

When I work with executives, they often have a breakthrough idea they are excited about. They begin to tell me what a great opportunity it is and how they are perfectly positioned to capitalize on it. However, when I begin to dig a little deeper it appears that there is some major barrier to making it happen. When I try to ask about it, they just shut down.

One reason that this happens is that there is a fundamental tension between innovation and operations. Operational executives tend to focus on identifying clear benchmarks to track progress. That’s fine for a typical project, but when you are trying to do something truly new and different, you have to directly confront the unknown.

At Google X, the tech giant’s “moonshot factory,” the mantra is #MonkeyFirst. The idea is that if you want to get a monkey to recite Shakespeare on a pedestal, you start by training the monkey, not building the pedestal, because training the monkey is the hard part. Anyone can build a pedestal.

The problem is that most people start with the pedestal, because it’s what they know and by building it, they can show early progress against a timeline. Unfortunately, building a pedestal gets you nowhere. Unless you can actually train the monkey, working on the pedestal is wasted effort.

The moral: Make sure you address the crux of the problem and don’t waste time with peripheral issues.

2. Don’t Get Taken In By Coin Flipping Monkeys

We live in a world that worships accomplishment. Sports stars who have never worked in an office are paid large fees to speak to corporate audiences. Billionaires who have never walked a beat speak out on how to fight crime (even as they invest in gun manufacturers). Others like to espouse views on education, although they have never taught a class.

Many say that you can’t argue with success, but consider this thought experiment: Put a million monkeys in a coin flipping contest. The winners in each round win a dollar and the losers drop out. After twenty rounds, there will only be two monkeys left, each winning $262,144. The vast majority of the other monkeys leave with merely pocket change.

How much would you pay the winning monkeys to speak at your corporate event? Would you invite them to advise your company? Sit on your board? Would you be interested in their views about how to raise your children, invest your savings or make career choices? Would you try to replicate their coin-flipping success? (Maybe it’s all in the wrist).

The truth is that chance and luck play a much bigger part in success than we like to admit. Einstein, for example, became the most famous scientist of the 20th century not just because of his discoveries but also due to an unlikely coincidence. True accomplishment is difficult to evaluate, so we look for signals of success to guide our judgments.

The moral: Next time you judge someone, either by their success or lack thereof, ask yourself whether you are judging actual accomplishment or telltale signs of successful coin flipping. It’s harder to tell the difference than you’d think.

3. The Infinite Monkey Theorem

There is an old thought experiment called the Infinite Monkey Theorem, which is eerily disturbing. The basic idea is that if there were an infinite amount of monkeys pecking away on an infinite amount of keyboards they would, in time, produce the complete works of Shakespeare, Tolstoy and every other literary masterpiece.

It’s a perplexing thought because we humans pride ourselves on our ability to recognize and evaluate patterns. The idea that something we value so highly could be randomly generated is extremely unsettling. Yet there is an entire branch of mathematics, called Ramsey Theory, devoted to the study of how order emerges from random sets of data.

While the infinite monkey theorem is, of course, theoretical, technology is forcing us to confront the very real dilemma’s it presents. For example, music scholar and composer David Cope has been able to create algorithms that produce original works of music that are so good even experts can’t tell they are computer generated. So what is the value of human input?

The moral: Much like the coin flipping contest, the infinite monkey theorem makes us confront what we value and why. What is the difference between things human produced and identical works that are computer generated? Are Tolstoy’s words what give his stories meaning? Or is it the intent of the author and the fact that a human was trying to say something important?

Imagining Monkeys All Around Us

G. H. Hardy, widely considered a genius, wrote that “For any serious purpose, intelligence is a very minor gift.” What he meant was that even in purely intellectual pursuits, such as his field of number theory, there are things that are far more important. It was, undoubtedly, intellectual humility that led Hardy to Ramanujuan, perhaps his greatest discovery of all.

Imagining ourselves to be heroes of our own story can rob us of the humility we need to succeed and prosper. Mistaking ourselves for geniuses can often get us into trouble. People who think they’re playing it smart tend to make silly mistakes, both because they expect to see things that others don’t and because they fail to look for and recognize trouble signs.

Parables about monkeys can be useful because nobody expects them to be geniuses, which demands that we ask ourselves hard questions. Are we doing the important work, or the easiest tasks to show progress on? If monkeys flipping coins can simulate professional success, what do we really celebrate? If monkeys tapping randomly on typewriters can create masterworks, what is the value of human agency?

The truth is that humans are prone to be foolish. We are unable, outside a few limited areas of expertise, to make basic distinctions in matters of importance. So we look for signals of prosperity, intelligence, shared purpose and other things we value to make judgments about what information we should trust. Imagining monkeys around us helps us to be more careful.

Sometimes the biggest obstacle between where we are now and the fabulous yonder we seek is just the few feet in front of us.

— Article courtesy of the Digital Tonto blog
— Image credit: Flickr

Subscribe to Human-Centered Change & Innovation WeeklySign up here to join 17,000+ leaders getting Human-Centered Change & Innovation Weekly delivered to their inbox every week.

Identity is Crucial to Change

Identity is Crucial to Change

GUEST POST from Greg Satell

In an age of disruption, the only viable strategy is to adapt. Today, we are undergoing major shifts in technology, resources, migration and demography that will demand that we make changes in how we think and what we do. The last time we saw this much change afoot was during the 1920s and that didn’t end well. The stakes are high.

In a recent speech, the EU’s High Representative for Foreign Affairs and Security Policy Josep Borrell highlighted the need for Europe to change and adapt to shifts in the geopolitical climate. He also pointed out that change involves far more than interests and incentives, carrots and sticks, but even more importantly, identity.

“Remember this sentence,” he said. “’It is the identity, stupid.’ It is no longer the economy, it is the identity.” What he meant was that human beings build attachments to things they identify with and, when those are threatened, they are apt to behave in a visceral, reactive and violent way. That’s why change and identity are always inextricably intertwined.

“We can’t define the change we want to pursue until we define who we want to be.” — Greg Satell

The Making Of A Dominant Model

Traditional models come to us with such great authority that we seldom realize that they too once were revolutionary. We are so often told how Einstein is revered for showing that Newton’s mechanics were flawed it is easy to forget that Newton himself was a radical insurgent, who rewrote the laws of nature and ushered in a new era.

Still, once a model becomes established, few question it. We go to school, train for a career and hone our craft. We make great efforts to learn basic principles and gain credentials when we show that we have grasped them. As we strive to become masters of our craft we find that as our proficiency increases, so does our success and status.

The models we use become more than mere tools to get things done, but intrinsic to our identity. Back in the nineteenth century, the miasma theory, the notion that bad air caused disease, was predominant in medicine. Doctors not only relied on it to do their job, they took great pride in their mastery of it. They would discuss its nuances and implications with colleagues, signaling their membership in a tribe as they did.

In the 1840s, when a young doctor named Ignaz Semmelweis showed that doctors could prevent infections by washing their hands, many in the medical establishment were scandalized. First, the suggestion that they, as men of prominence, could spread something as dirty as disease was insulting. Even more damaging, however, was the suggestion that their professional identity was, at least in part, based on a mistake.

Things didn’t turn out well for Semmelweis. He railed against the establishment, but to no avail. He would eventually die in an insane asylum, ironically of an infection he contracted under care, and the questions he raised about the prevailing miasma paradigm went unanswered.

A Gathering Storm Of Accumulating Evidence

We all know that for every rule, there are exceptions and anomalies that can’t be explained. As the statistician George Box put it, “all models are wrong, but some are useful.” The miasma theory, while it seems absurd today, was useful in its own way. Long before we had technology to study bacteria, smells could alert us to their presence in unsanitary conditions.

But Semmelweis’s hand-washing regime threatened doctors’ view of themselves and their role. Doctors were men of prominence, who saw disease emanating from the smells of the lower classes. This was more than a theory. It was an attachment to a particular view of the world and their place in it, which is one reason why Semmelweis experienced such backlash.

Yet he raised important questions and, at least in some circles, doubts about the miasma theory continued to grow. In 1854, about a decade after Semmelweis instituted hand washing, a cholera epidemic broke out in London and a miasma theory skeptic named John Snow was able to trace the source of the infection to a single water pump.

Yet once again, the establishment could not accept evidence that contradicted its prevailing theory. William Farr, a prominent medical statistician, questioned Snow’s findings. Besides, Snow couldn’t explain how the water pump was making people sick, only that it seemed to be the source of some pathogen. Farr, not Snow, won the day.

Later it would turn out that a septic pit had been dug too close to the pump and the water had been contaminated with fecal matter. But for the moment, while doubts began to grow about the miasma theory, it remained the dominant model and countless people would die every year because of it.

Breaking Through To A New Paradigm

In the early 1860s, as the Civil War was raging in the US, Louis Pasteur was researching wine-making in France. While studying the fermentation process, he discovered that microorganisms spoiled beverages such as beer and milk. He proposed that they be heated to temperatures between 60 and 100 degrees Celsius to avoid spoiling, a process that came to be called pasteurization

Pasteur guessed that the similar microorganisms made people sick which, in turn, led to the work of Robert Koch and Joseph Lister. Together they would establish the germ theory of disease. This work then led to not only better sanitary practices, but eventually to the work of Alexander Fleming, Howard Florey and Ernst Chain and development of antibiotics.

To break free of the miasma theory, doctors needed to change the way they saw themselves. The miasma theory had been around since Hippocrates. To forge a new path, they could no longer be the guardians of ancient wisdom, but evidence-based scientists, and that would require that everything about the field be transformed.

None of this occurred in a vacuum. In the late 19th century, a number of long-held truths, from Euclid’s Geometry to Aristotle’s logic, were being discarded, which would pave the way for strange new theories, such as Einstein’s relativity and Turing’s machine. To abandon these old ideas, which were considered gospel for thousands of years, was no doubt difficult. Yet it was what we needed to do to create the modern world.

Moving From Disruption to Resilience

Today, we stand on the precipice of a new paradigm. We’ve suffered through a global financial crisis, a pandemic and the most deadly conflict in Europe since World War II. The shifts in technology, resources, migration and demography are already underway. The strains and dangers of these shifts are already evident, yet the benefits are still to come.

To successfully navigate the decade ahead, we must make decisions not just about what we want, but who we want to be. Nowhere is this playing out more than in Ukraine right now, where the war being waged is almost solely about identity. Russians want to deny Ukrainian identity and to defy what they see as the US-led world order. Europeans need to take sides. So do the Chinese. Everyone needs to decide who they are and where they stand.

This is not only true in international affairs, but in every facet of society. Different eras make different demands. The generation that came of age after World War II needed to rebuild and they did so magnificently. Yet as things grew, inefficiencies mounted and the Boomer Generation became optimizers. The generations that came after worshiped disruption and renewal. These are, of course, gross generalizations, but the basic narrative holds true.

What should be clear is that where we go from here will depend on who we want to be. My hope is that we become protectors who seek to make the shift from disruption to resilience. We can no longer simply worship market and technological forces and leave our fates up to them as if they were gods. We need to make choices and the ones we make will be greatly influenced by how we see ourselves and our role.

As Josep Borrell so eloquently put it: It is the identity, stupid. It is no longer the economy, it is the identity.

— Article courtesy of the Digital Tonto blog
— Image credit: Unsplash

Subscribe to Human-Centered Change & Innovation WeeklySign up here to join 17,000+ leaders getting Human-Centered Change & Innovation Weekly delivered to their inbox every week.

What We See Influences How We’ll Act

What We See Influences How We'll Act

GUEST POST from Greg Satell

“Practical men, who believe themselves to be quite exempt from any intellectual influences, are usually slaves of some defunct economist,” John Maynard Keynes, himself a long dead economist, once wrote. We are, much more than we’d like to admit, creatures of our own age, taking our cues from our environment.

That’s why we need to be on the lookout for our own biases. The truth, as we see it, is often more of a personalized manifestation of the zeitgeist than it is the product of any real insight or reflection. As Richard Feynman put it, “The first principle is that you must not fool yourself—and you are the easiest person to fool. So you have to be very careful about that.”

We can’t believe everything we think. We often seize upon the most easily available information, rather than the most reliable sources. We then seek out information that confirms those beliefs and reject evidence that contradicts existing paradigms. That’s what leads to bad decisions. If what we see determines how we act, we need to look carefully.

The Rise And Fall Of Social Darwinism

In the 1860s, in response to Darwin’s ideas, Herbert Spencer and others began promoting the theory of Social Darwinism. The basic idea was that “survival of the fittest” meant that society should reflect a Hobbesian state of nature, in which most can expect a life that is “nasty, brutish and short,” while an exalted few enjoy the benefits of their superiority.

This was, of course, a gross misunderstanding of Darwin’s work. First, Darwin never used the term, “survival of the fittest,” which was actually coined by Spencer himself. Secondly, Darwin never meant to suggest that there are certain innate qualities that make one individual better than others, but that as the environment changes, certain traits tend to be propagated which, over time, can lead to a new species.

Still, if you see the world as a contest for individual survival, you will act accordingly. You will favor a laissez-faire approach to society, punishing the poor and unfortunate and rewarding the rich and powerful. In some cases, such as Nazi Germany and in the late Ottoman empire, Social Darwinism was used as a justification for genocide.

While some strains of Social Darwinism still exist, for the most part it has been discredited, partly because of excesses such as racism, eugenics and social inequality, but also because more rigorous approaches, such as evolutionary psychology, show that altruism and collaboration can themselves be adaptive traits.

The Making Of The Modern Organization

When Alfred Sloan created the modern corporation at General Motors in the early 20th century, what he really did was create a new type of organization. It had centralized management, far flung divisions and was exponentially more efficient at moving around men and material than anything that had come before.

He called it “federal decentralization.” Management would create operating principles, set goals and develop overall strategy, while day-to-day decisions were performed by people lower down in the structure. While there was some autonomy, it was more like an orchestra than a jazz band, with the CEO as conductor.

Here again, what people saw determined how they acted. Many believed that a basic set of management principles, if conceived and applied correctly, could be adapted to any kind of business, which culminated in the “Nifty Fifty” conglomerates of the 60’s and 70’s. It was, in some sense, an idea akin to Social Darwinism, implying that there are certain innate traits that make an organization more competitive.

Yet business environments change and, while larger organizations may be able to drive efficiencies, they often find it hard to adapt to changing conditions. When the economy hit hard times in the 1970s, the “Nifty Fifty” stocks vastly under-performed the market. By the time the 80s rolled around, conglomerates had fallen out of fashion.

Industries and Value Chains

In 1985, a relatively unknown professor at Harvard Business School named Michael Porter published a book called Competitive Advantage, which explained that by optimizing every facet of the value chain, a firm could consistently outperform its competitors. The book was an immediate success and made Porter a management superstar.

Key to Porter’s view was that firms compete in industries that are shaped by five forces: competitors, customers, suppliers, substitutes, and new market entrants. So he advised leaders to build and leverage bargaining power in each of those directions to create a sustainable competitive advantage for the long term.

If you see your business environment as being neatly organized in specific industries, everybody is a potential rival. Even your allies need to be viewed with suspicion. So, for example, when a new open source operating system called Linux appeared, Microsoft CEO Steve Ballmer considered it to be a threat and immediately attacked, calling it a cancer.

Yet even as Ballmer went on the attack, the business environment was changing. As the internet made the world more connected, technology companies found that leveraging that connectivity through open source communities was a winning strategy. Microsoft’s current CEO, Satya Nadella, says that the company loves Linux. Ultimately, it recognized that it couldn’t continue to shut itself out and compete effectively.

Looking To The Future

Take a moment to think about what the world must have looked like to J.P. Morgan a century ago, in 1922. The disruptive technologies of the day, electricity and internal combustion, were already almost 40 years old, but had little measurable economic impact. Life largely went on as it always had and the legendary financier lorded over his domain of corporate barons.

That would quickly change over the next decade when those technologies would gain traction, form ecosystems and drive a 50-year boom. The great “trusts” that he built would get broken up and by 1930 virtually all of them would be dropped as components of the Dow Jones Industrial average. Every face of life would be completely transformed.

We’re at a similar point today, on the brink of enormous transformation. The recent string of calamities, including a financial meltdown, a pandemic and the deadliest war in Europe in 80 years, demand that we take a new path. Powerful shifts in technology, demographics, resources and migration, suggest that even more disruption may be in our future.

The course we take from here will be determined by how we see the world we live in. Do we see our fellow citizens as a burden or an asset? Are new technologies a blessing or a threat? Is the world full of opportunities to be embraced or dangers we need to protect ourselves from? These are questions we need to think seriously about.

How we answer them will determine what comes next.

— Article courtesy of the Digital Tonto blog
— Image credit: Unsplash

Subscribe to Human-Centered Change & Innovation WeeklySign up here to join 17,000+ leaders getting Human-Centered Change & Innovation Weekly delivered to their inbox every week.

Why Big Ideas Often Fail to Survive Victory

Why Big Ideas Often Fail To Survive Victory

GUEST POST from Greg Satell

I still vividly remember a whiskey drinking session I had with a good friend in my flat in Kyiv in early 2005, shortly after the Orange Revolution had concluded. We were discussing what would come after and, knowing that I had lived in Poland during years of reform, he was interested in my opinion about the future. I told him NATO and EU ascension was the way to go.

My friend, a prominent journalist, disagreed. He thought that Ukraine should pursue a “Finnish model,” in which it would pursue good relations with both Russia and the west, favoring neither. As he saw it, the Ukrainian people, who had just been through months of political turmoil, should pursue a “third way” and leave the drama behind.

As it turned out, we were both wrong. The promise of change would soon turn to nightmare, ending with an evil, brutal regime and a second Ukrainian revolution a decade later. I would later find that this pattern is so common that there is even a name for it: the failure to survive victory. To break the cycle you first need to learn to anticipate it and then to prepare for it.

The Thrill Of A New Direction And An Initial Success

In the weeks after the Orange Revolution I happened to be in Warsaw and saw a huge banner celebrating democracy movements in Eastern Europe, with Poland’s Solidarity movement as the first and Ukraine’s Orange revolution as the last in the series. Everyone thought that Ukraine would follow its neighbor into peace and prosperity.

We were triumphant and it seemed like the forces of history were on our side. That’s one reason why we failed to see the forces that were gathering. Despite our enthusiasm, those who opposed our cause didn’t just melt away and go home. In fact they redoubled their efforts to undermine what we had achieved. We never really saw it coming.

I see the same thing in my work with organizational transformations. Once people get a taste of that initial success—they win executive sponsorship for their initiative, get a budget approved or even achieve some tangible progress on the ground—they think it will all get easier. It never does. In fact, it usually gets harder.

Make no mistake. Opposition doesn’t erupt in spite of an early success, but because of it. A change initiative only becomes a threat to the status quo when it begins to gain traction. That’s when the knives come out and, much like my friend and I after the Orange Revolution, most people working to bring about change are oblivious to it.

If you are working for a change that you believe in passionately, chances are you’re missing a brewing storm. Almost everyone does the first time around (and many never learn to recognize it).

Propagating Echo Chambers

One of the reasons we failed to see trouble brewing back then was that, as best we could tell, everyone around us saw things the same way we did. Whatever dissenting voices we did come across seemed like an aberration to us. Sure, some people were still stuck in the old ways, we thought, but with history on our side how could we fail?

Something similar happened in the wake of the George Floyd protests. The city council in Minneapolis, where the incident took place, voted to defund the police. Taking its cue, corporate America brought in armies of consultants to set out the new rules of the workplace. In one survey, 85% of CHRO’s said that they were expanding diversity and inclusion efforts. With such an outpouring of news coverage and emotion, who would dare to question them?

The truth is that majorities don’t just rule, they also influence in a number of ways. First, decades of studies show that we tend to conform to the views around us and that effect extends out to three degrees of relationships. Not only people we know, but the friends of their friends—most of whom we don’t even know—affect how we think.

It isn’t just what we hear but also what we say that matters. Research from MIT suggests that when we are around people we expect to agree with us, we’re less likely to check our facts and more likely to share information that isn’t true. That, in turn, impacts our informational environment, helping to create an echo chamber that reinforces our sense of certainty.

The Inevitable Backlash

Almost as soon as the new Ukrainian government took power in 2005, the opposition went on the offensive. While the new President, Viktor Yushchenko was seen positively, they attacked the people around him. His Prime Minister, Yulia Tymoshenko, was portrayed as a calculating and devious woman. When Yushchenko’s son got into trouble, questions were raised about corruption in his father’s administration.

A similar pattern took hold in the wake of the George Floyd protests. Calls for racial justice were portrayed as anti-police and law enforcement budgets across the country increased as “We Support Our Police” signs went up on suburban lawns. Critical Race Theory, an obscure legal concept rarely discussed outside of universities, became a political punching bag. Today, as layoffs increase, corporate diversity efforts are sure to take a hit.

These patterns are not exceptions. They are the rule. As Saul Alinsky pointed out, every revolution inspires a counter-revolution. That is the physics of change. Every reaction provokes a reaction. Every success impacts your environment and some of those changes will not be favorable to your cause. They will expose vulnerabilities that can be exploited by those who oppose your idea.

Yet Alinsky didn’t just identify the problem, he also pointed to a solution. “Once we accept and learn to anticipate the inevitable counter-revolution, we may then alter the historical pattern of revolution and counter-revolution from the traditional slow advance of two steps forward and one step backward to minimizing the latter,” he writes.

In other words, the key to surviving victory is to prepare for the backlash that is sure to come and build a strategy to overcome it.

Building A Shared Future Rooted In Shared Values

In the two decades I have been researching transformation and change, the failure to survive victory is probably the most consistent aspect of it. In fact, it is so common you can almost set your watch by it. Amazingly, no matter how many times change advocates experience it, they rarely see it coming. Many, in fact, seem to take pride in how many battles they have lost, seeing it as some kind of badge of honor.

The uncomfortable truth is that success doesn’t necessarily begat more success. Often it breeds failure. People mistake a moment for a movement and think that their time has finally come. Believing change to be inevitable, they get cocky and overconfident and miss the networks of unseen connections forming in opposition. They make sure to press a point, but fail to make a difference.

Lasting change always needs to be built on common ground. That’s what we failed to see all those years ago, when I began my journey. You can never base your revolution on any particular person, technology or policy. It needs to be rooted in shared values and if we truly care about change, we need to hold ourselves accountable to be effective messengers.

We can’t just preach to the choir. Sometimes we need to venture out of the church and mix with the heathens. We can be clear about where we stand and still listen to those who see things differently. That doesn’t mean we compromise. In fact, we should never compromise the values we believe in. What we can do, however, is identify common ground upon which to build a shared future.

These principles hold true whether the change you seek is in your organization, your industry, your community or throughout society as a whole. If you fail to learn and apply them, don’t be surprised when you fail to survive victory.

— Article courtesy of the Digital Tonto blog
— Image credit: Pexels

Subscribe to Human-Centered Change & Innovation WeeklySign up here to join 17,000+ leaders getting Human-Centered Change & Innovation Weekly delivered to their inbox every week.

Sometimes Ancient Wisdom Needs to be Left Behind

Sometimes Ancient Wisdom Needs to be Left Behind

GUEST POST from Greg Satell

I recently visited Panama and learned the incredible story of how the indigenous Emberá people there helped to teach jungle survival skills to Apollo mission astronauts. It is a fascinating combining and contrast of ancient wisdom and modern technology, equipping the first men to go to the moon with insights from both realms.

Humans tend to have a natural reverence for old wisdom that is probably woven into our DNA. It stands to reason that people more willing to stick with the tried and true might have a survival advantage over those who were more reckless. Ideas that stand the test of time are, by definition, the ones that worked well enough to be passed on.

Paradoxically, to move forward we need to abandon old ideas. It was only by discarding ancient wisdoms that we were able to create the modern world. In much the same way, to move forward now we’ll need to debunk ideas that qualify as expertise today. As in most things, our past can help serve as a guide. Here are three old ideas we managed to transcend.

1. Euclid’s Geometry

The basic geometry we learn in grade school, also known as Euclidean geometry, is rooted in axioms observed from the physical world, such as the principle that two parallel lines never intersect. For thousands of years mathematicians built proofs based on those axioms to create new knowledge, such as how to calculate the height of an object. Without these insights, our ability to shape the physical world would be negligible.

In the 19th century, however, men like Gauss, Lobachevsky, Bolyai and Riemann started to build new forms of non-Euclidean geometry based on curved spaces. These were, of course, completely theoretical and of no use in daily life. The universe, as we experience it, doesn’t curve in any appreciable way, which is why police ask us to walk a straight line if they think we’ve been drinking.

But when Einstein started to think about how gravity functioned, he began to suspect that the universe did, in fact, curve over large distances. To make his theory of general relativity work he had to discard the old geometrical thinking and embrace new mathematical concepts. Without those critical tools, he would have been hopelessly stuck.

Much like the astronauts in the Apollo program, we now live in a strange mix of old and new. To travel to Panama, for example, I personally moved through linear space and the old Euclidean axioms worked perfectly well. However, to navigate, I had to use GPS, which must take into account curved spaces for Einstein’s equations to correctly calculate distances between the GPS satellites and points on earth.

2. Aristotle’s Logic

In terms of longevity and impact, only Aristotle’s logic rivals Euclid’s geometry. At the core of Aristotle’s system is the syllogism, which is made up of propositions that consist of two terms (a subject and a predicate). If the propositions in the syllogism are true, then the argument has to be true. This basic notion that conclusions follow premises imbues logical statements with a mathematical rigor.

Yet much like with geometry, scholars began to suspect that there might be something amiss. At first, they noticed minor flaws that had to do with a strange paradox in set theory which arose with sets that are members of themselves. For example, if the barber who shaves everyone in town who doesn’t shave themselves, then who shaves the barber?

At first, these seemed like strange anomalies, minor exceptions to rules that could be easily explained away. Still, the more scholars tried to close the gaps, the more problems appeared, leading to a foundational crisis. It would only be resolved when a young logician named Kurt Gödel published his theorems that proved logic, at least as we knew it, is hopelessly broken.

In a strange twist, another young mathematician, Alan Turing, built on Gödel’s work to create an imaginary machine that would make digital computers possible. In other words, in order for Silicon Valley engineers to code to create logical worlds online, they need to use machines built on the premise that perfectly logical systems are inherently unworkable.

Of course, as I write this, I am straddling both universes, trying to put build logical sentences on those very same machines.

3. The Miasma Theory of Disease

Before the germ theory of disease took hold in medicine, the miasma theory, the notion that bad air caused disease, was predominant. Again, from a practical perspective this made perfect sense. Harmful pathogens tend to thrive in environments with decaying organic matter that gives off bad smells. So avoiding those areas would promote better health.

Once again, this basic paradigm would begin to break down with a series of incidents. First, a young doctor named Ignaz Semmelweis showed that doctors could prevent infections by washing their hands, which suggested that something besides air carried disease. Later John Snow was able to trace the source of a cholera epidemic to a single water pump.

Perhaps not surprisingly, these were initially explained away. Semmelweis failed to format his data properly and was less than an effective advocate for his work. John Snow’s work was statistical, based on correlation rather than causality. A prominent statistician William Farr, who supported the miasma theory, argued for an alternative explanation.

Still, as doubts grew, more scientists looked for answers. The work of Robert Koch, Joseph Lister and Louis Pasteur led to the germ theory. Later, Alexander Fleming, Howard Florey and Ernst Chain would pioneer the development of antibiotics in the 1940s. That would open the floodgates and money poured into research, creating modern medicine.

Today, we have gone far beyond the germ theory of disease and even lay people understand that disease has myriad causes, including bacteria, viruses and other pathogens, as well as genetic diseases and those caused by strange misfolded proteins known as prions.

To Create The Future, We Need To Break Free Of The Past

If you were a person of sophistication and education in the 19th century, your world view was based on certain axiomatic truths, such as parallel lines never cross, logical propositions are either true or false and “bad airs” made people sick. For the most part, these ideas would have served you well for the challenges you faced in daily life.

Even more importantly, your understanding of these concepts would signal your inclusion and acceptance into a particular tribe, which would confer prestige and status. If you were an architect or engineer, you needed to understand Euclid’s geometric axions. Aristotle’s rules of logic were essential to every educated profession. Medical doctors were expected to master the nuances of the miasma theory.

To stray from established orthodoxies carries great risk, even now. It is no accident that those who were able to bring about new paradigms, such as Einstein, Turing and John Snow, came from outside the establishment. More recently, people like Benoit Mandelbrot, Jim Allison and Katalin Karikó had to overcome fierce resistance to bring new ways of thinking to finance, cancer immunotherapy and mRNA vaccines respectively.

Today, it’s becoming increasingly clear we need to break with the past. In just over a decade, we’ve been through a crippling financial crisis, a global pandemic, deadly terrorist attacks, and the biggest conflict in Europe since World War II. We need to confront climate change and a growing mental health crisis. Yet it is also clear that we can’t just raze the global order to the ground and start all over again.

So what do we leave in the past and what do we bring with us into the future? Which new lessons do we need to learn and which old ones do we need to unlearn? Perhaps most importantly, what do we need to create anew and what can we rediscover in the ancient?

Throughout history, we have learned that the answer lies not in merely speculating about ideas, but in finding real solutions to problems we face.

— Article courtesy of the Digital Tonto blog
— Image credit: 1 of 950+ FREE quote slides from http://misterinnovation.com

Subscribe to Human-Centered Change & Innovation WeeklySign up here to join 17,000+ leaders getting Human-Centered Change & Innovation Weekly delivered to their inbox every week.

Strategy Lacking Purpose Will Always Fail

Strategy Lacking Purpose Will Always Fail

GUEST POST from Greg Satell

In 1989, just before the fall of the Berlin Wall, Francis Fukuyama published an essay in the journal The National Interest titled The End of History, which led to a bestselling book. Many took his argument to mean that, with the defeat of communism, US-style liberal democracy had emerged as the only viable way of organizing a society.

He was misunderstood. His actual argument was far more nuanced and insightful. After explaining the arguments of philosophers like Hegel and Kojeve, Fukuyama pointed out that even if we had reached an endpoint in the debate about ideologies, there would still be conflict because of people’s need to express their identity.

We usually think of strategy as a rational, analytic activity, with teams of MBA’s poring over spreadsheets or generals standing before maps. Yet if we fail to take into account human agency and dignity, we’re missing the boat. Strategy without purpose is doomed to fail, however clever the calculations. Leaders need to take note of that basic reality.

Taking Stock Of The Halo Effect

Business case studies are written by experienced professionals who are trained to analyze past situations from multiple perspectives. However, their ability to do that successfully is greatly limited by the fact that they already know the outcome of the situation they are studying. That can’t help but to color their analysis.

In The Halo Effect, Phil Rosenzweig explains how those perceptions can color conclusions. He points to the networking company Cisco during the dotcom boom. When it was flying high, it was said to have an unparalleled culture with people that worked long hours but loved every minute of it. When the market tanked, however, all of the sudden its culture came to be seen as “cocksure” and “naive.”

It is hard to see how a company’s culture could change so drastically in such a short amount of time, with no significant change in leadership. More likely, seeing Cisco’s success, analysts looked at particular qualities in a positive light. However, when things began to go the other way, those same qualities were perceived as negative.

When an organization is doing well, we may find its people to be “idealistic” and “values driven,” but when things go sour, those same traits come to be seen as “impractical” and “arrogant.” Given the same set of facts, we can—and often do—come to very different conclusions when our perception of the outcomes changes.

In most cases, analysts don’t have a stake in the outcome. From their point of view, they probably see themselves as objectively analyzing facts and following them to their most logical outcomes. Yet when the purpose for writing an analysis changes from telling a success story to lamenting a cautionary tale, their perception of events tends to change markedly.

Reassessing The Value Chain

For decades, the dominant view of business strategy was based on Michael Porter’s ideas about competitive advantage. In essence, he argued that the key to long-term success was to dominate the value chain by maximizing bargaining power among suppliers, customers, new market entrants and substitute goods.

Yet as AnnaLee Saxenian explained in Regional Advantage, around the same time that Porter’s ideas were ascending among CEOs in the establishment industries on the east coast, a very different way of doing business was gaining steam in Silicon Valley. The firms there saw themselves not as isolated fiefdoms, but as part of a larger ecosystem.

The two models are built on very different assumptions. The Porter model sees the world as made up of transactions. Optimize your strategy to create efficiencies, derive the maximum value out of every transaction and you will build a sustainable competitive advantage. The Silicon Valley model, however, saw the world as made up of connections and optimized their strategies to widen and deepen linkages.

Microsoft is one great example of this shift. When Linux first rose to prominence, Microsoft CEO Steve Ballmer called it a cancer. Yet more recently, its current CEO announced that the company loves Linux. That didn’t happen out of any sort of newfound benevolence, but because it recognized that it couldn’t continue to shut itself out and still be able to compete.

When you see the world as the “sum of all efficiencies,” the optimal strategy is to dominate. However, if you see the world as made up of the “sum of all connections,” the optimal strategy is to attract. You need to be careful to be seen as purposeful rather than predatory.

The Naïveté Of The “Realists”

Since at least the times of Richelieu, foreign policy theorists have been enthralled by the concept of Realpolitik, the notion that world affairs are governed by interests, not ideological, moral or ethical considerations. Much like with Porter’s “competitive advantage,” strategy is treated as a series of transactions rather than relationships.

Rational calculation of interests is one of those ideas that seems pragmatic on the surface, but is actually hopelessly academic and unworkable in the real world. How do you identify the “interests” you are supposed to be basing your decisions on if not by considering what you value? And how do you assess your values without taking into account your beliefs, morals and ethics?

To understand how such “realism” goes awry, consider the prominent political scientist John Mearsheimer. In March, he gave an interview to The New Yorker in which he argued that, by failing to recognize Russia’s role and interests as a great power, the US had erred greatly in its support of Ukraine.

Yet it is clear now that the Russians were the ones who erred. First, they failed to recognize that the world would see their purpose as immoral. Second, they failed to recognize how their aggression would empower Ukraine’s sense of nationhood. Third, they did not see how Europe would come to regard economic ties with Russia to be against their interests.

Nothing you can derive from military or economic statistics will give you insight into human agency. Excel sheets may not be motivated by purpose, but people are.

Strategy Is Not A Game Of Chess

Antonio Damasio, a neuroscientist who researches decision making, became intrigued when one of his patients, a highly intelligent and professionally successful man named “Elliot,” suffered from a brain lesion that impaired his ability to experience emotion. It soon became clear that Elliot was unable to make decisions..

Elliot’s prefrontal cortex, which governs the executive function, was fully intact. His memory and ability to understand events were normal as well. He was, essentially, a completely rational being with normal cognitive function, but no emotions. The problem was that although Elliot could understand all the factors that would go into making a decision, he could not weigh them. Without emotions, all options were all essentially the same.

In the real world, strategy is not a game of chess, in which we move inert pieces around a board. While we can make rational assessments about various courses of action, ultimately people have to care about the outcome. For a strategy to be meaningful, it needs to speak to people’s values, hopes, dreams and ambitions.

A leader’s role cannot be merely to plan and direct action, but must be to inspire and empower belief in a common endeavor. That’s what widens and deepens the meaningful connections that can enable genuine transformation.

— Article courtesy of the Digital Tonto blog
— Image credit: Pixabay

Subscribe to Human-Centered Change & Innovation WeeklySign up here to join 17,000+ leaders getting Human-Centered Change & Innovation Weekly delivered to their inbox every week.

Three Strategies for Overcoming Change Resistance

Three Strategies for Overcoming Change Resistance

GUEST POST from Greg Satell

Max Planck’s work in physics changed the way we were able to see the universe. Still, even he complained that “A new scientific truth does not triumph by convincing its opponents and making them see the light, but rather because its opponents eventually die, and a new generation grows up that is familiar with it.”

For most transformational efforts we need to pursue, we simply don’t have that kind of time. To drive significant change we have to overcome staunch resistance. Unfortunately, most change management strategies assume that opposition can be overcome through communication efforts that are designed to persuade.

This assumes that resistance always has a rational basis and clearly that’s not true. We all develop emotional attachments to ideas. When we feel those are threatened, it offends our dignity, identity and sense of self. If we are going to overcome our most fervent opponents we don’t need a better argument, we need a strategy. Here are three approaches that work:

Strategy 1: Designate An Internal Red Team

Resistance is never monolithic. While some people have irrational attachments based on their sense of identity and dignity, others are merely skeptical. One key difference between these two groups is that the irrational resistors rarely voice their opposition, but try to quietly sabotage change. The rational skeptics, on the other hand, are much more eager to engage.

While these are different groups, they often interact with each other behind the scenes. In many cases, it is the active, irrational opposition that is fueling the skeptics’ doubts. One useful strategy for dealing with this dynamic is to co-opt the opposition by setting up an internal red team to channel skepticism in a constructive way.

Red-teaming is a process in which an adversarial team is set up to poke holes in an operational or strategic plan. For example, red teams are used in airports and computer systems to see if they can find weaknesses in security. The military uses red teams to test battle plans. Perhaps most famously, a red team was used to help determine whether the conclusions that led to the raid on Osama bin Laden’s hideout were valid or if there was some other explanation.

Recruiting skeptics to be an internal red team provides two benefits. First, they can alert you to actual problems with your ideas, which you can then fix. Second, they not only voice their own objections, but also bring those of the irrational opposition out into the open (remember, irrational resisters rarely speak out.)

What’s key here is to make the distinction between rational skeptics and the irrational saboteurs. Engage with skeptics, leave the saboteurs to themselves.

Strategy 2: Don’t Engage And Quietly Gain Traction

Have you ever had this happen?: You’re in a meeting where things are moving slowly towards a consensus. Issues are discussed, objections raised and solutions devised. Toward the end of the meeting, just as things are shifting gears to next steps, somebody who had hardly said a word the whole time all of a sudden throws a hissy fit in the middle of the conference room and completely discredits themself.

There’s a reason why this happens. Remember saboteurs are not acting rationally. They have emotional attachments that they often can’t articulate, which is why they rarely give voice to their objections, but rather look for more discreet opportunities to derail the process. When they see things moving forward, they panic.

This doesn’t happen just in conference rooms. Those who are trying to sabotage change prefer to lurk in the background and hope they can quietly derail it. But when they see genuine progress being made, they will likely lash out, overreach and inadvertently further your cause.

This behavior is incredibly consistent. In fact, whenever I’m speaking to a group of transformation and change professionals and I describe this phenomenon to them, I always get people coming up to me afterwards. “I didn’t know that was a normal thing, I thought it was just something crazy that happened in our case!”

It’s important to resist the urge to respond to every attack. You don’t need to waste precious time and energy engaging with those who want to derail your initiative, which is more likely to frustrate and exhaust you than anything else. It’s much better to focus on empowering those who support change. Non-engagement can be a viable way to deal with opposition.

Strategy 3: Design A Dilemma Action

I once had a six-month assignment to restructure the sales and marketing operations of a troubled media company and the Sales Director was a real stumbling block. She never overtly objected, but would rather nod her head and then quietly sabotage progress. For example, she promised to hand over the clients she worked directly with to her staff, but never seemed to get around to it.

It was obvious that she intended to slow-walk everything until the six months were over and then return everything back to the way it was. As a longtime senior employee, she had considerable political capital within the organization and, because she was never directly insubordinate, creating a direct confrontation with her would be risky and unwise.

So rather than create a conflict, I designed a dilemma. I arranged with the CEO of a media buying agency for one of the salespeople to meet with a senior buyer and take over the account. The Sales Director had two choices. She could either let the meeting go ahead and lose her grip on the department or try to derail the meeting. She chose the latter and was fired for cause. Once she was gone, her mismanagement became obvious and sales shot up.

Dilemma actions have been around for at least a century. One early example was Alice Paul’s Silent Sentinels who picketed the Wilson White House with his own quotes in 1917. More recently, the tactic has been the subject of increasing academic interest. What’s becoming clear is that these actions share clear design principles that can be replicated in almost any context.

Key to the success of a dilemma action is that it is seen as a constructive act rooted in a shared value. In the case of the Sales Director, she had agreed to give up her accounts and setting up the meeting was aligned with that agreement. That’s what created the dilemma. She had to choose between violating the shared value or giving up her resistance.

How Change Really Happens

One of the biggest misconceptions about change is that it is an exercise in persuasion. Yet anyone who has ever been married or had kids knows how hard it can be to convince even a single person of something they don’t want to be convinced about. Seeking to persuade hundreds or thousands to change what they think or how they act is a tall order indeed.

The truth is that radical, transformational change is achieved when not when those who oppose it are convinced, but when they discredit themselves. It was the brutality of Bull Connor’s tactics in Birmingham that paved the way for the Civil Rights Act in 1964. It was Russia’s poisoning of Viktor Yushchenko in 2004 that set Ukraine on a different path. The passage of Proposition 8 in California created such controversy that it actually furthered the cause of same-sex marriage.

We find the same dynamic in our work with organizational transformations. Whenever you set out to make a significant impact, there will always be people who will hate the idea and seek to undermine it in ways that are dishonest, underhanded and deceptive. Once you are able to internalize that you are ready to move forward.

Through sound strategies, you can learn to leverage opposition to further your change initiative. You can co-opt those who are rationally skeptical to find flaws in your idea that can be fixed. For those who are adamantly and irrationally opposed to an initiative, there are proven strategies that help lead them to discredit themselves.

The status quo always has inertia on its side and never yields its power gracefully. The difference between successful revolutionaries and mere dreamers is that those who succeed anticipate resistance and build a plan to overcome it.

— Article courtesy of the Digital Tonto blog
— Image credit: Unsplash

Subscribe to Human-Centered Change & Innovation WeeklySign up here to join 17,000+ leaders getting Human-Centered Change & Innovation Weekly delivered to their inbox every week.

Unlearning is More Important Than Learning

Unlearning is More Important Than Learning

GUEST POST from Greg Satell

When I first went overseas to Poland in 1997, I thought I knew how the media business worked. I had some experience selling national radio time in New York and thought I could teach the Poles who, after 50 years of communism, hadn’t had much opportunity to learn how a modern ad market functioned. I was soon disappointed.

Whenever I would explain a simple principle, they would ask me, “why?” I was at a loss for an answer, because these were thought to be so obvious that nobody ever questioned them. When I thought about it though, many of the things I had learned as immutable laws were merely conventions that had built up over time.

As I traveled to more countries I found that even basic market functions, such as TV buying, varied enormously from place to place. I would come to realize that there wasn’t one “right” way to run a business but innumerable ways things could work. It was then that I began to understand the power of unlearning. It is, in fact, a key skill for the next era of innovation.

The One “True” Way To Innovate?

Innovation has become like a religion in business today, with “innovate or die” as its mantra. Much like televangelists preaching the prosperity gospel to gullible rubes, there’s no shortage of innovation gurus that claim to have discovered the secret to breakthrough innovation and are willing to share it with you, for an exorbitant fee, of course.

What I learned researching my book Mapping Innovation, however, is that there is no one “true” path to innovation. In fact, if you look at companies like IBM, Google and Amazon, although they are all world-class innovators, each goes about it very differently. IBM focuses on grand challenges that can take decades to solve, Google integrates a portfolio of innovation strategies and Amazon has embedded a customer obsession deep within its culture and practice.

What I found most interesting was that most people defined innovation in terms of how they’d been successful in the past, or in the case of self-described gurus, what they’d seen and heard to be successful. By pointing to case studies, they could “prove” that their way was indeed the “right” way. In effect, they believed that what they experienced was all there is.

Yet as I’ve explained in Harvard Business Review, innovation is really about finding novel solutions to important problems and there are as many ways to innovate as there are different types of problems to solve. Many organizations expect the next problem they need to solve to be like the last one. Inevitably, they end up spinning their wheels.

The Survival Of The Fittest?

The survival of the fittest is a thoroughly misunderstood concept. Although it arose out of Darwin’s work, it did not originate from him. It was coined by Herbert Spencer to connect Darwin’s work to his own ideas. Darwin’s theory was so novel and powerful at the time, it was difficult to articulate it clearly, and the phrase caught on.

All too often, people assume that Darwin’s theories predicted some sort of teleological end state in which one optimized form will dominate. If that were true, then the optimal strategy for every organism, as well as every business model and every organization, would be to strive to achieve that optimal state and dominate the competition.

Yet that’s not what Darwin meant at all. In fact, his theory rested on three pillars, limited resources, changing environments and super-fecundity, which is the tendency of organisms to produce more offspring that can survive. “Fittest” refers to a temporary state, not a permanent advantage. What is “fit” for one environment may be detrimental in another.

Eastern Europe was, for me, similar to the Galápagos Islands where Darwin first formed his famous theory. Seeing different business environments, in close proximity, give rise to so many different business models opened my eyes to new possibilities. Once I unlearned what I thought I knew, I was able to learn more than I could have imagined.

Turning The Page On Welchism

At the beginning of this century, Fortune magazine proclaimed Jack Welch to be the optimal manager of the last one. American industry had grown sclerotic and bureaucratic. It was in great need of some trimming down and Welch was truly an optimized fit for the environment.

Nicknamed “Neutron Jack” for his penchant of getting rid of all the people and only leaving the buildings standing, he voraciously cut through GE’s red tape. Profits soared, Welch became something of a prophet and “Welchism” a religion. Corporate boards heavily recruited GE executives as CEOs to replicate Welchism at their companies.

Yet as David Gelles explains in his book about Welch’s tenure at GE, The Man Who Broke Capitalism, not all was as it seemed. Yes, Welch made GE more efficient and profitable, but he also increased risk through “financializing” the industrial company, undermined engineering and innovation by moving manufacturing facilities overseas and cooked the books to make profits seem much smoother than they were.

GE would eventually implode, but the damage went much further and deeper than one company. Because Welchism was seen as the “one best way” to run a business, many other firms replicated its methods. The results have been alarming. In fact, a 2020 report by the Federal Reserve found that business dynamism in America has steadily declined since Jack Welch took the helm at GE in 1981.

Clearly we have some unlearning to do.

Moving Boldly Into An Uncertain Future

I’ve thought for some time that the 2020s would look a lot like the 1920s. That was the last time that we had such a convergence of technological, demographic and political forces at one time (and a pandemic as well!). Yet historical analogies can often be misleading. History is long and, if you look enough, you can find an analogy for almost anything.

It is certainly true that history seems to converge and cascade on particular moments and we seem to be at one of these moments now. We will need to unlearn much of what we thought we knew about shareholder value and other things as well. Yet correcting the mistakes of the past is never enough. We need to create anew.

The recently passed CHIPS Act is a good model for how to do this. Much of the $280 billion bill goes to tried and true programs that we under-invested in recent years, such as science programs at the NSF and the DOE as well as programs that support manufacturing and, of course, subsidies to support semiconductors. We know these things work.

Yet other programs are experiments. Some, such as a new Technology Directorate at the NSF are controversial. Others, such as $10 billion that will be spent on regional technology hubs and $1 billion that will go to a RECOMPETE pilot program to empower distressed communities, are new and innovative. We can almost guarantee that there will be hiccups and outright failures along the way.

It is tautologically true that the well-trod path will take us nowhere new. We need to unlearn the past if we are to learn how to build a new future.

— Article courtesy of the Digital Tonto blog
— Image credit: Unsplash

Subscribe to Human-Centered Change & Innovation WeeklySign up here to join 17,000+ leaders getting Human-Centered Change & Innovation Weekly delivered to their inbox every week.

Science Says You Shouldn’t Waste Too Much Time Trying to Convince People

Science Says You Shouldn't Waste Too Much Time Trying to Convince People

GUEST POST from Greg Satell

Experts have a lot of ideas about persuasion. Some suggest leveraging social proof, to show that people have adopted the idea and had a positive experience. Others emphasize the importance of building trust and using emotional rather than analytical arguments. Still others insist on creating a unified value proposition.

These are, for the most part, constructive ideas. Yet they are more a taxonomy than a toolbox. Human nature can be baffling and our behavior is rarely consistent. Sometimes we’ll dig in our heels on a relatively minor point and others we’ll give in on a major issue relatively easily, often without any constable rhyme or reason.

Yet consider this one simple science-based principle that explains a lot: The best indicator of what we think and what we do is what the people around us think and do. Once you internalize that, you can begin to understand a lot of otherwise bizarre behavior and work to spread the ideas you care about. Often it’s not opinions we need to shape, but networks.

Majorities Don’t Just Rule, They Also Influence

Consider a famous set of conformity studies performed by the psychologist Solomon Asch in the 1950s. The design was simple, but ingenuous. He merely showed people pairs of cards, asking them to match the length of a single line on one card with one of three on an adjacent card. The correct answer was meant to be obvious.

However, as the experimenter went around the room, one person after another gave the same wrong answer. When it reached the final person in the group (in truth, the only real subject, the rest were confederates), the vast majority of the time that person conformed to the majority opinion, even if it was obviously wrong!

The idea that people have a tendency toward conformity is nothing new, but that they would give obviously wrong answers to simple and unambiguous questions was indeed shocking. Now think about how hard it is for a more complex idea to take hold across a broad spectrum of people, each with their own biases and opinions.

The truth is that majorities don’t just rule, they also influence, even local majorities. So if you want people to adopt an idea or partake in an action, you need to take into account the communities they are already a part of—at home, at work, in their neighborhood and in other aspects of their social circles. That’s where their greatest influences lie.

The 3 Degrees of Influence Rule

In 1948, Congress authorized funding for the Framingham Heart Study, which would track the lifestyle and health habits, such as diet, exercise, tobacco use and alcohol intake, of 5209 healthy men and women. It was originally intended to last 20 years, but the results were so incredibly useful, it lasted for decades and even included the children of early cohorts.

More than a half century after the study began two researchers, Nicholas Christakis and James Fowler, began to suspect that the Framingham Heart Study could be used for a very different, but important purpose. What they noticed was that the data included not only information about people’s habits, but their social networks as well.

So they set out to see if they could identify causal links between people’s health and their social connections. Using 32 years of data, they were able to establish a strong effect in areas as diverse as happiness, smoking and even obesity. As it turns out, the people around us not only help to shape our opinions, but our health as well.

The really astounding discovery, however, was that the effect extended to three degrees of influence. So not only our friends’ friends, influence us deeply, but their friends too—people that we don’t even know. Wherever we go, we bring that long, complex web of influence with us and we, in turn, help to shape others’ webs of influence too.

So when set out to shape someone else’s opinion, we need to account for social networks. We may, for example, be able to play on a target’s emotions, give them all the facts and evidence and demonstrate strong social proof, but their communities — extending out to three degrees of influence — will always factor in. While we’re working to persuade, those invisible webs of influence may be working against us.

Thanksgiving Dinners And Earnings Guidance

There is no greater American tradition than the crazy uncle at Thanksgiving dinner. No matter what your political persuasion, you are bound to have some relative who holds very different opinions than the rest of the family and who feels no compunction about making clear to everyone at the table exactly where they stand.

As should be clear by now, the reason our crazy uncles are so impervious to persuasion is that we aren’t actually arguing with them at all, but the totality of their social networks. Their friends at work, buddies at the bar, people in their neighborhood and everybody else who they interact with on a regular basis, all get a say at our holiday table.

In much the same way, there isn’t any real reason for CEOs to provide earnings guidance for investors. Steve Jobs refused do it and Apple’s stock during his tenure. Same thing with Unilever under Paul Polman. In 2018, JP Morgan CEO Jamie Dimon and uber-investor Warren Buffett wrote a strong Op-Ed in the Wall Street Journal urging CEOs to end the practice.

During the pandemic many companies stopped giving earnings guidance to investors but, as soon as things began to stabilize, they started up again. It seems incredible, because all of the experts, even McKinsey, have advised against it. Still the vast majority of CEOs are unconvinced, despite all the contrary evidence. Could their networks be playing a role?

Don’t Try To Shape Opinions, Shape Networks

We like to think we can shape the ideas of others. It can sometimes seem like a puzzle. How can we conjure up the right combination of value proposition, analysis, emotive argument and social proof, to persuade our target?. There is, in fact, an enormous communication industry dedicated to exactly that proposition.

Decades of scientific research suggests that it’s not so easy. Our thoughts aren’t just the product of neurons, synapses and neurotransmitters reacting to different stimuli, but also our social networks. The best indicator of what people think and do is what the people around them think and do. While we’re trying to score debate points, those complex webs of influence are pushing back in often subtle, but extremely powerful ways.

We need to be far more humble about our persuasive powers. Anybody who has ever been married or had kids knows how difficult it is to convince even a single person of something. If you expect to shift the opinions of dozens or hundreds—much less thousands or millions—with pure sophistry, you’re bound to be disappointed.

Instead of trying to shape opinions, we’re often better off shaping networks. That’s why we advise our clients pursuing transformational change efforts to start with a majority, even if that majority is only three people in a room of five. You can always expand a majority out, but once you’re in the minority you’re going to get immediate pushback.

Rather than wordsmithing slogans, our time and efforts will be much better spent working to craft cultures, weaving the complex webs of influence that lead to genuinely shared values and shared purpose.

— Article courtesy of the Digital Tonto blog
— Image credit: Pexels

Subscribe to Human-Centered Change & Innovation WeeklySign up here to join 17,000+ leaders getting Human-Centered Change & Innovation Weekly delivered to their inbox every week.