Guiding Principles are the First Necessity – #TEDatIBM


Imagine A Better WorldCan you imagine a world where we could have such great trust in society that you no longer cared about your privacy? While some may already feel that way today, most of us could never imagine such a world, especially given what we have experienced over the past decade. At last year’s TED at IBM event, Marie Wallace addressed the challenge we face today in a brilliant speech with a very practical, evidence based solution in her talk, “Privacy by Design”. Her talk is where my belief that we should abandon hope of any true privacy was replaced by hope that there was a vital, and indeed better way.

In light of this year’s theme for TED at IBM being “Necessity and Invention”, I thought it important to revisit her talk and illuminate this topic from a fresh perspective. When it comes to the matter of privacy, Marie’s talk grounded me in the realization that our guiding principles are the first necessity. The theme, according to the conference web site reflects our common wisdom on the subject:

“Necessity is the mother of invention—or so we have been led to believe. We cannot help but suspect that our needs to create and to shape the world around us run much deeper than simple pragmatism.”

It’s true, it’s not pragmatism that is at the root of our inspiration to invent in my view, it is our ability to be imaginative, our ability to overcome challenges we face, our desire to not only survive, but to thrive. It is something innate in our very being. It is the unique combination of a bias towards action, a sense of greater purpose and a belief in our personal ability against all risks that drives many of us to invent. To create. To struggle against ‘the slings and arrows of outrageous fortune […] and by opposing end them’.

Unfortunately, as we have seen, particularly in the last century, absent a sense of true social responsibility, systems are designed and goods are brought to market out of a necessity that benefits a few at the expense of the many. All too often, it seems that data is being used to manipulate society broadly, and you specifically, instead of empowering us all.

Marie starts her talk with what I believe is one of the most powerful and important concepts of our modern era. It is also apparently from the Pope’s remarks last week during his US tour, a zeitgeist moment where many others are coming to this same realization. The idea that we can create a world by our own design, intentionally, not by inheritance or accident. As with the many lessons on life itself we have learned, we can choose to let it happen as it unfolds or we can direct it. So why not take an active role in shaping the world we’d like to see, and manifest it through thoughtful design. Today more than ever, this power is in our hands, not merely in the hands of a powerful few. But we must exercise it, not abdicate our rights out of a sense of helplessness. It’s a matter of intention and attention – do we want to create systems that are intentionally good for all, or to allow others to create systems that are used to manipulate us. Do we want to give attention and support to organizations and systems that are using their resources to manipulate us? I think not.

It is Marie Wallace’s central premise, that “How we approach privacy, will have the single greatest impact on [our future society].” It not only establishes the expectations of every human relative to their role in the market, but also their role in the work force. Will it continue to be based on suspicion and absent any meaningful degree of trust? Or could it instead be more trustworthy as a result? She and I agree in this case, it could be the latter, but it will take time and greater attention from us all in shaping this better future. As she said in her talk, “The reality is it doesn’t have to be like this. And I don’t think we want to live in a world where it will be like this.”

The alternative, is potentially pretty scary, to even the least educated of us on this subject, and perhaps even to the apathetic if they were to see how their data might be used to manipulate them instead of empower them. One of the examples she mentioned was relative to an organization learning from their data analytics that you had a problem with body image and this was used to sell you diet pills. Countering that approach, what if the system was open and transparent with you, and showed you the insights it generated and gave you options, not just taking advantage of your emotional state. It suggested healthier recipes, encouraged you to take a walk or to uplift your spirit and confidence. These are all possible with data analytics, but only if the controllers of the data and the insights are emboldened with positive intentions for you and society, instead of being motivated to sell you the most expensive diet pills on which they would make a profit.

Marie pointed out that even if you control your data with do not track and other mechanisms, there are still data leaks from one system to another that can provide companies with data you do not want them to have. Recently I was given a demo of a new product that did a deep personality profile on me based on my public social media presence. It developed a profile of my emotional state, and also my psychographic drivers. For many this is the holy grail of marketing, being able to tailor ads based on psychoanalysis to understand what motivates me and hit my ‘hot buttons’.

Indeed, there were a few things it suggested about my personality that were troubling, and many of which were flat our wrong because of how I have been managing my public image on Twitter proactively, and because of some recent #tweetfight I had. This is all data in the public domain, so I don’t necessarily have a problem with it, but I am concerned, as Marie is, about how others with less scrupulous intentions may use it. As a marketer, I am concerned about how a snapshot like the one it produced does not represent the whole of the person, and how even such advanced data analytics can still get it fundamentally wrong by basing such insights on a snapshot instead of the whole me and the deeper insights that would come from having a REAL Relationship with me.

Many will tell you that we need to accept all of this data is already out there and this is simply beyond our control. So the solution is to turn it off, to take ownership of our own data and to block advertising through our devices, as Apple recently enabled with the latest iOS update. But this is not a complete picture of the reality we face. As Marie points out, the current state is really a ‘privacy spaghetti’, or perhaps a ‘spaghetti monster’. We need to go to the root of the challenge and rethink our approach to this important topic from the ground up.

Which is exactly what she did in the IBM case study she shared on how they applied analytics to their vast trove of data generated from one of the oldest Enterprise Social Networks in existence. Instead of thinking of management and the employees as separate interests, she took the perspective of providing maximum value to all participants, not just the management. In taking the Privacy by Design approach, they built the foundation of their data analytics program rooted in privacy as a guiding principle that would rule all decisions and actions that followed, before writing a line of code.

In undertaking the project to better understand their employees, the IBM team embraced three core guiding principles. In my view, the pre-requisite before the invention.

  • A commitment to transparency and openness.
  • Embrace simplicity and ease of use.
  • Focus on personal empowerment.

By making these guidelines simple to understand and visible to all, they gave trust to gain trust. By giving employees actionable insights that would help them improve themselves as a principle benefit, and enabling them to choose whether or not to share those insights with others, it changed the dynamic of the relationship for the better in more ways then one. Still, management was given access to the aggregated analysis to understand the important trends, challenges and opportunities, but did not unnecessarily reveal the private details of a uniquely identifiable individual.

This resulted in a significant upside, the sort of upside that many of us have long been trying to prove to those who would choose to exploit the data instead of protecting and empowering individuals. According to Marie, “Demonstrating openness and transparency builds trust and it allows our users to engage more openly and freely with us and share more data. And more data means more value for them and for us. It’s a virtuous circle.” In short, their approach to Privacy by Design deepened their relationships with their employees. Instead of merely providing the other stakeholders with analytics, these deeper relationships provided increased engagement with interested employees that enabled the accomplishment of even more valuable outcomes.

To my original question – Can you imagine a world where we could have such great trust in society that you no longer cared about your privacy? I’d like to believe we could, by facilitating such transparency that we all knew what was available and where we had control over how it was being used. I’d like to intentionally design such a world, perhaps with you through my new community movement “We Are the Solution“. But in order to abandon my hope or interest in defending privacy, I first need greater confidence that unscrupulous people and companies who value profit above people are not able to use my data, or any data, in a manipulative way.

As you are probably keenly aware, this is not the world in which we live today. But it is a world we could design and build together if we choose to do so. As Scott Swhwaitzberg posits in “Trust me… there’s an app for that” the combination of technology and transparency can make this world a reality. But first, we need to ensure that no corporate desire trumps the guiding principles of our shared values across society. This is why the first necessity is to embrace a common set of guiding principles. This is why we must support organizations who share and operate under such values with our hands, hearts and wallets, and deny such support to those who don’t.

Watch Marie Wallace’s talk, “Privacy by Design” on YouTube, and visit the TED at IBM web site to learn more about the upcoming event.

DISCLOSURE:

Like many of you, I love the inspiration and big thinking that comes out of TED. It’s why I helped to produce BIL back in 2008 and why I spoke at BIL again in 2014. It’s a part of who I am, which is why I attend TEDx whenever I can (and hope to speak at a few next year) and why I was so grateful to be invited to TED at IBM last year and will be attending again on October 15, 2015 as their guest. This post, while not required of me in exchange for my invitation, was written as a part of the IBM New Way to Engage futurist program of which I am a part and is being promoted by IBM through that program. While they are paying to promote it in social media, the words above are completely my own, except as otherwise quoted, and do not reflect the position of IBM.

, , , , ,

  1. #1 by WarrenWhitlock - September 29th, 2015 at 00:15

    Whenever I think about the right to privacy, I like to flip it and consider the opportunity of openness. So much can be settled by holding institutions to this same standard… if they are collecting my data, the have to be open about it (not the data.. just the program). 

    Add in my right to not share and we have a system most of us can live with.

  2. #2 by ChrisHeuer1 - September 29th, 2015 at 15:37

    WarrenWhitlock That is seriously just an awesome perspective Warren. Thank you for sharing.

  3. #3 by marie_wallace - September 29th, 2015 at 15:59

    WarrenWhitlock I had this mad idea at one stage that “privacy” should be viral like the GPL license (for us software folks). So if I choose to be “open” to one application I am forcing it to be open back, and not only that but any other application that uses its data is also forced to be open.

  4. #4 by WarrenWhitlock - September 30th, 2015 at 18:15

    marie_wallace I like where you are coming from on this. Doc Searls explained VRM to me as a fundamental shift in who owns the data. I’m a CRM fan, I think the more we know, the better we can server.. so this took me a while to embrace. 

    With VRM, you own your data and the keys to it. I imagine it allowing a company to have a CRM record with all they can get, but with a privacy lock that prohibits them from doing anything not allowed… sort of how we see pop ups on Twitter/Facebook oAuth use. 

    Easier to explain as a thumbprint or thumbdrive that you hand to a doctor before getting help. All your records are there. You decide to turn on all medical, for an hour or indefinitely. 

    And revoke that access at will.

  1. No trackbacks yet.

Comments are closed.