Matthew Griffin, described as “The Adviser behind the Advisers” and a “Young Kurzweil,” is the founder and CEO of the World Futures Forum and the 311 Institute, a global Futures and Deep Futures consultancy working between the dates of 2020 to 2070, and is an award winning futurist, and author of “Codex of the Future” series. Regularly featured in the global media, including AP, BBC, CNBC, Discovery, RT, and Viacom, Matthew’s ability to identify, track, and explain the impacts of hundreds of revolutionary emerging technologies on global culture, industry and society, is unparalleled. Recognised for the past six years as one of the world’s foremost futurists, innovation and strategy experts Matthew is an international speaker who helps governments, investors, multi-nationals and regulators around the world envision, build and lead an inclusive, sustainable future. A rare talent Matthew’s recent work includes mentoring Lunar XPrize teams, re-envisioning global education and training with the G20, and helping the world’s largest organisations envision and ideate the future of their products and services, industries, and countries. Matthew's clients include three Prime Ministers and several governments, including the G7, Accenture, Bain & Co, BCG, Credit Suisse, Dell EMC, Dentons, Deloitte, E&Y, GEMS, Huawei, JPMorgan Chase, KPMG, Lego, McKinsey, PWC, Qualcomm, SAP, Samsung, Sopra Steria, T-Mobile, and many more.
WHY THIS MATTERS IN BRIEF
- Plagued by critics who accuse Google DeepMind of getting too much “free” access to NHS patient data, DeepMind’s new “bitcoin like” audit platform will let people know exactly how data is being accessed and used
Google’s AI powered health tech subsidiary, DeepMind Health, is planning to use a new technology loosely based on bitcoin to let hospitals, the NHS and eventually even patients track what happens to personal data in real-time.
Dubbed “Verifiable Data Audit”, the plan is to create a special digital ledger that automatically records every interaction with patient data in a cryptographically verifiable manner. This means any changes to, or access of, the data would be visible.
DeepMind has been working in partnership with London’s Royal Free Hospital to develop kidney monitoring software called Streams and has faced criticism from patient groups for what they claim are overly broad data sharing agreements. Critics fear that the data sharing has the potential to give DeepMind, and thus Google, too much power over the NHS.
In a blogpost, DeepMind co-founder, Mustafa Suleyman, and head of security and transparency, Ben Laurie, use an example relating to the Royal Free Hospital partnership to explain how the system will work.
“An entry will record the fact that a particular piece of data has been used, and also the reason why, for example, that blood test data was checked against the NHS national algorithm to detect possible acute kidney injury,” they write.
Suleyman says that development on the data audit proposal began long before the launch of Streams, when Laurie, the co-creator of the widely-used Apache server software, was hired by DeepMind.
“This project has been brewing since before we started DeepMind Health,” he said, “and it adds another layer of transparency. Our mission is absolutely central, and a core part of that is figuring out how we can do a better job of building trust. Transparency and better control of data is what will build trust in the long term.”
Suleyman pointed to a number of efforts DeepMind has already undertaken in an attempt to build that trust, from its founding membership of the industry group Partnership on AI to its creation of a board of independent reviewers for DeepMind Health, but argued the technical methods being proposed by the firm provide the “other half” of the equation.
Nicola Perrin, the head of the Wellcome Trust’s Understanding Patient Data taskforce, welcomed the verifiable data audit concept.
“There are a lot of calls for a robust audit trail to be able to track exactly what happens to personal data, and particularly to be able to check how data is used once it leaves a hospital or NHSDigital. DeepMind are suggesting using technology to help deliver that audit trail, in a way that should be much more secure than anything we have seen before.”
Perrin said the approach could help address DeepMind’s challenge of winning over the public.
“One of the main criticisms about DeepMind’s collaboration with the Royal Free was the difficulty of distinguishing between uses of data for care and for research. This type of approach could help address that challenge, and suggests they are trying to respond to the concerns.
“Technological solutions won’t be the only answer, but I think will form an important part of developing trustworthy systems that give people more confidence about how data is used.”
The systems at work are loosely related to the cryptocurrency bitcoin, and the blockchain technology that underpins it.
“Like blockchain, the ledger will be append-only, so once a record of data use is added, it can’t later be erased. And like blockchain, the ledger will make it possible for third parties to verify that nobody has tampered with any of the entries,” said Suleyman.
Laurie downplays the similarities.
“I can’t stop people from calling it blockchain related,” he said, but he described blockchains in general as “incredibly wasteful” in the way they go about ensuring data integrity: the technology involves blockchain participants burning astronomical amounts of energy – by some estimates as much as the nation of Cyprus – in an effort to ensure that a decentralised ledger can’t be monopolised by any one group.
DeepMind argues that health data, unlike a cryptocurrency, doesn’t need to be decentralised – Laurie says at most it needs to be “federated” between a small group of healthcare providers and data processors – so the wasteful elements of blockchain technology need not be imported over. Instead, the data audit system uses a mathematical function called a Merkle tree, which allows the entire history of the data to be represented by a relatively small record, yet one which instantly shows any attempt to rewrite history.
Although not technologically complete yet, DeepMind already has high hopes for the proposal, which it would like to see form the basis of a new model for data storage and logging in the NHS overall, and potentially even outside healthcare altogether.
“Right now,” says Suleyman, “it’s really difficult for people to know where data has moved, when, and under which authorised policy. Introducing a light of transparency under this process I think will be very useful to data controllers, so they can verify where their processes have used or moved or accessed data. That’s going to add technical proof to the governance transparency that’s already in place. The point is to turn that regulation into a technical proof.”
In the long run, Suleyman says, the audit system could be expanded so that patients can have direct oversight over how and where their data has been used, and it could even be expanded to other hospitals, but such a system would come a long time in the future, once concerns over how to secure access have been solved.