Merging of Humans and Machines

GregC_18Arati Prabhakar, director of the US Defense Advanced Research Projects Agency, recently wrote an article for Wired profiling two cutting-edge DARPA projects, including University of Utah bioengineering associate professor Greg Clark’s continued work on the Utah Slanted Electrode Array. The article is below.

Peter Sorger and Ben Gyori are brainstorming with a computer in a laboratory at Harvard Medical School. Their goal is to figure out why a powerful melanoma drug stops helping patients after a few months. But if their approach to human-computer collaboration is successful, it could generate a new approach to fundamentally understanding complexities that may change not only how cancer patients are treated, but also how innovation and discovery are pursued in countless other domains.

At the heart of their challenge is the crazily complicated hairball of activity going on inside a cancer cell – or in any cell. Untold thousands of interacting biochemical processes, constantly morphing, depending on which genes are most active and what’s going on around them. Sorger and Gyori know from studies of cells taken from treated patients that the melanoma drug’s loss of efficacy over time correlates with increased activity of two genes. But with so many factors directly or indirectly affecting those genes, and only a relatively crude model of those global interactions available, it’s impossible to determine which actors in the cell they might want to target with additional drugs.

That’s where the team’s novel computer system comes in. All Sorger and Gyori have to do is type in a new idea they have about the interactions among three proteins, based on a mix of clinical evidence, their deep scientific expertise, and good old human intuition. The system instantly considers the team’s thinking and generates hundreds of new differential equations, enriching and improving its previous analytical model of the myriad activities inside drug-treated cells. And then it spits out new results.

These don’t predict all the relevant observations from tumour cells, but it gives the researchers another idea involving two more proteins – which they shoot back on their keyboard. The computer churns and responds with a new round of analysis, producing a model that, it turns out, predicts exactly what happens in patients and offers new clues about how to prevent some cases of melanoma recurrence.

In a sense, Sorger and Gyori do what scientists have done for centuries with one another: engage in ideation and a series of what-ifs. But in this case, their intellectual partner is a machine that builds, stores, computes and iterates on all those hundreds of equations and connections.

The combination of insights from the researchers and their computer creates a model that does not simply document correlations – “When you see more of this, you’ll likely see more of that” – but rather starts to unveil the all-important middle steps and linkages of cause and effect, the how and why of molecular interactions, instead of just the what. In doing so, they make a jump from big data to deep understanding.

More than 3,220km away, another kind of human-machine collaboration unfolds at the University of Utah as Greg Clark asks Doug Fleenor to reach out and touch the image of a wooden door on a computer monitor.

Clark knows that Fleenor cannot physically touch this or any other object; Fleenor lost both his hands in a near-fatal electrical accident 25 years ago. But Fleenor’s arm has a chip in it that communicates with the computer, so when he moves his arm the image of a hand on the monitor also moves. He’s done this before – raising his arm, watching the cartoon hand move in sync and seemingly stroke 
the face of the door – but this time it’s different. He lurches back and gasps. “That is so cool!” he blurts.

What’s so cool is that as he guides his virtual hand across that virtual plank, he literally, biologically and neurologically, feels its wooden surface. Thanks to some new software and an array of fine electrical connections between another embedded chip and the nerves running up his arm to his brain, he experiences a synthesised sensation of touch and texture indistinguishable from a tactile event.

Read more in Wired

  Share this story:   Email this to someoneShare on FacebookShare on Google+Share on LinkedInPin on PinterestShare on RedditTweet about this on Twitter