Christopher Bishop has been on the forefront of Microsoft’s artificial intelligence applications for a while since he runs the corporate’s AI for Science research unit, which applies the powerful technology to the natural sciences.
Bishop sees the mission of the lab, which was founded in 2022, as accelerating scientific discovery using the technology. His team studies every little thing from how AI models may also help discover latest materials to how they will assist weather forecasting by predicting changes within the atmosphere.
In this conversation with the Financial Times’ AI editor Madhumita Murgia, he explains why he believes scientific discovery will prove to be the one most significant application of the technology.
Madhumita Murgia: Why did Microsoft found the AI for Science lab in 2022, with you heading it up in Cambridge? What’s its goal?
Christopher Bishop: The mission of the lab is to speed up scientific discovery with AI by science inside the natural sciences. Think chemistry, physics, biology and related areas like astronomy. The field is just not latest. My profession began off in physics after which plasma physics for fusion, and I moved into the sphere of neural (networks) 35 years ago.
What became clear to me is that the deep-learning revolution has increased the potential of machine learning and, subsequently, increased the potential for it to affect scientific discovery. So, within the 12 months or so leading as much as the formation of the team, we had quite a few projects scattered around Microsoft Research in relevant areas. It was clear we desired to speed up this.
I suggested pulling together a team, combining a mixture of existing projects, bringing those together under one umbrella after which growing with some latest hires, to make this a focus.
My view is that scientific discovery will prove to be the one most significant application of artificial intelligence.
AI Exchange
This spin-off from our popular Tech Exchange series of dialogues will examine the advantages, risks and ethics of using artificial intelligence, by talking to those on the centre of its development
Scientific discovery is so fundamental to human progress. It’s about gaining a greater understanding of the world, so that we are able to improve the human condition, whether it’s agriculture, industry, drug discovery, tackling healthcare, latest types of energy, sustainable types of energy or tackling climate change.
MM: You talked about moving from physics to AI. AI scientist Geoffrey Hinton won the (2024) Nobel Prize for physics, which some people found an unexpected classification because the two fields are quite separate. Hinton wasn’t a physicist to start with. How did you progress from physics to AI?
CB: As a young person, I used to be fascinated by the thought of artificial intelligence. I used to be fascinated by the brain. The brain continues to be the most important unsolved mystery within the universe. We understand the universe, yet there’s this kilo and a half of jelly that you could hold in your hand. It’s largely an entire mystery (but) is able to incredible feats of data processing and creativity.
Understanding that and the way we are able to recreate something along those lines in a machine, how we are able to amplify the capabilities of the human brain using artificial types of intelligence, that is intellectually very exciting. But on the time, I discovered the sphere of artificial intelligence uninteresting since it was about find out how to construct rules that you could program right into a computer which make the pc (seem) intelligent.
That, for me, was deeply unsatisfactory. It never appealed to me. And along got here Geoff and others, they usually were mavericks in a way. They were proposing this field — it was called connectionism on the time, or neural (networks) now — that was, not less than loosely, modelled on the thought of the brain.
I discovered that intellectually fascinating. That felt like a path to intelligence. You could never write a algorithm that may make a machine intelligent, but here was a path to artificial intelligence. I discovered that inspiring. When I look back, I will need to have been very courageous, because I had a decent, and successful, profession as a theoretical physicist.
I walked away into what, on the time, was seen as a fairly flaky field. It wasn’t proper, mainstream computer science, it wasn’t mainstream physics, but it surely was very inspiring; 35 years later, that appears like quite a very good decision.

MM: You’ve been on this field for 35 years. It’s evolved so much in that point. What have been the important thing moments of inflection — changing points which have stood out to you as having transformed the sphere? What has that journey looked like from the perspective of those inflections?
CB: If you take a look at it from 50,000ft, there have been three phases. The first phase was on a much smaller scale than today, but a number of excitement around neural nets (was) coming out of parents like Geoff Hinton and others in the sphere. The thing I believe I delivered to the sphere was recognising that, although these are inspired by neurobiology, what we were doing was statistics, albeit complex, non-linear statistics.
Then we found that these networks could solve interesting problems, but they were limited. They didn’t really have the performance, the accuracy, for real-world applications. They could do fun things within the laboratory, and it was impressive, but they sort of ran out of steam.
So, within the second phase, the sphere of neural nets went into the background. Quite a lot of people got inquisitive about other approaches.
The big breakthrough got here around in 2012 — and again Geoff was instrumental in that: the event of deep learning. It was a breakthrough in our ability to coach networks with many layers of processing. That was transformational, and in order that’s really the start of the trendy era. Problems that had eluded us for a decade or more suddenly became relatively much easier to resolve.
Not only that — the identical technology that led to a breakthrough in computer vision has led to a breakthrough in speech recognition. Then we began to apply it to other fields.
The rest is history. That’s the curve we’ve been on, driven by that fundamental ability to coach models which are very deep, because when you’ve got a model that has many layers of processing, it’s extremely general. Now you’ll be able to apply it to a complete host of various areas.
MM: When did you begin to grow to be impressed by what language models could do and to consider that they may be a next stage within the evolution of those systems?
CB: I used to be privileged, because I used to be certainly one of the relatively small number of individuals in Microsoft who was given early access to (Open AI’s model) GPT-4 while it was confidential. It was a rare moment to play with GPT-4. Back then (in 2023 when it was released), when no person — or only a few people — had ever seen technology like this, to grasp that it was a serious step forward in the power to generate languages.
It’s extremely good at generating human language, surprisingly good. But the second thing . . . the shocking thing, is that we had a system here which, for the primary time, can actually reason. It didn’t just generate relevant text. It understood what was occurring and will reason.
Now, its understanding, after all, is a shallower level than human understanding, but I liken it to (the primary time the Wright brothers flew a powered aeroplane in) 1903, standing on Kill Devil Hills at Kitty Hawk, and watching a few bicycle mechanics struggle into the air on this contraption.

You could have checked out it and said: ‘I’m not very impressed with that’. It only flew 120ft. Or you possibly can say: ‘Wow, that is the start of a brand new era’. It was that feeling — it was just like the hairs on the back of my neck standing up and (me) considering, for the primary time in my life, I’m interacting with a machine that’s showing — it’s sometimes called — the sparks of artificial intelligence. A protracted technique to go for human-level intelligence and beyond, but . . . like a primary encounter, in a way, at a private level. That was a remarkable moment.
For me, (using GPT-4) was more visceral. It was less about, we’ve run this benchmark and, look, it’s so a lot better than last 12 months’s benchmark. It was something quite different: reasonably qualitative and just realising that you possibly can have had a conversation with GPT-3 before, and you’d have had nice paragraphs and a pleasant conversation. But here . . . just the primary time, you realised you were coping with something that’s qualitatively different.
MM: Which points of science do you’re thinking that have been most modified by AI and where are you seeing practical progress?
CB: The thing that actually excites me is the differences between a big language model (LLM) used to assist other forms of information work and the character of science.
When you consider scientific discovery, let’s imagine you’re a pharmaceuticals company and also you’re attempting to develop a drug to tackle a selected disease. You’ve got a protein that you simply’re attempting to goal, and . . . the space of organic molecules that it’s essential to explore is about 10 to the ability 60.
It’s a gargantuan space, and it’s essential to explore that space to search out one or two molecules that may bind with the goal; that might be absorbed into the body; metabolise accurately; they’re not toxic; they might be synthesised, and all the remaining of it. So, you’re trying to search out that needle in a haystack. That’s not done by one person in a day. That’s a team of dozens or a whole lot of individuals working for a few years.
As a scientist, in a really perfect world, you’d have read every paper that’s ever been written, and absorbed and internalised it. That’s unimaginable for a human being, but that’s something a big language model can do.
But it’s rather more than that. An necessary aspect of science is that it involves experimentation. Fundamentally, it’s about evidence and about conducting experiments. In something like molecular research, you do plenty of experiments. You’re getting the outcomes of the experiments and also you’re refining those hypotheses, and also you’re going around that loop, often persistently.
You go around that iterative process, however the steps are being accelerated by AI — very dramatically. In a way, that’s the most important news of what’s real today, versus what may or may not occur in the long run.
MM: Why is Microsoft inquisitive about investing in science? What’s exciting, when it comes to progress, that may help corporations like Microsoft?
CB: Microsoft has a leadership position in artificial intelligence. We have tremendous infrastructure and (are) constructing that out straight away. Then the query is, where can that bring profit? My view — but I believe it’s one which’s shared broadly inside the company — is that scientific discovery is an area that may see significant acceleration and disruption through AI.
Secondly, it’s of tremendous value to society that it underpins human development — it’s that fundamental. Then, what’s Microsoft’s role as an organization? It’s to speed up the work and empower the work of others.
So, we take into consideration drug discovery. We take into consideration materials design, attempting to develop batteries, photoelectric cells, methods for capturing CO₂, and so forth. There are many organisations all over the world doing this. We consider that AI will likely be a giant accelerant. We consider we’ve got world-leading AI technology. To have the opportunity to bring that to customers and partners aligns with Microsoft’s mission of empowering others. Something we’d like to do in AI for Science is, through the research advances, to have the opportunity to create tools that might be utilized by scientists, possibly broad ranges of applications.
MM: Do you see the present paradigm, where AI models, particularly LLMs, need more (computing power), data and greater models to bring subsequent breakthroughs as a giant gamble? Or are there other ways by which we’re going to get to more powerful AI systems which are going to bring us to all these other applications?

CB: Machine learning is a wealthy field. We’ve seen a selected framework that’s been very successful — the LLM and the actual scaling, architecture approach and so forth. Because it’s been very successful, there’s been a number of attention interested in that. First of all, there are numerous different variants that might be explored for this. It’s not a hard and fast architecture.
There are so many various approaches and, after all, the sphere feels so much greater now. Rather a lot more persons are working on this area — more attention, more interest in the sphere. Just as in the times when neural nets ran out of steam and we saw that Cambrian explosion in that second phase of neural computing, I believe there’s an incredible amount of creativity that might be unlocked.
And so we begin to see the necessity for complementary approaches or variations of current techniques. There’s an incredible space to explore. The concept of a neural net is a general one, and we’ve seen certain architectures that work extremely well. We use LLM-like technology. For example, we’ve got generative models for drugs. We’ve recently published some work called the TamGen molecular generator, which has produced a brand new potential drug molecule for tackling tuberculosis that’s 100 times more practical at binding to the goal protein than the previous molecule.
It’s trained on the language of molecules. It can generate three-dimensional small organic molecules which are potential drug candidates. But we also use other architectures where appropriate. So it’s about the precise machine-learning tool for the job.
Building those laws of physics into the architecture leads you to something that’s reasonably different in structure from, say, GPT-4. It stays to be seen how far this scaling law takes us. But even when, in some unspecified time in the future, we discover that that’s not the optimal path to pursue, there are numerous alternatives to explore as well.
MM: What will we see within the domain of AI in science over the subsequent two to 5 years?
CB: The one thing that’s now clear is that this ability to take things that we knew find out how to do through lots and plenty of (computing power), and we are able to now do them much faster, corresponding to weather forecasting. This idea of an emulator, this thing we call the fifth paradigm, because it were, seems to be very robust. We’ve seen it in many various scenarios. It’s a general purpose template, if you happen to like, that we are able to apply in plenty of different places.
In the subsequent couple of years, we’ll see significant advances on this, probably in quite a few different domains. Very likely we’ll see those landing in practical ways in which scientists can use.