What does the public
think about AI?

UK, 2024

Authors

Jonathan Dupont

Vinous Ali

Amy Price

Seb Wride

Denise Baron

July 2024

Introduction & Summary

The last 12 months have seen Artificial Intelligence (AI) and the discourse around it continue to rapidly evolve.

Following up on last year’s research, we ran four extensive new nationally representative polls of adults across the US and the UK, asking the public their views on a range of AI issues: their feelings towards it, how they use AI today, how they expected it to evolve, and what they wanted the Government to do in response. This report explores the findings from our UK survey work.

>We asked for their views on everything from AI agents to misinformation, whether an AI could pass the Turing Test, and how important it was for the UK to maintain a technological lead ahead of China.

Here are some of the more interesting things that we found:

Today

  • The most commonly chosen emotion by UK residents for their feelings around AI wascuriosity– but they were also around a fifth more likely to choose negative emotions compared to last year.
  • Only around a third (31%) of UK adults said that theywere confident they could explain how modern AI models worked. In practice, people still seemed to be unsure about the AI models’ relative strengths and weaknesses: thinking that they were best at maths, and weakest at sounding sympathetic.
  • Onlyaround 13% of UK residents are currently using LLM based chatbots regularly. That said, awareness of ChatGPT specifically hasmore than doubled in the last year.

Uses

  • While UK adults thought accelerating science and improving economic productivity could be important benefits of AI, they were much less sure if they would benefit directly – just 5% thought it was likely to increase wages for workers.
  • A majority of UK residents under 35 said they would be interested in trying an AI personal tutor, personal assistant, workout coach or financial advisor. Over a quarter of UK adults under 35 say that they would be interested in an AI girlfriend or boyfriend.
  • Half of UK workers who use AI tools as part of their job worked out how to use those tools themselves. However,42% of all UK workers would be interested in formal training on how best to use AI tools.
  • Giving people choice in how and where AI is used is essential to maintain support.65% of UK adults say they would support AI diagnosis as long as people had a choice, whereas roughly two-thirds were opposed if it was forced on people.

Concerns

  • 66% of UK adults think it is likely that AI will increase unemployment.Those with a Bachelor’s Degree or Master’s were slightly more likely to believe that AI could do their job within twenty years than those with just a secondary school education.
  • That being said, while UK adults were most familiar with the risks of AI-led unemployment, creatingfake images or videos to hurt someone’s reputation is the risk they are currently most  worried about.
  • Over half of UK residents (53%) were not confident that they could detect fake AI generated content on the Internet – with confidence significantly falling the older our respondents got. 72% of Brits are worried that AI generated content could be used to manipulate an election.

Future

  • 47% of UK residents believe that AI could reach a human level of intelligence by the end of the 2030s, compared to 49% last year. By the end of the 2040s, 39% expect it to be 10x more intelligent.
  • Just 21% of UK adults think that trying to create a superintelligence is a good idea, with 40% admitting they were unaware that this was a goal of leading AI labs.

Policy

  • A majority of UK adults expressed support for arange of AI based policies, including requiring AI generated content to be clearly marked, making AI companies legally liable, restricting the export of advanced AI technology, and requiring AI companies to disclose details about their models and their training.
  • The only policy we tested which more people saw as a bad idea than a good idea wasbanning new research into AI.
  • When forced to choose  between the UK prioritising staying at the technological frontier or responsible development, even if this meant other countries like China taking the lead, we saw much less agreement. 39% of UK adults prioritised staying at the frontier, compared to 43% prioritising responsible development– with this disagreement cutting across demographics and political parties.

Foreword

Artificial intelligence (AI) has rapidly evolved from a futuristic concept to an everyday reality for millions of people in the United Kingdom (UK).

There are many new applications that promise to transform how people work, learn, communicate, and navigate the world around them, and even more on the horizon that offer enormous potential.

In the meantime, there are policy questions to consider. The UK set the tone for AI governance when it convened the AI Safety Summit at Bletchley Park last year, bringing together key stakeholders from across the globe to discuss how best to manage risks from recent advances in AI. UK policymakers quickly followed up by creating the first-of-its-kind AI Safety Institute to conduct research on how to test and evaluate advanced AI to ensure its safety. Since then, UK policymakers have been eager to deploy AI to transform public services like the National Health Service, where they hope to capitalise on the benefits AI offers for both productivity and healthcare outcomes.

Crucially, widespread AI adoption will require broad public acceptance of the technology. Technological advancements do not happen in a vacuum, but rather take place within a broader social and political context. The public’s perceptions, concerns, and priorities around AI will be a key driving force in shaping how the UK and other countries develop, deploy, and govern this technology.

This survey provides valuable insights into the current state of public opinion about AI in the UK. It is promising that a majority of adults remain optimistic about the impact of the technology, and it is understandable that the survey reveals a population that is curious and interested in AI, but also concerned about its impact. Some see AI as a force for good that will improve productivity, education, healthcare, and research, while others view it as a threat to jobs, privacy, and even democracy. This duality is reflected in how the UK has approached AI governance, with its focus on balancing concerns with promoting innovation.

Government and industry leaders interested in maintaining the UK’s status as a global leader in AI should be closely attuned to public sentiment about the technology because political support for forward-thinking AI policies will ultimately hinge on public acceptance of the technology. The UK has positioned itself as a proponent of responsible AI innovation, but this survey shows that amongst the UK public there is disagreement on whether to focus on responsible AI development, even if that means letting countries like China take the lead with a less restrained approach, or prioritise staying at the frontier of AI development. The UK will have to decide which path forward it will choose.

What is certain, however, is that more people are choosing to engage with the technology, laying the groundwork for the UK public being ready and willing to benefit from AI. As government and industry work together to address public concerns and ensure that AI development aligns with the values and aspirations of the British people, they will open the doors to widespread AI deployment and the opportunities that will come with it.

Ayesha Bhatti

Policy Analyst, Center for Data Innovation

Today

How do people feel about AI?

In 1950, mathematician and computer scientist Alan Turing proposed the imitation test: a test for intelligent behaviour in a machine whereby a human evaluator has to communicate with an entity in a text chat and decide whether they are talking to another human, or an AI.


We have not reached the point where AIs can reliably pass a Turing test. But we are getting closer: 45% of UK adults told us that they wouldn’t be confident that they could tell whether a chat was with an AI or not in less than a minute. Roughly a third said that they wouldn’t be confident that they could tell in under 10 minutes.


Overall, 54% of UK adults said that AI was developing faster than they expected. That’s up over a third compared to when we asked the same question last year.


When asked about the nearest potential historical comparator to AI as a technology, UK adults point to the computer or the Internet. They don’t see it as transformative as the printing press or electricity – but they also expect it to be significantly more important than, say, social media on its own.

Thinking about the potential importance of AI as a future technology which, if any, of the following past technologies is the closest match in importance?

This rapid development has led to mixed feelings. As with last year, when asked to describe how they feel about AI, the most commonly chosen emotion by UK residents was curiosity – with a mix of positive and negative emotions after that.

What was noticeable was that negative emotions have slightly ticked up compared to last year.

53% of UK adults reported being optimistic about the impact of technology on the economy and society in the future, with only 16% saying they felt pessimistic.


This correlated with their feelings when we asked specifically about artificial intelligence. Brits were moderately more likely to have positive expectations than negative ones, although a significant proportion were unsure or simply felt it would have no effect on them personally.

Do you think AI will make better, make worse, or have no impact for…

Do people understand how AI works?

In some ways, AI is more intuitive than other technologies: often the best way to interact with it is to talk to it how you would another human. In other ways, it is very complex, and even the world’s leading AI experts today do not fully understand how a transformer model works the way it does.

In our polling, onlyaround a third (31%) of UK adults said that they were confident they could explain how modern AI models worked. When we pushed on this further by asking around a range of terms related to AI, we saw even lower levels of awareness.

Which, if any, of the following terms related to AI are you familiar with?

Interestingly, there was very mixed awareness of the relative strengths and weaknesses of today’s models. Many people in our poll seemed to think of AI models as having the traditional strengths and weaknesses of a computer: good at maths and with a perfect memory, but weak at common sense reasoning and sounding empathetic. In practice, this is almost the opposite of the strengths and weaknesses of today’s LLM based models.

As far as you are aware which, if any, of the following abilities would you say that current AI models tend to be strong/weak at?

How are UK Adults using AI today?

Putting aside abstract impressions, how much are UK adults concretely aware and using AI tools that are available today?


In our polling, the highest awareness was for existing AI tools that have been around for a long time: Amazon Alexa, Google Assistant and Apple’s Siri.


That said, ChatGPT was not far behind the big three – and compared to last year’s poll, awareness of OpenAI’s tool had nearly doubled.

Have you personally heard of any of the following?

For ChatGPT, we can also compare usage year on year – with the proportion who say they have used it multiple times increasing from 19% to 43%.

Have you personally used ChatGPT?

Although awareness may be high, this has not yet necessarily turned into regular usage for everyone. In our polling, just 13% of UK adults said they were using one of the LLM based chatbots regularly, with a considerable gradient across both age and gender.

% regularly using ChatGPT/Gemini/Claude/Copilot

In our polling, we saw evidence that usage may continue to grow reasonably fast.On average, over 40% of users of the tools said they had only started using them in the last 3 months.


Those who are using these tools find them overwhelmingly helpful, if not yet essential, to their day-to-day life. Of UK adults using LLM based chatbots:

0 %
say they find them helpful
0 %

say they have become an essential tool they use regularly

0 %

say they use them from time to time, but would not miss them if they didn’t exist

When we asked what use cases people had tried, the most common was to help explain something, with around two thirds of users saying they had done this. After that, around a half of users said that they used them to help brainstorm ideas or write text.

Which, if any, of the following have you used an AI tool to do? Please select all that apply

Opportunities

The Potential Benefits of AI

AI is likely to be one of the most significant economic drivers in the next twenty years. The IMF this year estimated that AI could boost productivity in an advanced economy like the UK by 1.5%,1 similar to predictions last year by Goldman Sachs for the US.2

In our polling, when we asked about the potential benefits from AI we saw an interesting dichotomy: while the most widely recognised benefits were accelerating scientific advancement and increasing productivity across the economy, respondents were much less likely to believe that this would translate into increased wages for workers, with this being the least popular choice.

Which, if any, of the following do you think could be important benefits from AI?

When it came to personal use cases, however, we saw a widespread interest in at least giving AI a try in a variety of roles: from basic research to giving early warning of a new medical condition.

2024 is likely to be a year where there is an increased focus on the creation of agents. Agents are designed not to just to be able to answer questions, but to actually carry out basic tasks for you. Both OpenAI3 and Google4 have been explicit that this is the next leap forward.

In our polling, we asked about a range of potential AI use cases, from acting as a personal assistant to acting as a virtual workout coach. Overall, we saw more caution here than for the more generic AI use cases above, although younger adults were more prepared to give AI agents a go.A majority of UK residents under 35 said they would at least try an AI personal tutor or personal assistant.

Which of the following, if any, would you be interested in using AI agents for in your day-to-day life?
One of the more controversial AI agent use cases has been using them for companionship, as a virtual friend or even boyfriend and girlfriend. In 2023, AI companion app Replika was already reporting over 10 million users worldwide.5 In our polling, only 17% of UK adults  said that they were interested in engaging with AI as a companion. However, there was a big difference across demographics. 29% of women under 35 say that they would be interested in an AI girlfriend or boyfriend.
% interested in trying a virtual boyfriend or girlfriend

At Work

The current wave of AI hype was largely driven by the arrival of ChatGPT – but to what extent are people actually using LLM based chatbots like it at work?

In our poll, just over a third of UK workers told us that they had used a chatbot at work – but over two-thirds of those who had used them said they found them helpful or very helpful.

0 %

of UK workers have used an LLM chatbot tool at work

0 %

of UK workers using LLM based chatbots say that they find them helpful or very helpful

0 %

of UK workers using LLM based chatbots say that they have become an essential tool they use regularly

Overall, only around 13% of workers said they were using these tools regularly. Although the majority of this group are using them at least multiple times a week.

% Regularly using LLM chatbots at work

Those workers who are already using AI tools seem to be classic early adopters: around half of them said they had respectively decided to use these tools on their own, worked out how to use them and say they learn best from exploring and experimenting themselves.

0 %

of UK workers using LLM based chatbots say that they worked out how to use those AI tools themselves

0 %

of UK workers using LLM based chatbots say that they decided to use those AI tools themselves

0 %

of UK workers using LLM based chatbots say that they learn best from exploring and experimenting with AI tools themselves

Which of the following is closer to your view?

Health

Alongside the economy, one of the most significant opportunities from AI is to speed up the diagnosis and treatment of health conditions.

Given the many sensitivities in this space, UK adults are understandably unsure about using AI to diagnose illnesses. When first asked, opinions are fairly evenly split.

0 %

of UK adults support using AI to diagnose patients

0 %

of UK adults say that they oppose using AI to diagnose patients

0 %

of UK adults say they are unsure

Reliability is the most significant concern here, with 77% saying they worried the AI system would give incorrect diagnoses. 59% also worried that an AI would not treat patients in a sympathetic and caring way.

However, with basic protections in place, we saw that it was possible to overcome concerns about AI diagnosis.

For example, 73% of UK residents say they would support AI diagnosis if it was double checked by a human doctor.Giving people the choice whether they used it or not increased support to 65%.

The only situation where we saw strong opposition was in a scenario where patients would be outright forced to use the system, which over two-thirds of UK adults opposed.

Would you support or oppose an AI making diagnoses in the following circumstances?

Concerns

Risks

Ever since science fiction writers first conceived the idea of artificial intelligence, we have been inundated with stories about the many ways they can go wrong. It is therefore perhaps no surprise that we saw a reasonably high level of self-reported familiarity across a range of risks, with the most common being the potential for unemployment.

On a scale from 1-7, how familiar or unfamiliar are you with the following potential risks?

Across the range of harms we presented, from hurting someone’s reputation with embarrassing videos through to human extinction, UK adults do seem to believe that AI represented a significant increase in risk.

For each of the following, in your opinion, how much does AI increase risk compared to existing tools and technologies?

Misinformation and Deepfakes

This survey work was carried out before the Prime Minister called the General Election. Nevertheless, even before the campaign got underway, 53% of UK adults were worried about the potential impact of misinformation on the UK General election.

Perhaps unsurprisingly, voters were more likely to think that the “other” side would benefit most from misinformation: Conservative supporters thought Keir Starmer would be helped most, and Labour supporters thought Rishi Sunak would.

Who do you think misinformation would be most likely to help in the campaign?

While 72% were worried AI generated content would be used to manipulate an election, this was just one of their worries. There were also concerns around the potential for AI to con people out of their money or create sexually explicit deep fakes without consent.

How worried or unworried are you about the following potential risks from AI-generated content?

It was also clear that respondents were more worried about criminals, terrorists, and foreign governments than any domestic political party.

Which, if any, of the following are you most worried about using AI to manipulate people?

When asked whether AI generated content would exacerbate the spread of misinformation, 68% said they thought it would make the problem significantly worse.

This concern is likely, in part, because half of UK adults (53%) were not confident that they could detect fake AI generated content on the Internet – with confidence significantly falling as the age of respondents increased.

Overall there was strong support for better labelling, with 66% of UK adults saying that Governments and companies need to do much more to better label and restrict misleading AI generated content.

How confident, if at all, are you that you can detect fake AI generated content on the Internet?

When we asked indirectly, it was the elderly and children that were seen to be at the highest risk of being misled, with 53% and 45% pointing to each group respectively. Only 1% of respondents claimed not to be worried about anyone at all.

Could AI tools be part of the solution to misinformation, helping to spot and counter it? At present, it seems UK adults need more convincing on this point. Just 35% of UK adults said that they think it is likely that new AI tools could help reduce misinformation.

Automation

New technologies have always changed the structure of the economy – but one of the more unusual things about AI is that there is significantly more uncertainty about who it is likely to affect and how.

We asked people to give a score out of 10 regarding how likely they thought an AI could do their job as well as them in the next 20 years – with an average score of 4.3.

On a scale from 0-10, how likely do you think it is that an AI or robot will be able to do your job as well in the next twenty years?

This score does not vary very much by income level or education – although those with a Bachelor’s Degree or Masters were slightly more likely to believe that AI could do their job than those with just a secondary education.


However, there is a significant difference between younger and older respondents with the average score declining as age increases. Those aged 18-24 have an average score of 5.22, which drops to 2.99 for individuals aged 65 and above.

When we asked our poll respondents to rate what jobs they thought might get automated, computer programming, routine manufacturing jobs and customer services agents were at the top. By contrast, UK adults were less convinced that AIs would be able to take on the roles of scientists, musicians, actors or doctors.

Which of the following jobs, if any, do you think AI or robots would be able to do instead of a human in the next 20 years?

Corresponding with this, UK adults thought that AI was likely to reduce the relative importance of data analysis, coding and graphic design skills – while raising the importance of persuading other humans.

Which, if any, of the following types of general work skills do you think the rise of AI will make it more/less important for humans to be good at?

Overall, over half of those we polled thought both that AI would likely increase unemployment and that Governments should actively seek to counter this.

0 %

of UK adults say that they think it likely AI will increase unemployment

0 %

of UK adults say that governments should try to prevent human jobs from being taken over by AI or robots

0 %

of UK adults say that the Government and companies should offer formal retraining and skills programs to people like me to help them to transition to different careers

This all being said, people remained relatively optimistic about their personal outlook: only around a quarter (24%) expected their job to disappear entirely, while 28% thought they would take on other responsibilities, 29% thought they would oversee the AI, and 29% thought they would work fewer hours.

Future

Artificial General Intelligence

Given the speed of advances in AI, how long is it until AIs reach a capability level equivalent to that of a human?

As with last year’s poll, we asked our respondents by which decade they thought a human level AGI was most likely to be developed. We saw remarkable consistency – 47% believe it would happen by the end of the 2030s, compared to 49% last year. A fifth of the population (20%) thought that this had already happened.

In your opinion, when is an AI - either a computer program or a robot - likely to be first developed that is as smart as a human?

While the 2030s are not very far away, this would suggest that the public are roughly aligned with prediction markets, which also suggest that a date in the 2030s is most likely.

Extending the question this year, we then went on to ask the public how long they thought it would take for an AI to significantly exceed human level intelligence by at least 10x.

On this metric, a significantly smaller proportion of the public thought we had already hit this threshold, while moderately more thought it would never happen. That said, even taking this into account, around half the public thought we would see an AI significantly smarter than a human in the 2040s.

In your opinion, when is an AI - either a computer program or a robot - likely to be first developed that is as smart as a human?

As in last year’s report, we saw that many people did not see intelligence in purely analytical terms, with 50% believing that an AI would have to be capable of feeling emotions to be as smart as a human. This is only a small amount below the level that thought an AI would have to feel emotions to feel conscious.

Which of the following, if any, would an AI have to do to be as smart as a human / conscious in your view?

Superintelligence

If a superintelligent AI was created – an AI significantly more intelligent than any human – what would this mean for the world? Such an AI could develop many new powerful technologies, but could in itself be a significant risk.

In our polling, we saw that UK adults were more wary than welcoming of the idea of a superintelligence:

0 %

of UK adults say that trying to create a superintelligence is a good idea

0 %

of UK adults say that trying to create a superintelligence is a bad idea

0 %

of UK adults say that trying to create a superintelligence is dangerous

70% of UK adults thought superintelligent AI would be used to create new weapons, whereas 38% thought that it would actively seek to destroy human civilisation.By contrast, only around 27% thought it likely to lead to an radical acceleration of economic growth, and just 16% thought it would lead to an end to war.

Suppose we succeeded in developing a highly advanced super-intelligent AI in the next thirty years. How likely or unlikely do you think the following consequences of this would be?

Given both the potential benefits and risks of a superintelligence, only a small minority of UK adults thought we should try to accelerate its development – while more than a third thought respectively that we should stay at today’s pace or actively slow down.

Considering the potential benefits and risks from advanced AI:

0 %

of UK adults say that we should accelerate the development of this technology

0 %

of UK adults say that we should develop it around the same pace as we are now

0 %

of UK adults say that we should look to slow its development

Almost a quarter of UK adults (24%) believe that there is a greater than 10% chance that a superintelligence will cause humans to go extinct in the next 100 years. Compared to other potential existential risks for this same risk threshold, it is seen as ten percentage points more likely than an asteroid, but ten percentage points behind climate change.

Risk of causing humans to go extinct in next 100 years

Policy

How should the Government react?

As part of our poll, we asked our respondents their views on a wide range of policies that other people have suggested: everything from clear labelling to a pause on new research.

In order to get a better view on how urgent a particular issue might be, we allowed them to indicate if they didn’t think it was necessary now,  but were open to it later on.

Across the population we saw a majority of respondents supporting a wide range of policies that they believed should happen now. The top ten most popular actions were as follows.

  1. 76% were in favour of requiring AI generated content to be clearly marked.
  2. 70% were in favour of making AI companies legally liable for any negative outcomes from the use of their technology.
  3. 68% were in favour of requiring AI companies to disclose details about their AI models, such as the dataset used to train their systems and their safety testing.
  4. 68% were in favour of creating an intergovernmental organisation to monitor AI safety and security.
  5. 65% were in favour of banning the export of advanced AI technology to hostile Governments.
  6. 63% were in favour of keeping track of any data centres or advanced microchips that could be used to develop advanced AI.
  7. 61% were in favour of holding regular global conferences with international governments and technologies companies to monitor the progress of the technology.
  8. 57% were in favour of banning the use of AI generated content for political campaigns.
  9. 57% were in favour of providing stronger social safety net policies, such as unemployment insurance, job transition assistance, and job placement services for workers likely to be affected by AI.
  10. 56% were in favour of offering government sponsored retraining to people who are at risk of losing their job to AI.

Despite supporting this range of policies, 56% of UK adults also agreed that we needed tomove cautiously before creating new laws and regulations to avoid creating unintended consequences.

In your opinion, which of the following policies would you be open to?

The only policy which more people saw as a bad idea than thought it should be implemented now was banning new research into AI (40%, compared to 17%). However, around half the population were open to them being necessary at some point down the line.

Net difference between individuals aged 18-25 and those aged 65+ who say each policy should happen now

International Competition

While people instinctively support more regulation overall, we wanted to understand how strong this support is in practice. Most pressingly, do they maintain this view even if it would have a material impact on AI progress in the UK – and threaten other countries taking the technological lead?

When we asked people to make a forced decision between the two, we saw mixed opinions.

0 %

of UK adults said the UK should seek to stay at the technological frontier, developing new AI systems rapidly to ensure it has the world’s most powerful systems

0 %

of UK adults said the UK should develop new AI systems responsibly, even if this means slowing down and letting other countries like China take the lead

0 %

of UK adults said they didn’t know

When we gave people a list of arguments for both sides – prioritising staying at the lead, or responsible development, we saw almost equal agreement across all of them.

How persuasive or unpersuasive do you find the following arguments?

To wrap up the poll, we asked people to explain their views in their own words. In general, we saw that those who believed it was important that the UK remain at the technological frontier had relatively similar views on why it was dangerous to let other countries such as China get ahead – whereas those who prioritised safety had a broader range of reasons why they feared moving too fast with AI.

Those who believe The UK should seek to stay at the technological frontier, developing new AI systems rapidly to ensure it has the world’s most powerful systems
We need to be at the forefront of any technology otherwise we will fall economically behind others countries

Male, 53


We don't want to fall behind the rest of the world.

Female, 32


The United States needs to be the #1 leader in AI to keep the world safe

Male, 45


AI advancement will determine the next super powers.

Male, 46
We need to stay ahead of the game to keep ourselves secure.

Female, 50


The UK should stay at the forefront in technology and build and rely on its own technologies and not those of other countries who may have subversive intentions.

Male, 61


AI is going to be massive in the future and the UK should not be left behind

Male, 34


China is far too powerful as it is!

Female, 76
We need to keep up with advances in tech so we are able to deal with any threats.

Male, 26


We need to rely more on ourselves rather than other countries taking the lead

Female, 36


Because AI is already here and it's going to be part of our lives.

Female, 53
Previous slide
Next slide
Those who believe the UK should develop new AI systems responsibly, even if this means slowing down and letting other countries like China take the lead
I don't think people understand how dangerous AI could be.

Female, 31


I think a considered approach will help regulation and allow a full understanding of risk and how to minimise it.

Female, 29


Developing too fast means more mistakes.

Female, 35


We must think carefully about what we want from AI before we go too far.

Male, 44
I would rather the technology take longer but it be responsible, ethical and safe.

Female, 40


I think when decisions are rushed, it is easy to move in the wrong direction and not consider the wider implications

Female, 69


The potential risks of unchecked and unregulated AI development outweigh the benefits.

Male, 24


I think the country has more important issues at the moment.

Female, 55
There is much to fear and to feel positive about with AI. I think a considered but slower approach will help regulation and allow a full understanding of risk and how to minimise it.

Female, 29


I believe such game changing technology should only be introduced with a responsible approach, because it is much harder to do this retrospectively. Doing something properly is more important than doing it first

Male, 61
Previous slide
Next slide
  1. https://www.imf.org/en/Publications/WEO/Issues/2024/04/16/world-economic-outlook-april-2024
  2. https://www.goldmansachs.com/intelligence/pages/ai-may-start-to-boost-us-gdp-in-2027.html
  3. https://www.technologyreview.com/2024/05/01/1091979/sam-altman-says-helpful-agents-are-poised-to-become-ais-killer-function/
  4. https://blog.google/intl/en-africa/products/explore-get-answers/google-io-2024-an-io-for-a-new-generation/
  5. https://www.hbs.edu/faculty/Pages/item.aspx?num=63508