logo
What International AI Safety report says on jobs, climate, cyberwar and more

What International AI Safety report says on jobs, climate, cyberwar and more

The Guardian29-01-2025

The International AI Safety report is a wide-ranging document that acknowledges an array of challenges posed by a technology that is advancing at dizzying speed.
The document, commissioned after the 2023 global AI safety summit, covers numerous threats from deepfakes to aiding cyberattacks and the use of biological weapons, as well as the impact on jobs and the environment.
Here are some of the key points from the report chaired by Yoshua Bengio, a world-leading computer scientist.
In a section on 'labour market risks', the report warns that the impact on jobs will 'likely be profound', particularly if AI agents – tools that can carry out tasks without human intervention – become highly capable.
'General-purpose AI, especially if it continues to advance rapidly, has the potential to automate a very wide range of tasks, which could have a significant effect on the labour market. This means that many people could lose their current jobs,' said the report.
The report added that many economists believe job losses could be offset by the creation of new jobs or demand from sectors not touched by automation.
According to the International Monetary Fund, about 60% of jobs in advanced economies such as the US and UK are exposed to AI and half of these jobs may be negatively affected. The Tony Blair Institute has said AI could displace up to 3m private-sector jobs in the UK, though the ultimate rise in unemployment will be in the low hundreds of thousands because growth in the technology will create new roles in an AI-transformed economy.
'These disruptions could be particularly severe if autonomous AI agents become capable of completing longer sequences of tasks without human supervision,' the report says.
It adds that some experts have pointed to scenarios where work is 'largely' eliminated. In 2o23 Elon Musk, the world's richest person, told the former UK prime minister, Rishi Sunak, that AI could ultimately replace all human jobs. However, the report said such views were controversial and there was 'considerable uncertainty' over how AI might impact labour markets.
The report described AI's impact on the environment as a 'moderate but rapidly growing contributor' as datacentres – the central nervous systems of AI models – consume electricity to train and operate the technology.
Datacentres and data transmission account for about 1% of energy-related greenhouse gas emissions, said the report, with AI constituting up to 28% of datacentre energy consumption.
It added that models are using more energy as they become more advanced and warned that a 'significant portion' of global model training relies on high-carbon energy sources such as coal or natural gas. Use of renewable energy by AI firms and improvements in efficient have not kept pace with rising demand for energy, said the report, which also points to tech firms admitting that AI development is harming their ability to meet environmental targets.
The report also warns that water consumption by AI, used for cooling equipment in datacentres, could pose a 'substantial threat to the environment and the human right to water'. However, the report adds that there is a shortage of data about the environmental impact of AI.
An all-powerful AI system evading human control is the central concern of experts who fear the technology could extinguish humanity. The report acknowledges those fears but says opinion varies 'greatly'.
'Some consider it implausible, some consider it likely to occur, and some see it as a modest-likelihood risk that warrants attention due to its high severity,' it says.
Bengio told the Guardian that AI agents, which carry out tasks autonomously, are still being developed and so far are unable to carry out the long-term planning necessary for those systems to eradicate jobs wholesale – or evade safety guidelines. 'If an AI cannot plan over a long horizon, it's hardly going to be able to escape our control,' he said
The report states that new models can create step-by-step guides to creating pathogens and toxins that surpass PhD-level expertise. However, it cautions that there is uncertainty over whether they can be used by novices.
There is evidence of advancement since an interim safety report last year, the experts say, with OpenAI producing a model that could 'meaningfully assist experts in the operational planning of reproducing known biological threats'.
A fast-growing threat from AI in terms of cyber-espionage is autonomous bots being able to find vulnerabilities in open-source software, the term for code that is free to download and adapt. However, relative shortcomings in AI agents mean the technology is not able to plan and carry out attacks autonomously.
The report lists an array of known examples of AI deepfakes being used maliciously, including tricking companies into handing over money and creating pornographic images of people. However, the report says there is not enough data to fully measure the amount of deepfake incidents.
'Reluctance to report may be contributing to these challenges in understanding the full impact of AI-generated content intended to harm individuals,' the report says. For example, institutions often hesitate to disclose their struggles with AI-powered fraud. Similarly, individuals attacked with AI-generated compromising material about themselves may stay silent out of embarrassment and to avoid further harm.'
The report also warns that there are 'fundamental challenges' to tackling deepfake content, such as the ability to remove digital watermarks that flag AI-generated content.

Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Drivers floored to discover meaning behind name of Elon Musk's Tesla Motors
Drivers floored to discover meaning behind name of Elon Musk's Tesla Motors

Daily Mirror

time37 minutes ago

  • Daily Mirror

Drivers floored to discover meaning behind name of Elon Musk's Tesla Motors

Tesla is one of the most famous car companies in the world - but the name has a very special meaning few people know about thanks to Elon Musk. Here's what you need to know Tesla, the American trailblazer for electric vehicles and clean energy, has been talked about a lot this year. Established in 2003 as Tesla Motors by Martin Eberhard and Marc Tarpenning, it wasn't long before the charismatic Elon Musk powered through the initial investment phase in February 2004, taking charge as chairman and eventually climbing to the role of CEO in 2008. While Tesla's innovative streak is well known, there's still a veil of intrigue surrounding the origins of its name. It was recently brought up on social media when someone posed the question: "What is the meaning of the word 'Tesla' in a business context? How did the company get its name?" ‌ The query sparked a flurry of speculation but, unknown to many, the company pays homage to Nikola Tesla, an electrical engineer and illustrious inventor whose accomplishments were overshadowed during his time. Musk's decision to honour him with the company name keeps Tesla's spirit alive, contributing to his belated recognition. ‌ An article by Seattle Times highlighted: "In the age of Edison, Westinghouse, Marconi and J.P. Morgan, Tesla was a giant of innovation because of his contributions in the fields of electricity, radio and robotics." Many are now recognising that Tesla is finally receiving the recognition he deserves, largely attributed to Elon Musk. Marc Seifer, author of Wizard: The Life and Times of Nikola Tesla, stated: "It's a sociological fact that Elon Musk took the Tesla name and launched Nikola Tesla into the stratosphere. Tesla's risen to the surface again, and now he's getting his due." Despite gracing the cover of Time magazine in 1931, the inventor died impoverished in 1943 after years spent on underfunded projects. Today, some of Tesla's most significant inventions are gaining renewed attention. In 1884, Tesla arrived in New York with hopes that Thomas Edison would back his alternating-current (AC) motor and electrical system. The Seattle Times remarked: "But Edison was investing in highly inefficient direct-current (DC) systems, and he had Tesla re-engineer a DC power plant on Pearl Street in Lower Manhattan. The men soon parted company over a financial dispute. ‌ "George Westinghouse provided funding for Tesla's AC induction motors and devices, which soon came to dominate manufacturing and urban life. Unlike the DC motors of the time, Tesla's motors didn't create sparks or require expensive permanent magnets to operate. "Instead, they used a rotating magnetic field that used power more efficiently in a basic design that is still the core of most electric motors." In 1896, Tesla's creation of the power-generating system at Niagara Falls marked a massive leap for his AC system. Soon after cities started to adopt AC power following Westinghouse's victory over Edison, the staunch DC advocate, in a rivalry depicted in The Current War with Benedict Cumberbatch portraying Edison. Besides, Tesla was pivotal in evolving radio technology, experimenting from 1892 to 1894 and famously describing the radio as an "oscillator" that transformed electricity into high-frequency radio waves for transmitting energy, sound and further communications across vast distances. Envisioning a global transmission network, Tesla conducted promising tests in Colorado in 1899 and then attempted to build a so-called "World System" in Shoreham, Long Island, aiming for the wireless powering of vehicles, ships and planes. Just like smartphones fetch data, sound and visuals from thin air today, Tesla imagined universal accessibility to electricity, yet financial constraints and the cessation of support from financier J.P. Morgan Jr. thwarted his revolutionary aspirations. The Seattle Times remarked on the bittersweet legacy of a brilliant mind, stating: "His achievements remained incomplete. He created the AC energy system and the basics of radio communication and robotics but wasn't able to bring them all to fruition. His life shows that even for a brilliant inventor, innovation doesn't happen in a vacuum. It requires a broad spectrum of talents and skills. And lots of capital."

Learn How to Evaluate Large Language Models for Performance
Learn How to Evaluate Large Language Models for Performance

Geeky Gadgets

time43 minutes ago

  • Geeky Gadgets

Learn How to Evaluate Large Language Models for Performance

What if you could transform the way you evaluate large language models (LLMs) in just a few streamlined steps? Whether you're building a customer service chatbot or fine-tuning an AI assistant, the process of assessing your model's performance often feels like navigating a maze of technical jargon and scattered tools. But here's the truth: without proper evaluations, even the most advanced AI can fail to deliver accurate, reliable, and meaningful results. In this quick-start guide, Matthew Berman demystifies the art of LLM evaluations, showing you how to set up a robust process that ensures your AI solutions are not just functional but exceptional. With a focus on Retrieval-Augmented Generation (RAG) evaluations and Amazon Bedrock, this guide promises to make a once-daunting task surprisingly accessible. By the end of this tutorial, Matthew Berman explains how to configure a secure AWS environment, build a knowledge base, and implement structured evaluation metrics—all while using Amazon Bedrock's powerful tools like prompt management and safety guardrails. Along the way, you'll learn how to compare models, pinpoint weaknesses, and refine your AI for optimal performance. Whether you're a seasoned developer or just starting out, this guide offers actionable insights to help you evaluate LLMs with confidence and clarity. Ready to discover how a well-designed evaluation process can elevate your AI projects from good to new? Let's explore the possibilities together. LLM Evaluation with Amazon Bedrock The Importance of Model Evaluations Model evaluations are the cornerstone of building dependable AI systems. They ensure your AI delivers accurate, coherent, and contextually relevant results. For instance, if you're deploying a chatbot to answer questions about a 26-page hotel policy document, evaluations are essential to verify that the responses are both correct and meaningful. Evaluations also serve several key purposes: Benchmarking: Track your model's performance over time to monitor improvements or regressions. Track your model's performance over time to monitor improvements or regressions. Identifying weaknesses: Pinpoint areas where the model requires refinement. Pinpoint areas where the model requires refinement. Model comparison: Evaluate multiple models to determine the best fit for your specific use case. Without thorough evaluations, it becomes challenging to measure the effectiveness of your AI or ensure it meets user expectations. Understanding Amazon Bedrock Amazon Bedrock is a fully managed service designed to simplify working with LLMs. It provides access to a variety of AI models from providers such as Amazon, Meta, and Anthropic, along with tools to assist evaluation and deployment. Key features of Amazon Bedrock include: Agents: Automate workflows and repetitive tasks efficiently. Automate workflows and repetitive tasks efficiently. Safety guardrails: Ensure ethical and secure AI usage by preventing harmful or biased outputs. Ensure ethical and secure AI usage by preventing harmful or biased outputs. Prompt routing: Optimize query handling to improve response accuracy. Optimize query handling to improve response accuracy. Knowledge base integration: Seamlessly connect external data sources for enhanced contextual understanding. Seamlessly connect external data sources for enhanced contextual understanding. Prompt management: Organize, test, and refine prompts to improve model performance. These features make Amazon Bedrock an ideal platform for evaluating and optimizing LLMs, particularly in scenarios requiring external data integration and robust evaluation metrics. Setup LLM Evaluations Easily in 2025 Watch this video on YouTube. Check out more relevant guides from our extensive collection on Large Language Models (LLMs) that you might find useful. Practical Use Case: Chatbot for a Hotel Policy Document Imagine you are tasked with creating a chatbot capable of answering questions about a detailed hotel policy document. This scenario underscores the importance of integrating external knowledge bases and conducting thorough evaluations. By following the steps outlined below, you can set up and assess the chatbot's effectiveness, making sure it provides accurate and helpful responses to users. Step 1: Configure Your AWS Account Begin by setting up your AWS account. Create IAM users with the necessary permissions to access Amazon Bedrock, S3 buckets, and other AWS services. Ensure that permissions are configured securely to prevent unauthorized access. If required, adjust Cross-Origin Resource Sharing (CORS) settings to enable resource access from different origins. Proper configuration at this stage lays the foundation for a secure and efficient evaluation process. Step 2: Set Up S3 Buckets Amazon S3 buckets serve as the storage backbone for your evaluation process. Create and configure buckets to store essential resources, including: Knowledge base: The hotel policy document or other reference materials. The hotel policy document or other reference materials. Test prompts: A set of queries designed to evaluate the chatbot's responses. A set of queries designed to evaluate the chatbot's responses. Evaluation results: Data generated during the evaluation process for analysis. Implement proper access controls to secure sensitive data and ensure compliance with privacy standards. Step 3: Build the Knowledge Base Upload the hotel policy document to an S3 bucket and convert it into a vector store. A vector store transforms the document into a searchable format, allowing efficient querying by the LLM. Once the knowledge base is prepared, sync it with Amazon Bedrock to allow the model to access it during evaluations. This step ensures the chatbot can retrieve relevant information to answer user queries accurately. Step 4: Set Up RAG Evaluation Retrieval-Augmented Generation (RAG) evaluation combines the generative capabilities of LLMs with an external knowledge base to produce accurate and contextually relevant responses. In Amazon Bedrock, configure the following components: Inference models: Select the LLMs you wish to evaluate. Select the LLMs you wish to evaluate. Evaluation metrics: Define criteria such as correctness, coherence, and helpfulness to measure performance. Define criteria such as correctness, coherence, and helpfulness to measure performance. Test prompts: Use a diverse set of queries to evaluate the chatbot's ability to handle different scenarios. Store the evaluation results in your designated S3 bucket for further analysis. This structured approach ensures that the evaluation process is both comprehensive and repeatable. Step 5: Analyze Evaluation Results Once the evaluation is complete, review the results to assess the model's performance. Focus on key metrics such as correctness, coherence, and helpfulness to determine how effectively the chatbot answers questions. Compare the model's outputs with reference responses and ground truth data to identify discrepancies. Use performance distributions and other analytical tools to pinpoint areas that require improvement. This step is crucial for refining the model and making sure it meets user expectations. Step 6: Compare Models If you are testing multiple models, such as Nova Pro and Nova Premiere, use the evaluation results to compare their performance. Visualize differences in metrics to identify which model aligns best with your specific requirements. This comparison enables you to make an informed decision about which model to deploy, making sure optimal performance for your use case. Key Takeaways Evaluating LLMs is an essential step in deploying reliable and effective AI solutions. Amazon Bedrock simplifies this process by providing tools to test and compare models, integrate external knowledge bases, and customize evaluation metrics. By following this guide, you can optimize your AI implementations, making sure they meet user needs and deliver consistent, high-quality results. Media Credit: Matthew Berman Filed Under: AI, Guides Latest Geeky Gadgets Deals Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, Geeky Gadgets may earn an affiliate commission. Learn about our Disclosure Policy.

Customer service AI startup Decagon raises $131 million
Customer service AI startup Decagon raises $131 million

Reuters

timean hour ago

  • Reuters

Customer service AI startup Decagon raises $131 million

SAN FRANCISCO, June 23 (Reuters) - Decagon, a startup providing customer service solutions powered by artificial intelligence, said on Monday it had raised $131 million in a funding round that valued it at $1.5 billion. The San Francisco company's Series C round was led by venture capital firms Accel and Andreessen Horowitz, with existing investors A*, Bain Capital Ventures and BOND participating. Avra, Forerunner and Ribbit Capital also took part in the round. The fundraise comes less than a year after a Series B round, when it raised $65 million at a $650 million valuation. The funding establishes Decagon as one of the most highly valued AI startups providing AI customer support solutions. Its competitors include giants like Salesforce (CRM.N), opens new tab and startups like Sierra, helmed by OpenAI board chairman and former Salesforce co-CEO Bret Taylor. In October, Sierra raised $175 million in a funding round giving it a $4.5 billion valuation. Since OpenAI's ChatGPT burst on the scene, investor interest in funding AI technology has gradually shifted from expensive foundation models to applications that generate steady revenue. Decagon serves clients such as Hertz, Duolingo (DUOL.O), opens new tab, Eventbrite (EB.N), opens new tab and Chime (CHYM.O), opens new tab. The company will use its funding to grow the team and sell into more enterprises, its CEO Jesse Zhang told Reuters. Financial services company Chime saw a 60% reduction in contact center costs from using Decagon and a doubling of its net promoter score, a measure of customer satisfaction, Zhang said. "When AI can take action and solve things, customers can get what they want much faster and more consistently," he said. The company provides both text-based AI customer support solutions like chat and email as well as AI voice agents that are capable of end-to-end customer support calls. Decagon's software tools also allow companies to create their own AI customer support solutions, Zhang said.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store