OpenAI and other firms are using synthetic data to train AI models

Skirts complaints related to IP abuse, privacy and data access

clock • 3 min read
Advanced AI models now being trained using computer-made 'synthetic' data
Image:

Advanced AI models now being trained using computer-made 'synthetic' data

Major tech firms developing generative AI models are actively exploring a new approach to acquiring the vast amounts of information they need for their advanced models: creating it from scratch using computer-generated data.

Players like Microsoft, OpenAI, and Cohere are employing synthetic data to train their large language models (LLMs), primarily due to the constraints in the availability of human-made data. Micr...

To continue reading this article...

Join Computing

  • Unlimited access to real-time news, analysis and opinion from the technology industry
  • Receive important and breaking news in our daily newsletter
  • Be the first to hear about our events and awards programmes
  • Join live member only interviews with IT leaders at the ‘IT Lounge’; your chance to ask your burning tech questions and have them answered
  • Access to the Computing Delta hub providing market intelligence and research
  • Receive our members-only newsletter with exclusive opinion pieces from senior IT Leaders

Join now

 

Already a Computing member?

Login

You may also like
Microsoft, Google and Snap report strong quarterly results, IBM and Intel less so

Finance and Reporting

Microsoft and Google see AI investments bearing fruit

clock 26 April 2024 • 4 min read
Big Tech's AI spending spree worries investors

Artificial Intelligence

Zuckerberg says building a leading AI system will take several years and require significant investment

clock 26 April 2024 • 3 min read
BlueVoyant chooses Leeds for new Security Operations Centre

Security

The new SOC will boost customer compliance with regulations like NIS2 and DORA

clock 25 April 2024 • 4 min read

More on Big Data and Analytics

Even CERN has to queue for GPUs. Here's how they optimise what they have

Even CERN has to queue for GPUs. Here's how they optimise what they have

'There's a tendency to say that all ML workloads need a GPU, but for inference you probably don't need them'

John Leonard
clock 17 April 2024 • 4 min read
Partner Content: Why good data is the foundation of AI success

Partner Content: Why good data is the foundation of AI success

Does your organisation have the right quantity and quality of data to make its AI ambitions a reality?

Arrow
clock 04 April 2024 • 2 min read
Partner Content: Human-in-the-loop - How AI can boost your organisational culture

Partner Content: Human-in-the-loop - How AI can boost your organisational culture

Why it’s vital to consider your organisation’s people when implementing AI

Arrow
clock 26 March 2024 • 2 min read