Welcome to Nural's newsletter where you will find a compilation of articles, news and cool companies, all focusing on how AI is being used to tackle global grand challenges.

Our aim is to make sure that you are always up to date with the most important developments in this fast-moving field.

Packed inside we have

  • OpenAI takes image generation to the next level with DALL-E-2
  • DeepMind discovers the optimal balance for large language models
  • and, what AlphaFold did next ...

If you would like to support our continued work from £1 then click here!

Graham Lane & Marcel Hedman

Key Recent Developments

DALL-E-2, the future of AI research, and OpenAI’s business model

DALL-E 2, the future of AI research, and OpenAI’s business model
DALL-E 2 shows how far the AI research community has come toward harnessing the power of deep learning and addressing some of its limits.

What: OpenAI have introduced a new version of DALL·E, a deep learning network that generates images from text descriptions.  DALL-E-2 produces much more accurate images and has an uncanny knack of correctly interpreting the text prompts. What is more, the new version can edit images, for example adding a vase of flowers to a coffee table.
Key Takeaways: The capabilities of the new version have impressed even world-weary AI experts. One of the early users discusses the strengths and weaknesses of the new system. Despite these advances, however, OpenAI acknowledges that they are still struggling to address many know issues. For example, all nurses, personal assistants and flight attendants generated by the model are female, whereas the lawyers, CEOs and builders are all male. For this reason, the system is not yet available to the public. There is still a long way to go …
Blog:  DALL·E 2

DeepMind has found a way to efficiently scale large language models

Training Compute-Optimal Large Language Models: DeepMind’s 70B Parameter Chinchilla Outperforms 530B Parameter Megatron-Turing | Synced
Today’s extreme-scale language models have demonstrated astounding performance on natural language processing tasks, attributed mainly to their ever-expanding size, which can surpass 500 billion parameters. But as such models have scaled dramatically in recent years, the amount of data used in their…

What: DeepMind researchers investigated the optimal relation between the size of an AI model (number of different parameters) and the size of the training data (the number of “tokens”). They found that current large language models are “undertrained” because the size of the model has been increased whereas the size of the training data has been kept constant. Through experimentation they found that, for best results, the size of the model and the size of the training data should be scaled equally. They then trained a model, called Chinchilla, with the optimal ratio (using a smaller model but more training data) and found that it “significantly outperformed” well-known large language models on a range of tasks when both models were trained using the same amount of computing resource.
Key Takeaways: Large Language Models are criticised as being responsible for unsustainable levels of carbon emissions and being hugely expensive. In an attempt to address this, DeepMind has developed a way of finding the optimal balance between parameters and data that uses the least computing resource.
Paper: An empirical analysis of compute-optimal large language model training

What's next for AlphaFold and the AI protein-folding revolution

What’s next for AlphaFold and the AI protein-folding revolution
DeepMind software that can predict the 3D shape of proteins is already changing biology.

What: In July 2021 DeepMind released AlphaFold2 that predicts how proteins fold into complex 3D shapes. It soon announced that it had predicted the 3D structure of nearly every protein made by humans and has subsequently predicted nearly half of all known proteins. The technology is now being applied in a wide range of different areas, for example examining the Nuclear Pore Complex, and the structure of bacterial proteins that promote the formation of ice.
Key Takeaways: AlphaFold2 is best used to provide an initial approximation to be validated by experiment because it isn’t always fully accurate. The researchers working specifically on drug discovery seem a little hesitant in their assessment – one talks of “critical optimism” and another sees AlphaFold2 speeding up the experimental process rather than replacing it. Accuracy and details are important in developing new drugs.
In late 2021, DeepMind launched a spin-off, IsoMorphic Labs, to apply AlphaFold2 to drug discovery. Since then, the company has said little about its work. We wait to see how that is developing ...

Image from the linked article: What's next for AlphaFold and the AI protein-folding revolution

AI Ethics

🚀 To truly target hate speech, moderation must extend beyond civility

A study has shown that white supremacists avoid vulgar and toxic language that trigger automated systems, instead emphasising their whiteness by appending “white” to many terms, such as “white children”, “white women” and “the white race".

🚀 New technology, old problems

The missing voices in Natural Language Processing.

🚀 California suggests taking aim at AI-powered hiring software

A proposed amendment to California's hiring discrimination laws would make it illegal for businesses and employment agencies to use automated-decision systems to screen out applicants who are considered a protected class.

Other interesting reads

🚀 Daedalus special edition - AI & Society

Daedalus is an open access journal published by the American Academy Of Arts & Sciences. The Spring 2022 edition is dedicated to the theme of AI and Society featuring 25 in-depth articles by leading figures in the field.

🚀 New ML models can predict adverse outcomes after abdominal hernia surgery with high accuracy

Procedure-specific risk calculator has potential to encourage changes in patient behaviors before surgery to improve success in abdominal wall reconstruction operations.

🚀 What AI can do for climate change, and what climate change can do for AI

To tackle the climate crisis, AI is becoming more open and democratic.

Cool companies found this week


Diligent - provide robots to support patient healthcare teams. Their robot called Moxi has been the feature of numerous promotional videos. The company has raised $30 million in round B funding.


Lucinity - provides a system that combines human expertise and AI technology to fight against money laundering. It has been recognised as one of 10 startups riding the wave of AI innovation and last received venture funding of $6 million in July 2020.

Data privacy

Lightbeam - provides a data privacy automation platform using AI to identify personal data. The company has emerged from stealth with $4.5 million in seed funding.

And finally ...


AI/ML must knows

Foundation Models - any model trained on broad data at scale that can be fine-tuned to a wide range of downstream tasks. Examples include BERT and GPT-3. (See also Transfer Learning)
Few shot learning - Supervised learning using only a small dataset to master the task.
Transfer Learning - Reusing parts or all of a model designed for one task on a new task with the aim of reducing training time and improving performance.
Generative adversarial network - Generative models that create new data instances that resemble your training data. They can be used to generate fake images.
Deep Learning - Deep learning is a form of machine learning based on artificial neural networks.


Marcel Hedman
Nural Research Founder

If this has been interesting, share it with a friend who will find it equally valuable. If you are not already a subscriber, then subscribe here.

If you are enjoying this content and would like to support the work financially then you can amend your plan here from £1/month!