Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
  • AI
  • Tech
  • Cybersecurity
  • Finance
  • DeFi & Blockchain
  • Startups
  • Gaming
Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
Dataconomy
No Result
View All Result

Rain AI shook hands with OpenAI

Neuromorphic processors developer RainAI reaches $51 million deal with AI giant OpenAI

byEmre Çıtak
December 4, 2023
in Artificial Intelligence, News
Home News Artificial Intelligence

OpenAI, the leading non-profit AI research company, has signed a letter of intent with Rain AI, a startup developing neuromorphic processors (NPUs) specifically tailored for AI applications. The agreement outlines OpenAI’s commitment to purchasing $51 million worth of Rain AI’s NPUs upon their commercial availability.

In recent years, the need for specialized hardware has become increasingly evident for AI technologies. AI companies are pushing their limits to offer personalized experiences to their customers. Rain AI’s NPUs, dubbed “Digital Dendrites,” are inspired by the structure and function of the human brain. They are designed to mimic the brain’s ability to process information efficiently and with low power consumption, making them a promising solution for the demanding computational needs of AI tasks.

At the heart of this partnership lies a letter of intent signed in 2019, outlining OpenAI’s intention to purchase $51 million worth of Rain AI’s NPUs upon their commercial availability. This substantial investment underscores OpenAI’s belief in the transformative potential of Rain AI’s technology.

Stay Ahead of the Curve!

Don't miss out on the latest insights, trends, and analysis in the world of data, technology, and startups. Subscribe to our newsletter and get exclusive content delivered straight to your inbox.

Rain AI
The agreement outlines OpenAI’s commitment to purchasing $51 million worth of Rain AI’s NPUs upon their commercial availability (Image credit)

What is Rain AI?

Rain AI is a startup company that is developing neuromorphic processors (NPUs), which are a type of computer chip that is designed to mimic the structure and function of the human brain. NPUs are seen as a promising technology for artificial intelligence (AI) applications because they have the potential to be much more energy efficient than traditional GPUs, which are currently the most widely used type of AI accelerator.

Rain AI’s NPUs, dubbed “Digital Dendrites,” are still in development, but the company has claimed that they could potentially be 100 times more powerful and 10,000 times more energy efficient than GPUs. The company has also said that its NPUs could be used to train AI models as well as run them once they are deployed.

Rain AI has received funding from a number of high-profile investors, including OpenAI CEO Sam Altman and Aramco Ventures, the venture capital arm of Saudi Aramco. The company is also reportedly working with a number of large technology companies, including Google, Oracle, Meta, Microsoft, and Amazon.

Despite its promising technology and high-profile backers, Rain AI faces a number of challenges. The company’s NPUs are still in development, and it is unclear when they will be commercially available. Additionally, Rain AI will need to compete with a number of established players in the AI hardware market, including Nvidia, Google, and Amazon.

How do NPUs work?

As mentioned before, neuromorphic processors (NPUs) are a type of artificial intelligence (AI) accelerator designed to mimic the structure and function of the human brain. Unlike traditional processors that rely on sequential instructions, NPUs are built on a network of interconnected artificial synapses, similar to the neural connections in the brain.

This architecture allows NPUs to process information in a parallel and distributed manner, making them well-suited for the computationally intensive tasks involved in AI applications.

Here is a detailed breakdown of how NPUs work:

  1. Input and synapses: Information is fed into the NPU through its input layer, where it is represented as spikes or pulses of electrical activity. These spikes are then transmitted to the artificial synapses, which act as the basic computational units of the NPU
  2. Synaptic weighting: Each synapse has a weight associated with it, which determines the strength of the connection between the input and output neurons. These weights are dynamically adjusted during the training process, allowing the NPU to learn from the data and improve its performance
  3. Neuronal activation: When a spike arrives at a synapse, it contributes to the activation of the postsynaptic neuron. If the cumulative activation exceeds a certain threshold, the neuron fires, generating a spike of its own. This process is replicated throughout the network, allowing the NPU to perform complex computations
  4. Learning and adaptation: NPUs are capable of learning and adapting to new data by adjusting the weights of their synapses. This process, known as synaptic plasticity, mimics the way the human brain learns and forms new memories
  5. Parallel processing: Unlike traditional processors that execute instructions sequentially, NPUs perform computations in parallel across the network of interconnected neurons. This parallel architecture allows NPUs to achieve significantly higher processing speeds for AI workloads

In summary, NPUs offer several advantages over traditional processors for AI applications, including:

  • Parallel processing for faster computations
  • Synaptic plasticity for learning and adaptation
  • Energy efficiency for power-constrained devices
Rain AI
Rain AI’s NPUs aim to efficiently process information with low power consumption, addressing the demanding computational needs of AI tasks (Image credit)

How can Rain chips help with the development of AI?

Neural processing units (NPUs) are specialized hardware accelerators designed to handle the computationally demanding tasks of artificial intelligence (AI), particularly machine learning and deep learning. They offer several advantages over traditional CPUs and GPUs, making them increasingly crucial for the development of AI.

Enhanced performance and efficiency

NPUs are specifically designed to accelerate the mathematical operations commonly used in AI, such as matrix multiplications and convolutions. This specialized architecture allows them to perform these operations significantly faster and more efficiently than CPUs and GPUs, leading to substantial improvements in AI model training and inference speed.

Lower power consumption

NPUs are designed to be highly energy-efficient, consuming significantly less power than CPUs and GPUs while delivering comparable or even superior performance. This energy efficiency is particularly important for edge devices, such as smartphones and IoT devices, where battery life is a critical constraint.

Specialized features for AI workloads

NPUs often incorporate specialized features tailored for AI workloads, such as hardware-accelerated activation functions and support for various data formats commonly used in AI models. These features further enhance the performance and efficiency of AI applications.

Reduced latency

NPUs offer lower latency compared to CPUs and GPUs, meaning they can process data and produce results more quickly. This is particularly important for real-time AI applications, such as autonomous vehicles and natural language processing.

Rain AI
Rain AI claims its NPUs could be 100 times more powerful and 10,000 times more energy-efficient than traditional GPUs (Image credit)

Enabling more complex AI models

The enhanced performance and efficiency of NPUs enable the development and execution of more complex and demanding AI models that would not be feasible with conventional CPUs and GPUs. This enables the advancement of areas such as natural language processing, image analysis, and machine learning.

In short, NPUs are a critical technology for the development of AI. They offer a number of advantages over traditional hardware, making them well-suited for complex and demanding AI tasks.

As AI technology continues to advance, NPUs are expected to play an even more important role in enabling the next generation of AI applications.

Altman returns

Sam Altman’s eventful weeks with OpenAI are over and he’s back to work. While the AI company was expected to take a lot of damage in the process, on the contrary, the unveiling of Q-star, ChatGPT with voice, and now Rain AI investment once again shows many in the industry why OpenAI is a successful company.

Artificial intelligence companies have a huge stake in our future and every day we are faced with a new invention. The widespread use of AI, which started with slow and malfunctioning chatbots at the beginning of 2023, is a matter of curiosity as to what more it will show us a month before 2024.

As users and reporters, all we can do is hope that a fully autonomous life in the cartoons awaits us, not the chaotic future of sci-fi.


Featured image credit: Rain AI.

Tags: FeaturedopenAI

Related Posts

Texas Attorney General files lawsuit over the PowerSchool data breach

Texas Attorney General files lawsuit over the PowerSchool data breach

September 5, 2025
iPhone 17 Pro is expected to arrive with 48mp telephoto, variable aperture expected

iPhone 17 Pro is expected to arrive with 48mp telephoto, variable aperture expected

September 5, 2025
AI chatbots spread false info in 1 of 3 responses

AI chatbots spread false info in 1 of 3 responses

September 5, 2025
OpenAI to mass produce custom AI chip with Broadcom in 2025

OpenAI to mass produce custom AI chip with Broadcom in 2025

September 5, 2025
When two Mark Zuckerbergs collide

When two Mark Zuckerbergs collide

September 5, 2025
Deepmind finds RAG limit with fixed-size embeddings

Deepmind finds RAG limit with fixed-size embeddings

September 5, 2025

LATEST NEWS

Texas Attorney General files lawsuit over the PowerSchool data breach

iPhone 17 Pro is expected to arrive with 48mp telephoto, variable aperture expected

AI chatbots spread false info in 1 of 3 responses

OpenAI to mass produce custom AI chip with Broadcom in 2025

When two Mark Zuckerbergs collide

Deepmind finds RAG limit with fixed-size embeddings

Dataconomy

COPYRIGHT © DATACONOMY MEDIA GMBH, ALL RIGHTS RESERVED.

  • About
  • Imprint
  • Contact
  • Legal & Privacy

Follow Us

  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
No Result
View All Result
Subscribe

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy Policy.