NeuReality has said that it is looking to redefine today’s outdated AI system architecture by developing an AI-centric inference platform based on a new type of System-on-Chip (SoC).
NR1-P is NeuReality’s first implementation of this new AI-centric architecture, with other implementations to follow. This prototype platform validates the technology and allows customers to integrate it in orchestrated data centres and other facilities
The company said that it was highly focused on its follow-on NR1 SoC, which will leverage its unique fusion architecture to bring a 15X improvement in performance per dollar compared to the available GPUs and ASICs offered by deep learning accelerator vendors.
The prototype platform aims to accelerate the growth of real-life AI applications such as public safety, e-commerce, social networks, medical use-cases, digital assistants, recommendation systems and Natural Language Processing (NLP).
The NR1-P solution targets cloud and enterprise datacentres, alongside carriers, telecom operators and other near edge compute solutions and NeuReality said that it expects these enterprise customers to hit scalability and cost barriers soon after completing their shift from pilots to production deployment of their AI use cases.
NeuReality’s NR1-P is built in a 4U server chassis equipped with sixteen Xilinx Versal Adaptive Compute Acceleration Platform (ACAP) cards, highly integrated and fully software programmable compute platforms that can adapt to evolving and diverse algorithms and go far beyond traditional FPGAs. The NR1-P SoC leverages these capabilities to implement the AI-centric compute engine while fusing unique system functions with datapath functions and the neural network processing function. All these come together to optimise critical AI-inference flows while removing system bottlenecks.
According to Moshe Tanach, NeuReality’s Founder and CEO, “Datacenter owners and AI as-a-service customers are already suffering from the operational cost of their ultra-scale use of compute and storage infrastructure. This problem is going to be further aggravated when their AI based features will be deployed in production and widely used. For these customers, our technology will be the only way to scale their deep learning usage while keeping the cost reasonable.
"Our mission was to redefine the outdated system architecture by developing AI-centric platforms based on a new type of System-on-Chip. Our platforms provide ideal compute solutions for these new types of AI use cases. With NR1-P, customers can finally experience this new type of solution, leverage it in their solutions, and prepare to integrate our upcoming NR1 product once it is in production”.
According to NeuReality’s own benchmark testing, the NR1-P prototype platform delivers more than a 2X improvement in Total Cost of Ownership (TCO) compared to NVIDIA’s CPU-centric platforms running disaggregated deep learning service. The current achievement also paves the way towards additional dramatic improvements that will be integrated into NR1, NeuReality’s flagship product.
“Our technology provides a huge reduction in both CAPEX and OPEX for infrastructure owners. Customers using large scale use cases of image processing such as object detection or face recognition can already get 50% saving in the cost of the technology or twice the capacity for the same price when using NR1-P," added Tanach.
NeuReality emerged from stealth in February 2021, with $8m in seed investment from Cardumen Capital, OurCrowd and Varana Capital, and the appointment of AI luminary Dr. Naveen Rao, former General Manager of Intel’s AI Products Group, to NeuReality’s Board of Directors.