Advanced Computing in the Age of AI | Saturday, July 20, 2024

NeuReality Launches Developer Portal for NR1 Inference Platform, Expanding Affordable AI Access 

SAN JOSE, Calif., April 16, 2024 -- NeuReality, an AI infrastructure technology company, announced today the release of a new software developer portal and demo for easy installation of its full software stack and APIs. This marks a significant milestone for NeuReality since delivery and activation of its 7nm AI inference server-on-a-chip, the NR1 NAPU, and successful bring up of its entire NR1 AI hardware and software system in the first quarter.

The NR1 AI Inference Solution enables businesses and governments to run new AI training models and existing AI applications without over-investing millions in scarce GPUs. Regardless of AI accelerator performance, CPUs remain the primary performance bottlenecks in AI Inference, resulting in excessive power consumption and cost, making the most exciting AI innovations impossible for the majority of organizations to install and operate today.

The NR1 system was deemed customer-ready in Q1 2024 after the NAPU arrived from TSMC Taiwan in December, followed by a successful bring-up and system integration in just 90 days. “Completing the seamless bring-up and integration of a sophisticated server-on-a-chip device and a comprehensive hardware/software AI system within a small startup team is nothing short of remarkable," said Ilan Avital, chief R&D officer at NeuReality.

The system successfully met target functionality and performance requirements, covering server-on-chip (SOC), IP, and software aspects. This achievement marked its readiness for early customer pilots, particularly with cloud service providers, financial services, and healthcare sectors for computer vision, automatic speech recognition and natural language processing – laying an affordable foundation for generative AI, multi-modality, and more advanced technologies to come. NeuReality attributed the swift bring-up process of the NR1 system to robust architecture and rigorous emulation testing conducted before the 2023 tape out in collaboration with Synopsys.

The accompanying Software Development Kit (SDK) is designed exclusively for high-volume, high-variety AI workloads in enterprise data centers. It contains hierarchical tools for all types of compute engines and XPUs, along with optimized partitioning – making it easy to install, manage, and scale while giving developers their time back from the traditional complexity of deploying AI Inference.

NeuReality’s solution delivers an unprecedented developer experience with significant flexibility to deploy the most advanced and complex AI pipelines more easily, based on the specific needs of their projects. It empowers developers with toolchain for complete AI pipeline acceleration, orchestration, provisioning and inference runtime APIs to streamline AI deployment workflow. All of this and more is now documented in a new technical whitepaper to inspire innovators to focus urgently on end-to-end data center efficiency and resource optimization to enable affordable AI deployments.

Citing a 35% AI adoption rate globally and lower than 25% rate in the U.S., NeuReality is focused on lowering market barriers to mainstream industries. “It’s simply out of reach to the majority of businesses,” added Avital. “We can start changing that now by reducing high power consumption at the source – and educating customers that the ideal AI Inference servers require fundamentally different and more efficient server configurations than big supercomputers and high-end GPUs used in AI training.”

For example, NeuReality's NR1-S AI Inference Appliance outperforms the Nvidia DGX H100 System with the same deep learning processing performance but packed with 6x the data processing performance, half the price, one-third of the energy consumption, and half the physical space – all without requiring a host CPU in the system. The NR1 engineering involved packing 6.5x more processing power onto the NR1 NAPU, equivalent to 830 CPU cores in a single 4u chassis, while at the same time having enough power to host 10 Nvidia GPUs or any AI accelerator.

“At Cleveland Avenue, we see the immense potential of disruptive AI technologies like NeuReality to revolutionize the retail landscape, particularly within our focus on investments in restaurants, foodtech, and beverage,” said Mingu Lee, managing partner, technology investments at Cleveland Avenue.

“With NeuReality's breakthroughs in running large-scale AI models such computer vision for retail analytics, conversational AI for in-store and online virtual assistants and personalized recommendations, and generative AI-powered drive-throughs, we're not just investing in technology; we're investing in democratizing AI for ‘have not businesses’ ensuring that even those with razor-thin profit margins can deliver exceptional customer experiences and efficient business processes with the help of AI," added Lee.

Access the new Software Developer Portal and learn more at:

About NeuReality

Founded in 2019 by a seasoned team of system engineers, NeuReality Ltd. is an AI technology innovation company that creates purpose-built AI Inference system architecture, silicon, hardware, and software for the ultra-scalability of current and future AI applications. Its cutting-edge technology transforms how companies operate daily AI inferencing with its holistic, ready-to-use NR1 AI Inference Solution that supports limitless deep learning models and customer choice in hardware providers and open-source software. In its quest to democratize AI and unleash greater human achievements, NeuReality AI solutions are easily accessible, adaptable, and affordable for all governments and businesses large and small – with a robust set of leading industry partners to deliver and deploy. For more information, visit

Source: NeuReality