Categories
Artificial Intelligence Hardware What I’m Up To

Specs for NVIDIA’s GB10 chip, which powers HP’s ZGX Nano G1n AI workstation

I’m currently working with Kforce as a developer relations consultant for HP’s new tiny desktop AI powerhouse, the ZGX Nano (also known as the ZGX Nano G1n). If you’ve wondered about the chip powering this machine, this article’s for you!

The chip powering the ZGX Nano is NVIDIA’s GB10, a combination CPU and GPU where “GB” stands for “Grace Blackwell.” The chip’s two names stand for each of its parts…

Grace: The CPU

The part named “Grace” is an ARM CPU with 20 cores, arranged in ARM’s big.LITTLE (DynamIQ) architecture, which is a mix of different kinds of cores for a balance of performance and efficiency:

    • 10 Cortex-X925 cores. These are the “performance” cores, which are also sometimes called the “big cores.” They’re designed for maximum single-thread speed, higher clock frequencies, and aggressive out-of-order execution, their job is to handle bursty, compute-intensive workloads such as gaming and rendering, and on the ZGX Nano, they’ll be used for AI inference.
    • 10 Cortex-A725 cores. These are the “efficiency” cores, which are sometimes called the “little cores.” They’re designed for sustained performance per watt, running at lower power and lower clock frequencies. Their job is to handle background tasks, low-intensity threads, or workloads where power efficiency and temperature control matter more than peak speed.

Blackwell: The GPU

The part named “Blackwell’ is NVIDIA’s GPU, which has the following components:

    • 6144 neural shading units, which act as SIMD (single-instruction, multiple data) processors that act as “generalists,” switching between standard graphics math and AI-style operations. They’re useful for AI models where the workloads aren’t uniform, or with irregular matrix operations that don’t map neatly into 16-by-16 blocks.
    • 384 tensor cores, which are specialized matrix-multiply-accumulate (MMA) units. They perform the most common operation in deep learning, C = A × B + C, across thousands of small matrix tiles in parallel. They do so using mixed-precision arithmetic, where there are different precisions for inputs, products, and accumulations.
    • 384 texture mapping units (TMUs). These can quickly sample data from memory and do quick processing on that data. In graphics, these capabilities are use to resize, rotate, and transform bitmap images, and then paint them onto 3D objects. When used for AI, these capabilities are used to perform bilinear interpolation (used by convolutional neural network layers and transformers) and sample AI data.
    • 48 render output units (ROPs). In a GPU, the ROPs are the final stage in the graphics pipeline — they convert computed fragments into final pixels stored in VRAM. When used for AI, ROPs provide a way to quickly write the processing results to memory and perform fast calculations of weighted sums (which is an operation that happens with all sorts of machine learning).

128 GB of unified RAM

There’s 128GB of LPDDR5X-9400 RAM built into the chip, a mobile-class DRAM type designed for high bandwidth and energy efficiency:

  • The “9400” in the name refers to its memory bandwidth (the speed at which the CPU/GPU can move data between memory and on-chip compute units) of 9.4 Gb/s per pin. Across a 256-bit bus, this provides almost 300 GB/s peak bandwidth

  • LPDDR5X is more power-efficient than HBM but slower; it’s ideal for compact AI systems or edge devices (like the ZGX Nano!) rather than full datacenter GPUs.

As unified memory, the RAM is shared by both the Grace (CPU) and Blackwell (GPU) portions of the chip. That’s enough memory for:

  • Running large-language-model inference up to 200 billion parameters with 4-bit weights

  • Medium-scale training or fine-tuning tasks

  • Data-intensive edge analytics, vision, or robotics AI

Because the memory is unified, it means that the CPU and GPU share a single physical pool of RAM, which eliminates explicit data copies.

The RAM is linked to the CPU and GPU sections using NVIDIA’s C2C (chip-to-chip) NVLINK , their low-power interconnector that lets CPU/GPU memory traffic move at up to 600 GB/s aggregate. That’s faster than PCIe 5! This improves latency and bandwidth for workloads that constantly exchange data between CPU preprocessing and GPU inference/training kernels.

Double the power with ConnectX

If the power of a single ZGX Nano wasn’t enough, there’s NVIDIA’s ConnectX technology, which is based on a NIC that provides a pair of 200 GbE ports, enabling the chaining/scaling out of workload across  two GB10-based units. The doubles the processing power, allowing you to run models with up to 400 billion parameters!

The GB10-powered ZGX Nano is a pretty impressive beast, and I look forward to getting my hands on it!

 

Categories
Artificial Intelligence Current Events Tampa Bay What I’m Up To

“Back to the Future of Work” covered in Tampa Bay Business & Wealth!

Last week’s panel event, Back to the Future of Work, was featured in Tampa Bay Business and Wealth!

Taking place at the Reliaquest Auditorium in Tampa startup space Embarc Collective, the event featured a discussion about different ways to think about how we measure the value of work in the new world of AI, remote work, ubiquitous internet, and economic uncertainty.

On the panel were:

Check out the article, From hours to outcomes: Tampa panel explores the future of work in an AI world!

 

Categories
Artificial Intelligence Meetups Tampa Bay What I’m Up To

I’ll be speaking at the Indo-US Chamber of Commerce’s “Practical AI” panel — Tuesday, September 16 at 6:30 p.m.!

Promotional poster for “Practical AI: Turning Technology into Business Value” featuring moderator Sam Kasimalla and panelists Joey de Villa, Sudeep Sarkar, and Priya Balasundaram.
Click the image to reserve your ticket!

Want to learn how AI can be used in your business or career, meet key people from Tampa Bay’s dynamic South Asian community, and enjoy some Indian food? Then you’ll want to attend the Indo-U.S. Chamber of Commerce’s panel, Practical AI: Turning Technology into Business Value, taking place next Tuesday, September 16th at 6:30 p.m. at Embarc Collective!

The tl;dr

Why Attend?

  • Learn how AI can be applied beyond theory to solve real business challenges.

  • Hear from leaders in academia, entrepreneurship, and applied technology.

  • Network with Tampa Bay’s growing AI and tech community.

  • Enjoy complimentary Indian cuisine while connecting with innovators and peers.

Speakers & Moderator

Register for this event!

Once again, this is a free event, and there’ll be a complimentary Indian dinner. Register now!

Categories
Artificial Intelligence Tampa Bay What I’m Up To

Scenes from Fractio’s “Back to the Future of Work” event (September 4, 2025)

Last night, Fractio hosted Back to the Future of Work, an event built around a panel discussion about changing the way we assign value to work in the age of AI.

I arrived early to set up my computer to run the pre-panel video…

…and check that the panel seats and mics were set up properly…

…then had some quick breakfast-for-dinner (which was symbolic of how our understanding of paying for time was about to be turned upside-down):

The event took place at Embarc Collective, who’d set up the room in a way that would let people comfortably eat “brinner” while watching the panel…

…and when the doors open, a room-packing crowd came in.

After a little time to let people get their food, breakfast cocktail, and mingle, they were seated…

…and the panel got under way!

Fatin Kwasny, organizer of the panel and Fractio CEO, moderated…

…and the panel got started.

From left to right, the panelists were:

I enjoyed participating on the panel, and it appears that my fellow panelists did as well! I also heard from many audience members who found the event informative and entertaining.

Thanks to Florida CFO Group for sponsoring breakfast-for-dinner and breakfast cocktails…

…and to Byron Reese for providing us with copies of his book, We are Agora, to give to attendees!

 

Categories
Hardware What I’m Up To

The room where it happens

Joey de Villa’s home office. It has a shiny hardwood floor, two desks in an L-shaped configuration, monitors, keyboards, synthesizers, and other gear. A large octopus art piece looms over the back wall.
Tap to view at full size.

For the curious, here’s a recent pic of my home office, a.k.a. “The Fortress of Amplitude.” The gear configuration changes every now and then, but it generally looks like this. It’s where the magic happens!

Categories
Artificial Intelligence Hardware What I’m Up To

Quick announcement: I’m doing developer relations for HP’s new ZGX Nano AI computer!

Just so you know: today’s my first day at Kforce doing developer relations for HP! More specifically, for HP’s ZGX Nano, a tiny computer designed specifically for running large AI models right on your desktop…and not on someone else’s computers!

The ZGX Nano packs a ridiculous amount of power into a tiny space…

Powered by NVIDIA’s GB10 GPU and a 20-core ARM CPU sharing 128GB of RAM, the ZGX Nano performs at 1,000 teraflops (1 petaflop), which is 1015 floating-point operations per second. It’ll support an AI model taking in 200 billion parameters — 400 billion if you connect two ZGX Nanos together.

I’m getting set up for day one on the job as I write this, so I’m keeping this post short and ending with this gem from a little while back: HP’s Rules of the Garage:

Categories
Career What I’m Up To

The interview where I ALMOST succeeded

Last week, I interviewed for a developer relations leadership role at a company whose product I genuinely use and admire.

I made it to round 2 of 3, but ultimately wasn’t selected.

While I didn’t land the job and a chance to work with an amazing company and incredible team, I’m honored to have been considered and incredibly proud of the work I put in:

  • 30+ hours of research and preparation
  • 100+ slides across two presentations
  • Some of the most meaningful conversations I’ve had with a team in years

I could simply throw up my hands in resignation and leave all that work and content to languish in a folder on a backup drive or in the cloud…

…but instead, I’m sharing it here. Why?

Because:

  • Good ideas deserve to circulate! Maybe there’s a framework, approach, or creative solution in my presentations that could help someone else.
  • We’re not alone in this. The job market is tough right now, and I want people to know they’re not the only ones putting in extraordinary effort. I know you’re out there, giving it your all!
  • Transparency builds community and helps others. Real examples of strategic work are worth way more than hand-waving abstract advice.

I’m sharing the slides that outline my complete developer relations strategy presentation plus my tactical execution plan, anonymized and annotated with speaker notes. You’ll see my “Foundation / Focus / Flywheel” framework, community engagement strategies, and how I approached everything from attribution tracking to expanding into Europe.

  • If you’re job searching: Take what’s useful here and build on it! No attribution necessary. We’re all in this together, and your success doesn’t diminish mine.
  • If you found this valuable: Please share it! It helps me and others going through this.
  • If you’re hiring: This shows how I think about DevRel strategy. If your team builds for developers, let’s chat!

My philosophy is that either I win or I learn. I learned a lot from this process, and I’m grateful for the opportunity to have engaged with such a thoughtful team.

Next time, I just might win.

Here’s the link to my presentations.