Categories
Artificial Intelligence Meetups

Notes from Pratik Patel’s presentation, “AI Native Architecture” (Feb. 19, 2026 @ Tampa Java User Group / Tampa Bay AI Meetup)

Last Thursday, February 19th, Tampa Java User Group welcomed Pratik Patel, Java Champion and Director of Developer Relations at Azul Systems, to give his AI Native Architecture talk at Kforce headquarters. Tampa Bay AI Meetup was happy to partner with Tampa JUG, and we thank Ammar Yusuf for the invite!

We had a pretty full room…

…followed by an accordion number…

…followed by Pratik’s presentation.

Here are my notes from the presentation:


Three kinds of AI

Here’s a fun “icebreaker” game to try at your next tech gathering: ask the room to name the three fundamental types of AI, and watch what happens.

When tried on the crowd at last Thursday’s Tampa Java User Group / Tampa Bay AI Meetup, a lot of people called “generative AI,” which was hardly a surprise.

We came close, but didn’t directly name, the second kind: predictive analysis. It’s the kind of AI that’s been quietly running inside every credit card transaction you’ve made for the past decade. It saved me a lot of headache last year when someone used my credit card number to buy enough gas to fill an F-250 in rural Georgia while I was having a poke bowl in St. Pete. A neural network detected the mismatch between the gas-guzzler purchase and my usual spending and location patterns, which led to a text from the credit card company, and my immediate “That wasn’t me” response.

None of us got the third one: time-series AI. It’s the branch that looks at data across time to spot trends and make forecasts. Not “Will Joey buy 50 gallons of gas in rural Georgia?” but “What has Joey been buying every Friday evening for the past two years, and what does that predict about next Friday?”

Pratik kicked off his talk on AI-native architecture with this. By the time he was done, we’d gotten a serious rethink of not just what kinds of AI exist, but what it actually means to build an application with AI at its core, as opposed to just bolting AI onto the side and hoping for a stock price bump.

Your data is your moat

One of the central arguments Pratik made is that data is what separates a defensible business from one that can be replicated by a developer with a generous cloud credit and a free afternoon.

He used Penske Truck Leasing as his example. Anyone can, theoretically, buy a bunch of trucks and stand up a website. What you can’t easily replicate is a decade of auction data, bidding history, customer behavioral patterns, and operational intelligence. That data is what lets Penske do something like: identify a customer who bid on a truck but didn’t win the auction, then automatically reach out to offer them a similar vehicle. The data made it obvious, and a system acted on it.

This is why the old saying “data is the new oil” is actually more apt than it sounds. Raw oil isn’t useful until it’s refined. Raw data sitting in an S3 bucket isn’t useful either until it’s refined toom by cleaning it, structuring it, and using it to power an application that your competitors simply don’t have the history to replicate. This kind of advantage that sets you apart is referred to as a moat.

In this new world, where anyone can vibe-code a decent SaaS clone in an afternoon using AI tools, your proprietary data may be that moat protecting you from someone in their mom’s basement with good taste and ambition.

The architecture stack (or: Where all this stuff actually lives)

Pratik laid out a three-layer view of what an AI application architecture actually looks like in the real world. It was a helpful maps to the “who does what” question that comes up whenever engineering teams start building this stuff.

On the left side is data acquisition and preprocessing. This comprises tools like Apache Kafka for event streaming, Apache Iceberg as a data layer that lets multiple teams share the same underlying datasets without tripping over each other, and Spark for processing data at scale. This is where collection, cleaning, and transformation happen. It’s also where most AI projects quietly die, because the data turns out to be messier than anyone admitted during planning.

In the middle is model building and fine-tuning. Pratik was direct here: your company is almost certainly not going to train its own large language model from scratch. The estimates for what it cost to train GPT-5 range from $100 million to over a billion dollars in GPU time. Unless “Uncle Larry” is personally funding your AI initiative, you’re going to use an off-the-shelf model, like OpenAI, Gemini, Claude, or one of the increasingly capable open-weight models like DeepSeek or Alibaba’s Qwen3. The Python ecosystem owns this tier for now, thanks to its long history in data science and extensive libraries, though Java options like Deep Learning4J are maturing.

On the right is inference and integration, which is where most application developers will actually spend their time. This is the code you write to orchestrate models, retrieve relevant context, handle the results, and deliver a useful experience to users. This is also where AI-native thinking diverges sharply from “AI bolted on,” which Pratik spent considerable time on.

The most important thing Pratik said all evening

Here it is: LLMs are non-deterministic, and that changes everything about how you build software.

Traditional software is built on deterministic foundations. If you write a database query that asks for a specific user profile, you will get the exact same answer every time: that user’s profile. The result is deterministic, and it’s  reliable in a way that software developers have spent the previous decades taking for granted.

LLMs don’t work that way. Ask the same question twice and you may get meaningfully different answers. That’s just a fundamental property of how token prediction and the attention mechanism work. The model doesn’t do the deterministic thing and look up an answer. Instead, it generates an answer based on probabilistic similarity to everything it has ever been trained on.

When the generated answer is wrong, we call it hallucination. But the more accurate framing is that hallucination is the shadow side of the same capability that makes these models useful at all.

(Joey’s note: I like to say “All LLM responses are hallucinations. It’s just that some hallucinations are useful.”)

For casual applications, such as “Find me a bar with karaoke near downtown Tampa,” we can put up with a certain amount of “wrongness.” You go there, find out there’s no karaoke, drink anyway, call it a night. However, for a system that’s analyzing medical imaging and flagging potential tumors, our tolerance for wrongness is zero, and “the model felt pretty confident” is not an acceptable answer.

The emerging approaches to this are interesting: evaluation frameworks built into tools like Spring AI and LangChain that let you run suites of tests against model outputs; and something called “LLM as a judge,” where you use a second model to evaluate the outputs of the first. Ask OpenAI a question, get an answer, hand both the question and the answer to Gemini and say: “Does this look right?” It’s new, it’s imperfect, and it’s the current state of the art.

The good news, as Pratik put it: everyone is early. You are not behind.

About those costs

Don’t let the $20/month subscription price fool you into thinking AI inference is cheap at scale.

Pratik made the case that inference costs are not going to come down dramatically anytime soon, and offered some uncomfortable data points in support. Moore’s Law, the Intel cofounder’s observation that transistor density on chips doubles every 18 months, is effectively dead. We’re at the sub-nanometer level of chip fabrication and at that level of miniaturization, you’re really starting to fight the laws of physics.

GPU prices have gone in the opposite direction of what you might hope: the Nvidia 5090, the top consumer-grade card, has gone from roughly $2,000 at launch to $4,000 on the secondary market. RAM prices have spiked because every data center on Earth is buying it for AI workloads. When Pratik noticed RAM prices shooting up, he moved money into Western Digital and Seagate stock. He may be onto something.

The practical upshot for developers building applications: if you’re running hundreds of evaluation tests per hour during development (which is what you should be doing, given the non-determinism problem described above)  burning frontier model tokens for all of that is going to get expensive fast.

Pratik’s solution is to do the bulk of development testing against locally-run open-weight models via Ollama. His current recommendations: qwen3-coder for coding-adjacent tasks (and it legitimately does not phone home, I’ve run Wireshark to confirm), and nemotron from Nvidia for more general work. Then switch to the frontier model for production and final evaluation. Your laptop handles the iteration, and the cloud handles the deployment.

AI-native vs. bolting AI on (or: What actually matters)

You’ve heard this story before, even if you don’t immediately recognize it.

Pratik brought up an old term: sneakernet. That’s from  the era when all software was executables running on your machine, and deploying software meant physically walking to a user’s desk with a floppy disk. Then came the cloud, and suddenly continuous deployment became a thing, and anyone still doing quarterly releases felt like a relic.

But here’s what’s easy to forget: cloud native wasn’t just about faster deploys. It forced a complete rethink of how applications are designed, how they’re operated, and how they fail. The servers went from being pets (named, tended, mourned when they died) to being cattle (anonymous, disposable, replaced without ceremony). This called for a different approach.

Pratik’s central argument is that we’re at exactly that same inflection point with AI, and that most companies are going to blow it, at least initially.

When your boss comes in and says “put some AI in the product so our stock price goes up” (Pratik confirmed this is a real conversation people are having in real offices, not a joke), the tempting response is to bolt on a RAG endpoint, add “AI-powered” to the marketing copy, and call it a day. Retrieve some relevant documents. Stuff them into a prompt. Return a plausible-sounding answer. Ship it!

That’s not AI-native. That’s sneakernet with an LLM duct-taped to it.

An AI-native system learns, adapts, and acts autonomously. Not when a user presses a button. Proactively, in response to new data, with judgment that improves over time.

Pratik described the evolution of his own download analytics system as a concrete example. It started as “AI bolted on,” with a natural language interface that let people query a Spark cluster without writing SQL. Useful. Not native.

Over the past year and a half, he rebuilt it into something different: a system that monitors weekly data feeds, detects when something has changed (for example, a spike in Java 17 downloads), connects that to relevant context from an internal knowledge base (there was a critical security patch), and proactively sends him a synthesized briefing before he even thinks to ask. He still reviews it. But the thinking now happens without him.

The hotel booking example he used to illustrate the idea is even more vivid. Pratik has a specific, consistent set of hotel preferences: he wants to be within walking distance of wherever he’s speaking, the gym needs to be a real gym (not a treadmill and a motivational poster — Hotel 5 in Seattle, I’m lookin’ right at you), and he always searches by exact address rather than city name. He does this exact sequence of clicks every single time he books a hotel. An AI-native Marriott system would see this behavioral pattern, learn from it, and surface the right three options without him having to do any of that manual filtering. Not because someone programmed “Pratik likes gyms” into a rule engine, but because the system observed his behaviors, inferred some patterns, and generalized.

Could you do all of this algorithmically? Technically, yes. But think about it: you’d be writing bespoke preference logic for millions of users with different, compounded, evolving preferences, and you’d be doing it forever. The whole point of using an LLM here is that you’re borrowing its capacity for generalization instead of hand-coding every case yourself.

Agents, fine-tuning, and a grain of salt

Pratik offered a measured take on the current agentic AI frenzy. Agents can act, but do they actually learn from what they’ve done? That’s the gap between today’s agentic frameworks and a genuinely AI-native system. Agents are probably not going away because they’re real and useful, but the framing will shift again in six months ( that’s just how this space works). The best approach is to build the fundamentals, not the hype.

On fine-tuning: if you need a model that’s deeply specialized for a domain, you don’t have to build an LLM from scratch. Low-Rank Adaptation (LoRA) lets you take an existing large model and attach a domain-specific adapter that shifts its weights toward your area of expertise. OpenAI’s recently released finance-specific model that they built in collaboration with Goldman Sachs, trained on a large corpus of financial data is exactly this. The base model does the heavy lifting. The adapter makes it fluent in corn futures.

On RAG: retrieval-augmented generation is essentially fancy-pants prompt stuffing. You find the documents most relevant to a user’s query, pull them in, and let the model reason over them. It’s the right approach for a lot of use cases, it’s not magic, and it works best when your underlying data is actually clean and well-structured. Remember the greybeard saying: “Garbage in, garbage out,” a principle that the age of AI has managed to make both more important and more dangerous, since we can now generate garbage at industrial scale.

The take-away

If you walked away from Pratik’s talk with one thing, it should probably be this: the fundamental shift AI requires isn’t technical. It’s conceptual. Just like cloud native forced you to stop thinking about servers as permanent fixtures and start thinking about them as fungible infrastructure, AI native requires you to stop thinking about AI as a feature you add to an application and start thinking about it as the substrate the application is built on.

The application that learns. The application that adapts. The application that wakes up when new data arrives and starts thinking before you ask it to.

That’s the goal. We’re early. The tools are changing fast. But the direction is clear, and the developers who internalize that shift now, rather than bolting features on and hoping for a stock price bump, are going to be the ones building the interesting stuff.

Sample apps!

If you’d like to dive deeper into what Pratik was talking about, he has companion sample apps. The details are in this picture:

Categories
Artificial Intelligence Career Security What I’m Up To

A Fake Recruiter Tried to Scam Me — I Caught Him Using ChatGPT

The newest video on the Global Nerdy YouTube channel is now online! It’s called A Fake Recruiter Tried to Scam Me — I Caught Him Using ChatGPT. Watch it now!

It’s the story of how a scammer posing as an executive recruiter tried to con me out of hundreds (and possibly thousands) of dollars using AI-generated emails, a fake job description, and a fabricated “internal document” from OpenAI.

He had me… for thirty seconds, and then I thought about it.

The short version

A “recruiter” emailed me out of the blue about a developer relations role. This isn’t out of the the ordinary; this has happened before, and it’s happened a couple of times in the past couple of months.

However, this role stood out: it was Director of Developer Relations role at OpenAI. Remote-first, $230K–$280K base, Python-primary, and AI-focused. It was basically my dream job on paper.

Over the course of several emails, he asked for my resume and salary expectations while giving me nothing concrete in return: no company name, no hiring manager, no specifics.

When I finally got suspicious and asked three simple verification questions:

  1. Who’s your contact at OpenAI?
  2. Is this a retained or contingency search?
  3. What’s your formal relationship with the hiring organization?

He went silent for over a day, then came back with a wall of text that answered none of them.

Then came the real play: he told me that OpenAI required three purportedly “professional documents” before I could interview, and they had to be ready in the next 48 hours:

  1. An “Executive Impact Matrix,”
  2. A “Technical Leadership Competency Assessment,” and
  3. A “Cross-Functional Influence & Initiative Report”

The descriptions of these documents made it look as if they were complex and would take hours to prepare. The recruiter “helpfully” offered to connect me with a “specialist” who could prepare them for a fee.

None of these documents are real. No company asks for them. It’s a document preparation fee scam, and the whole weeks-long email exchange was just the runway to get me to that moment.

But the best part? When I didn’t bite, he followed up with a fake “OpenAI Candidate Review” document showing my name alongside other “candidates” with star ratings. This would be a massive HR violation if it were real:

But it wasn’t real! He generated it with ChatGPT. And he left behind evidence — the dumbass forgot to crop out the watermark.

How the AI gave him away

One of the most interesting things about this scam is how AI was both the scammer’s greatest tool and his undoing.

Every email he sent me was written in polished, flawless corporate English.

But in the one paragraph where he steered me toward paying the “specialist,” the grammar suddenly fell apart:

“a professional I have known for years that specialise in this kind of documents with many great and positive result.”

The AI wrote the con. But the human wrote the close. And the seam between the two is where the truth leaked out.

This is a pattern worth watching for. As AI-powered scams become more common, the tell is going to be a shift in quality at the moment where the scammer needs to speak in their own words. You’ll see well-written text, abruptly followed by different writing style marked by poor, non-idiomatic grammar (because they’re communicating with you in a language they don’t know  well). Keep an eye out for that sudden transition.

The 3 questions real recruiters can answer

If you’re job searching right now and a recruiter reaches out, ask them these three questions:

  1. Who is your contact at the hiring company?
  2. Is this a retained or contingency search?
  3. What is your formal relationship with the hiring organization?

A real recruiter answers these in seconds. A fake one dodges, deflects, or disappears.

8 fake recruiter red flags

Based on my experience, here are eight things to watch out for:

  1. The job seems tailor-made for you. LLMs make it trivially easy to generate a convincing “JD” (job description) from someone’s LinkedIn profile. If it checks every single box, ask why.
  2. The information only flows one direction. They ask for your resume, salary, and preferences. They give you nothing concrete: no company name, no hiring manager, no search terms.
  3. The email footer doesn’t add up. Gmail addresses or mismatched domains, vague or incomplete street addresses, and an “alphabet soup” of certifications are all warning signs.
  4. They dodge verification questions. Real recruiters are proud of their client relationships. Fake ones ghost you when you ask for specifics.
  5. They ask you to pay for documents or preparation. No legitimate employer requires this. Ever. This is always the scam.
  6. Watch for the grammar shift. Polished emails that suddenly drop in quality when money enters the conversation? That’s AI-generated content with a human-written sales pitch sloppily stitched in.
  7. Check the metadata. If they send you an “official” document, look at every corner, every file property, every detail. Scammers are playing a numbers game, and as a result, they’re often rushed and sloppy. Sometimes they literally leave the watermark.
  8. The emotional setup is part of the scam. Flattery, validation, and the sense that someone finally sees your worth is intoxicating, especially when you’ve been job hunting for months. That’s by design. The best time to be skeptical is when you most want to believe.

Why this matters right now

This isn’t just my problem. It’s an epidemic:

AI tools are making these scams more polished, more personalized, and harder to detect. The “spray and pray” emails with obvious typos are being replaced by tailored, multi-email campaigns that build trust over weeks before making their move.

If you’re job searching (or know someone who is), please share this post and the video. The more people know what to look for, the less effective these scams become.

Watch the video

Once again, here’s the video, where I walk through the entire scam step by step, from the first email to the ChatGPT watermark:

And if you haven’t already, subscribe to the Global Nerdy YouTube channel. There’s more coming soon, and I promise it’ll be less infuriating than this one. Probably.

Report it

If this has happened to you, here’s where to report it:

And if you’ve got your own story about a fake recruiter, drop me a line on LinkedIn! Let’s make these scams harder to pull off.

Categories
Artificial Intelligence Humor

Remember to occasionally sass back at LLMs

The screenshot above is another regular reminder from Yours Truly that the LLM isn’t always right, but the final decision is always yours. Sometimes, you need to sass back — not necessarily to get better results, but to remind yourself not to abdicate completely to AI.

In case the first line in my prompt sounds familiar, but you can’t place it, here’s the source:

Here’s the video version:

https://www.youtube.com/shorts/NPXjPWIj31g

Categories
Artificial Intelligence Current Events Humor

Claude’s Super Bowl ads are so funny that Sam Altman’s crashing out over them

How upset is Sam Altman about Anthropic’s Super Bowl ads for Claude, which poke fun at ChatGPT and their inclusion of ads? Upset enough to call them “authoritarian,” in the same way a tween would call their parents “fascist” because they wouldn’t give them permission to go to a slumber party.

But daaaaamn, are they memorable and funny.

There are four such ads, each one featuring two actors, with one playing the part of the user, and the other playing the part of ChatGPT. The acting is perfect, with the user clearly in need of answers, and ChatGPT with slightly delayed responses delivered in a saccharine tone and a creepy smile at the end (“Give me your creepiest fake smile!” must’ve been part of the audition process). All the ads end with a snippet of the rap version of Blu Cantrell’s 2003 number, Breathe, which features Sean Paul and one of the best beats from that era.

I’ve posted the four ads below, from my least to most favorite. Each one features a common LLM use case.

Here’s Treachery, where a student is asking ChatGPT to evaluate her essay:

Deception features ChatGPT providing advice on the user’s business idea:

Violation’s user wants a six-pack — the muscle kind, not the beard kind — and is about to regret telling ChatGPT his height:

And my favorite, Betrayal, starts with the user trying to get closer to his mom, and ends on a cougar-riffic note:

OpenAI CEO and owner of the world’s most punchable voice Sam Altman is, as the kids say, crashing out over these ads, calling them “dishonest” (they’re more hyperbolic) and “authoritarian” (which is Altman himself being hyperbolic):

…and the most blunt headline of the bunch:

Categories
Artificial Intelligence Business Career Work

The key to thriving in the AI age is beating the bottlenecks

One of Nate B. Jones’ recent videos has the title Why the Smartest AI Bet Right Now Has Nothing to Do With AI (It’s Not What You Think). While the title is technically correct, I think it should be changed to In the Age of AI, You Have to Beat the Bottlenecks.

Bottleneck: a definition

Many Global Nerdy readers aren’t native English speakers, so here’s a definition of “bottleneck”:

A bottleneck is a specific point where a process slows down or stops because there is too much work and not enough capacity to handle it. It is the one thing that limits the speed of everything else.

Imagine a literal bottle of water.

  • The body of the bottle is wide and holds a lot of water.

  • The neck (the top part) is very narrow.

  • When you try to pour the water out quickly, it cannot all come out at once. It has to wait to pass through the narrow neck.

In business or technology, the “bottleneck” is that narrow neck. No matter how fast you work elsewhere, everything must wait for this one slow part.

Elon is often wrong, but you can learn from his wrongness

My personal rule is that when Elon Musk says something, and especially when it’s about AI, turn it at least 90 degrees. At the most recent World Economic Forum gathering in Davos, he talked a great “abundance” game, with sci-fi claims that AI would create unlimited economic expansion and plenitude for all:

Nate Jones watched the talk with Musk, but came to the conclusion that Musk’s take is the wrong frame for the immediate future. The current AI era will be one of bottlenecks, not abundance. I agree, as I’ve come to that conclusion about any grandiose statement that Musk makes; after all, he is Mr. “we’ll have colonies on Mars real soon now.

Here are my notes from Jones’ video…

Notes

Instead of abundance, Nate suggests that what we are entering is a “bottleneck economy.” While AI capability is growing, the actual value it produces won’t automatically flow everywhere and benefit everyone. Instead, it will concentrate around specific areas based on AI’s constraints and limitations [00:00].

Research from Cognizant claims AI could unlock $4.5 trillion in U.S. labor productivity (and yes, you need to take that figure with a huge grain of salt), and it comes with a massive caveat: businesses must implement AI effectively. Currently, there’s a wide gap between AI models and the hard work of integrating them into business workflows. This “value gap” means that the trillion-dollar impact won’t materialize until organizations figure out how to bridge the distance between models can do in general and what they can specifically do for a company’s operations [01:01].

Physical infrastructure is the first bottleneck. AI capability is increasingly constrained by things it needs from the physical world, specifically land, power, and skilled trade workers. Building the data centers required to train and run models takes years, and not just for the building process, but also permitting and connections to the power grid. This creates a wedge between the speed of software development and building infrastructure  [03:56].

Beyond just buildings and power, the hardware supply chain is the second bottleneck. Access to compute, high-bandwidth memory, and advanced chip manufacturing (controlled largely by TSMC) determines who gets a seat at the table. Companies that understand this are securing resources years in advance and treating regions with stable power and friendly permitting as strategic assets. This creates a market where value is captured by those navigating physical constraints in addition to building better algorithms [06:02].

The third bottleneck is one you might not have thought of: the cost of trust. As the cost of generating content collapses to near zero, the cost of trust is skyrocketing. Jone highlights what he calls a “trust deficit,” calling it a major coordination bottleneck. When any content can be fabricated, the ability to verify and authenticate information becomes expensive and crucial. Value will shift to institutions, platforms, or individuals who can mediate trust and provide a reliable signal in world rapidly filling with synthetic media slop [07:36].

For organizations, there’s the bottleneck of applying general AI to specific contexts. A general AI model won’t know a company’s private code base, board politics, or competitive dynamics. The bridge between “AI can do this” and “AI does this usefully here” requires tacit knowledge; that is, the practices and relationships that aren’t written down but live in the heads of the company’s employees. Companies that solve this integration problem will unlock productivity, while those that don’t will spend lots of money on tools they never use [09:55].

The fifth bottleneck is another one you might not have though of: the increasing value of taste. For individuals,  and especially for those in tech, the bottlenecks are shifting from acquiring skill to getting good at making judgment calls. AI is commoditizing hard skills like programming (it’s cutting down the time to proficiency from years to months), the really valuable skills are going to be taste and curation. The ability to distinguish between AI output that’s “good enough” versus AI output that’s extraordinary will become the differentiator. Developing taste takes experience, time, and observation. This is going to create a dangerous race for early-career professionals, whose entry-level work is being devalued [14:52].

The combination of problem-finding and execution are the sixth bottleneck. When problem-solving becomes automated, finding the problem and executing on the solution become the new moats. The market will reward those who can frame the right questions and navigate the ambiguity of implementing appropriate solutions. Jones emphasizes that while AI can generate a strategy or a plan, it can’t execute the “grinding work” of follow-through, holding people accountable, and navigating organizational politics. Success depends on identifying these new personal bottlenecks rather than optimizing for old skills that AI is turning into commodities [16:50].


Tips for techies and developers to beat the bottlenecks

  • Cultivate a sense for taste in addition to a skill for syntax. As coding moves from purely “grind” to at least partially “vibe” (see my vibe code vs. grind code post), your value shifts from writing code to reviewing AI-generated code. You need to refine your sense of what makes code good to differentiate yourself from the flood of AI output, which tends towards the average. [15:06]
  • Specialize! To beat the “good enough” standard of AI, pick a niche, and specialize in it. The window for being a generalist is closing, and extraordinary depth allows you to spot quality that AI (which once again, tends towards the average) misses. [16:16]
  • Pivot to problem finding. AI makes a lot of problem solving cheap, which makes problem finding the rare and precious thing. Stop defining yourself solely as a problem solver. Focus on defining the right problems to solve, framing the architecture, and determining direction. This management-level skill is harder for AI to replicate than execution. [16:50]
  • Value tacit knowledge and context. Tacit knowledge is the “soft” knowledge of how an organization works, and it’s almost never documented (at least directly), but lives in the heads of the people working there. Knowing why a legacy codebase exists or understanding specific stakeholder needs is a “context moat” that general AI models can’t easily infer. [17:36]
  • Focus on execution and follow-through. AI can generate the plan/code, but it can’t navigate the friction of deployment. The “grinding work” of implementation, such as convincing teams, fixing integration bugs, and finalizing products, is where the real value now lies. [18:47]
  • Build your tolerance for ambiguity. This has always been good for real life, but now it’s also good for tech work, which used to live in rigid, well-defined, unambiguous spaces… but not anymore! The tech landscape is shifting rapidly, and the ability to remain functional and productive while “metabolizing change” and dealing with uncertainty is a critical soft skill that separates leaders from people who freeze when things become ambiguous. [20:01]
  • Audit your personal bottlenecks: Be honest about what is actually constraining your career right now. It might not be learning a new framework (the old bottleneck). Instead, it might be your ability to integrate AI tools into your workflow or your ability to communicate complex ideas. Find those bottlenecks and come up with strategies to overcome them! [21:25]
Categories
Artificial Intelligence

A quick intro to OpenClaw (formerly MoltBot, formerly ClawdBot) and 7 tips for getting started

When I was asked about what AI tools I was trying out in my recent interview on the Enlightened Fractionals podcast, one of the tools I named was Clawdbot. But I was already out-of-date enough to have used the incorrect name, because it had been changed to Moltbot. Or maybe it had been re-renamed to its current name (at least at the time of writing), OpenClaw.

Clawdbot, Moltbot, OpenClaw: What is this thing?

OpenClaw is an open-source AI assistant that went from launch to viral sensation to full-on crisis management mode in just five days. It originally went by the name I used, Clawdbot, but then rebranded twice:

  1. From ClawdBot to Moltbot after Anthropic raised trademark concerns about the name’s similarity to Claude. Let’s face it, the name “ClawdBot” was a reference to Claude, and the misspelling was intentionally meant to prevent the kind of IP violation concern that they ended up running into. “Moltbot” is a reference to molting, which is when a lobster sheds its outer shell and emerges with a new, soft shell as its exoskeleton.
  2. From Moltbot to OpenClaw after creator Peter Steinberger simply decided he didn’t like the interim name.

Throughout the chaos, the project now know as OpenClaw has attracted over 144,000 GitHub stars, along with crypto scammers, handle-sniping bots, and a lot of cybersecurity practitioners’ attention.

What makes OpenClaw different?

  1. Unlike traditional AI chatbots that live on dedicated websites, OpenClaw integrates directly into to a number of messaging apps, and it’s pretty likely you already use at least one of them. You can interact with it using WhatsApp, Telegram, iMessage, Slack, Discord, or Signal. Using OpenClaw is like texting or messaging a friend, and it routes your messages to whichever LLM you choose while handling task automation locally.
  2. OpenClaw runs on a computer (real or virtual) that you control and gives the LLM access, allowing it to take action on your behalf.

The promise of a real AI assistant

OpenClaw offers three standout capabilities:

  1. Persistent memory: OpenClaw remembers from session to session and doesn’t forget everything when you close the app. It learns your preferences, tracks ongoing projects and actually remembers conversations you had and what you tell it.
  2. Proactive notifications: OpenClaw notifies you about important things, such as daily briefings, deadline reminders and email triage summaries. You can wake up to a text saying, “Here are your three priorities today,” without having to ask the AI first — it does so proactively.
  3. Real automation: Because you can grant OpenClaw read and write access to your local filesystem and browser access, it has been described as “an LLM with hands.” It can schedule tasks, read and re-organize your files, fill out forms, search  and reply to your email, generate reports, and control smart home devices. It’s been used for thinngs like achieving “inbox zero” to handling research threads that run for days, habit tracking, and providing automated weekly recaps of what they shipped.

Real talk: Should you try OpenClaw or wait?

At this point, I feel the need to remind you that Clawdbot/Moltbot/OpenClaw is an open source project moving at AI speed that’s been in use by early adopters for only a week. And it that time, the project has faced the threat of cancellation via trademark lawyers, and some of its user base have fallen prey to crypto scammers while others have failed to grasp its security implications and have exposed their private information to the ’net at large.

If you need something that “just works” and has something like a one-click install, I suggest waiting. The things OpenClaw does are too cool and convenient to be ignored. If the OpenClaw people don’t make a safer, simpler version, someone else most definitely will (and get rich in the process).

Serious security considerations

Just Google “security” and “openclaw” (or “clawdbot” or “openmolt”) and you’ll see articles written by all manner of security experts who’ve flagged significant risks with OpenClaw’s architecture. It runs on your local computer and can interact with emails, files, and credentials on that computer. If you configure it the wrong way, you can unintentionally expose private data such as API keys.

Researchers have already discovered numerous publicly accessible OpenClaw instances that have little or no authentication. OpenClaw also creates what one security analyst called a “hybrid identity” problem, where it operates as you, using your credentials after you’ve logged off. This kind of “digital twinning” was largely in the realm of science fiction until last week, and ,ost security systems aren’t designed to handle it.

The current OpenClaw situation (which is subject to change very, very quickly)

Despite the initial hiccups (and there will be more),  OpenClaw continues to grow. It’s got an active Discord community, it keep collecting GitHub stars, and the  team appears to have learned some lessons about viral success and security practices. Expect to see more posts and stories about it over the next few weeks.

7 tips for getting started with OpenClaw

  1. If you’re feeling confident about trying it out, go to openclaw.ai and review the documentation thoroughly. Before installing anything, read through the official guides to understand the architecture, requirements, and how the message routing to LLM providers works. This will help you make informed decisions about your setup.
  2. Complete the security checklist before deployment. This is new software in a new field where we learn new things every day. Given the documented vulnerabilities in early deployments, prioritize authentication configuration, ensure your instance isn’t publicly accessible, and never expose API keys. Consider using a dedicated machine or virtual environment rather than your primary computer. (I’m currently using a Raspberry Pi 500 for this purpose.)
  3. Beware of Mac Mini scams. Speaking of dedicated machines, the Mac Mini, thanks to its fast Apple Silicon processors and fantastic memory bandwidth, has become the preferred AI development machine and the preferred OpenClaw platform. Enterprising con artists have found out how in-demand Mac Minis are and have been posting scam ads on places like Facebook Marketplace. I’ll write an article about my own experiences with such scammers soon.
  4. Choose and configure your LLM backend. Decide if you want to use one of the bigger paid services like Claude, ChatGPT, or Gemini, and understand the associated costs before connecting them to OpenClaw (you might want to consider DeepSeek). You can also go with a local model, which is what I’m doing.
  5. Start with a single messaging integration. Don’t go nuts. Pick one messaging platform to use with OpenClaw to test the waters (I suggest Discord). This limits your exposure while you learn how OpenClaw behaves and what permissions it actually needs.
  6. Limit its destructive capability and start by giving OpenClaw only read-only automation. Start by letting OpenClaw summarize emails or provide briefings before  giving it “write” access to send messages, modify files, or execute commands on your behalf. Begin slowly and safely, then gradually expand its permissions as you become more certain about your security configuration and how OpenClaw behaves.
  7. As a reminder of the dangers of letting an AI agent run wild on your behalf, I strongly recommend you watch the Sorceror’s Apprentice part of the Walt Disney animated film Fantasia. In case you don’t have a Disney+ account, I’ve posted it in the YouTube embeds below:

Categories
Artificial Intelligence Process Programming

Projects I’m vibe coding, projects I’m grind coding, and projects in-between

A couple of weeks back, I wrote about how coding happens on a spectrum whose opposite ends are:

  • Vibe coding, a term coined by Andrej Karpathy, is where where developers use natural language prompts to have LLMs or LLM-based tools generate, debug, and iterate on code. Vibe coding is declarative, because you describe what you want.
  • Grind coding, my term for traditional programming, where you specify how a program performs its tasks using a programming language. Grind coding is imperative, because you specify how the thing you want works.

I myself have been writing code for different purposes, on different parts of this spectrum (see the diagram at the top of this article for where they land on the spectrum):

  • The Tampa Bay Tech Events utility: This is the Jupyter Notebook I use to gather event info from online listings and build the tables that make up the event listings I post every week here on Global Nerdy. I wrote the original code myself, but I’ve called on Claude to take the tedious stuff, including analyzing the obfuscated HTML in Meetup’s event pages to find the tags and classes containing event information.
  • MCP server for my current client: This is a project that started before I joined, and was written using a code generation tool. The client is a big platform connected to some big organizations; my job is to be the human programmer in the loop.
  • Picdump poster: Every week, I post “picdump” articles on the Global Nerdy and Accordion Guy blogs. Over the week, I save interesting or relevant images to specific folders, and the picdump poster utlity builds a blog post using those images. It’s a low-effort way for me to assemble some of my most-read blog posts, and it’s more vibe-coded than not, especially since I don’t specialize in building WordPress integrations.
  • Copy as Markdown: Here’s an example of using vibe coding as a way to have custom software built on demand. I wanted a way to copy text from a web page, and then converting that copied text into Markdown format. This one was purely vibe-coded; I simply told Gemini what I wanted, and it not only generated the code for me, but also gave me instructions on how to install it.