Subscriptions

Ctrl + K

Sign inCreate account

Home

AllListenPaidSavedHistory

Sort byPriorityRecent

Generative Value

Generative Value

10:05 AM

The AI & Healthcare Opportunity

Part Two of How AI Can Help Improve the US Healthcare System

Eric Flaningam∙11 min read

💎DiamantAI

💎DiamantAI

9:38 AM

Why AI Experts Are Moving from Prompt Engineering to Context Engineering

How giving AI the right information transforms generic responses into genuinely helpful answers

Nir Diamant∙8 min read

Platforms, AI, and the Economics of BigTech

Platforms, AI, and the Economics of BigTech

8:30 AM

How to intellectually debate AI while completely missing the point

Why both sides of the argument on AI miss the point

Sangeet Paul Choudary∙7 min read

AI CFO Office

AI CFO Office

8:30 AM

How to use AI to get out of Excel Hell

I spoke to a CFO of a $300 million company who said, Every financial metric is bullshit.

AI CFO Office∙4 min read

The Leverage

The Leverage

8:02 AM

OpenAI Released an Agent, XAI Released an e-girl

The Weekend Leverage, July 20.

Evan Armstrong∙8 min read

SEMI VISION

SEMI VISION

7:15 AM

Important Notice: Beware of Unauthorized Sales of SemiVision Reports

By SemiVision Research

SEMI VISION∙1 min read

AI Supremacy

AI Supremacy

6:57 AM

Deep Dive on Thinky’s Seed Round (Part II)

This is a Multimodal Sovereign AI startup. The first of its kind.

Michael Spencer∙24 min read

Derek Thompson

Derek Thompson

6:00 AM

The Sunday Morning Post: ‘Science Is the Belief in the Ignorance of Experts’

The people who built the technology we rely on were often crazy, wrong, or both.

Derek Thompson∙7 min read

The Intellectual Edge

The Intellectual Edge

2:00 AM

The Case For Multi-disciplinary Thinking

The best Intellectual Edge we can have.

The Intellectual Edge∙11 min read

AI & Tech Insights

AI & Tech Insights

12:08 AM

Human Grit vs. Smart Code: Psyho Wins, AI Is Right Behind

🚩 What Happened

AI & Tech Insights∙3 min read

How to AI

How to AI

Jul 19

16 chatgpt myths.

No, ChatGPT does not ‘read’ the entire internet.

Ruben Hassid∙6 min read

Product Growth

Product Growth

Jul 19

How Linear Built a $1.25B Unicorn with Just 2 PMs

Linear’s Head of Product Nan Yu goes deep on the Linear Method

Aakash Gupta∙ 1 hr listen

The Founders Corner®

The Founders Corner®

Jul 19

Jensen Huang on AI & Jobs🤖, 12 VCs Took Half the Money in H1🏦, Murati’s $12B AI Lab Has No Product Yet🧬

If you’re building, investing, or just trying to stay ahead of the curve, you’re in the right place.

Ruben Dominguez Ibar and Chris Tottman∙4 min read

Sabrina Ramonov 🍄

Sabrina Ramonov 🍄

Jul 19

Best AI Prompt to Humanize AI Writing

Free AI prompt you can use with any LLM to help reduce telltale signs of AI text.

Sabrina Ramonov 🍄∙3 min read

Investing 101

Investing 101

Jul 19

Seek Sunlight

Here We Are Again At The Twilight

Kyle Harrison∙4 min read

ByteByteGo Newsletter

ByteByteGo Newsletter

Jul 19

EP172: Top 5 common ways to improve API performance

What other ways do you use to improve API performance?

ByteByteGo∙6 min read

Market Sentiment

Market Sentiment

Jul 19

Betting on GLP-1

Buying the haystack

Market Sentiment and Rebound Capital∙10 min read

Nate’s Substack

Nate’s Substack

Jul 19

Will AI Really Doom Us? 3 Hard Facts That Say ‘Don’t Panic’

A data-driven, plain-English letter to friends who lie awake fearing an AI apocalypse—why today’s systems fall short of doomsday hype and where the real risks really lie

Nate∙ 14 min watch

Mostly metrics

Mostly metrics

Jul 19

The CEO’s #1 Job: Capital Allocation

Today we’re talking about capital allocation, and how it’s one of the most powerful, yet least discussed, skills a CEO needs to learn.

CJ Gustafson∙ 14 min watch

Decoding ML

Decoding ML

Jul 19

I read Claude’s prompt. Here are 5 tips to master prompt engineering.

The simplest way to find the right LLMs. Copying code from ChatGPT won’t make you an AI Engineer.

Paul Iusztin∙5 min read

Get app

Joshua Saxe

Subscribe

Joshua Saxe

AI security notes 6/27

AI generated code security risks stem from basic LLM limitations; we should frame prompt injection models like other detection and response tools

Joshua Saxe's avatar

Joshua Saxe

Jun 27, 2025


Sources of vibe coding risk

We know AI coding agents can write blatantly insecure code or cause slopsquatting risks. But a deeper and more enduring issue will be AI’s lack of access to tribal, implicit organizational cultural knowledge, often passed down by oral tradition, or in fragmented, poor documentation, required to understand what ‘secure code’ looks like in specific organizational contexts.

Another enduring source of risk: when user requests are underspecified, AI ‘fills in the blanks’ in ways that may be misaligned with the engineer’s intentions:

Of course, even when full context is available, AI coding systems also have poorer code security understanding than human engineering teams

As an aside, these are all reasons that AI won’t replace human engineers in the foreseeable future and will remain a ‘power tool’ requiring significant hand holding and human review.

Gaps in achieving superhuman performance at secure coding

Much reduces to the alignment problem in AI security, including problems in secure AI coding

Opportunities in vuln discovery and remediation

Lots of new and interesting work targets vulnerability discovery and remediation. Notably, most existing research targets finding and fixing generic bugs that don’t depend on specific organizational context. A couple papers I found interesting recently:

Prompt‑injection guardrails: instrumentation, not silver bullets

Prompt injection detection models can be easily bypassed by AI redteamers. The headlines tend to suggest this means the these models are “broken,” but that framing misses how these tools are meant to be used.

Prompt injection models are for AI detection and response

Google’s GenAI Security Team blogged Mitigating Prompt Injection with a Layered Defense Strategy detailing how they situate prompt injection defense within their larger strategy.

PI‑Bench is an interactive benchmark that mutates prompts in real time; baseline detectors drop from F1 0.82 on familiar attacks to 0.41 on unseen variants. The study quantifies how brittle today’s guardrails are when adversaries shift tactics—a direct validation of why SOC‑style iterative training is needed.

SoK: Evaluating Jailbreak Guardrails for LLMs surveyed 40 guardrail papers, proposed a six‑dimension taxonomy, and showed that mixing statistical and rule‑based defenses yields the best risk–utility trade‑off.


Subscribe to Joshua Saxe

Launched 3 months ago

Machine learning, cyber security, social science, philosophy, classical/jazz piano. Currently at Meta working at the intersection of Llama and cybersecurity

Subscribe

By subscribing, I agree to Substack’s Terms of Use, and acknowledge its Information Collection Notice and Privacy Policy.

Travis Truman's avatar

Mohamed Kifayathulla's avatar

Prashant Kulkarni's avatar

4 Likes∙

1 Restack

4

2