NHacker Next
login
▲OpenAI models coming to Amazon Bedrock: Interview with OpenAI and AWS CEOsstratechery.com
292 points by translocator 18 hours ago | 97 comments
Loading comments...
spindump8930 17 hours ago [-]
Remember that models on different inference platforms might not necessarily give exactly the same results, adding another axis of non-determinism to development. Things like quantization, custom model serving silicon, batching, or other inference optimizations might mean a model from the original provider performs differently from the hosted one :/

This paper isn't the exact same scenario, since it's an auditable open weight llama model, but shows the symptoms of this: https://arxiv.org/pdf/2410.20247

bossyTeacher 17 hours ago [-]
Anyone who has used gpt-x via openai vs microsoft has experienced this very clearly.
energy123 7 hours ago [-]
Which one is better?
dannyw 5 hours ago [-]
For OpenAI, OpenAI direct has always been better; except maybe early 2023-era when OpenAI Platform was not that stable or reliable yet.

For Anthropic, it can vary based on model and time. For Opus 4.7, Bedrock is the clear winner in TPS by leaps: https://artificialanalysis.ai/models/claude-opus-4-7/provide...

spindump8930 13 minutes ago [-]
That artificial analysis page has some great references for this, thanks for sharing.
weli 6 hours ago [-]
As a rule of thumb inference offered by the model labs are closer to the "true implementation" compared to third parties. They have other problems though.
redsocksfan45 4 hours ago [-]
[dead]
zmmmmm 16 hours ago [-]
Availability through Bedrock has been a major driver in use of Anthropic in my org. And I am betting there is actual margin in it as well.

I wonder if this is directly linked to the split up with Microsoft. Just from my anecdata, OpenAI is getting completely ignored in serious enterprise deployments because what they offer on Azure sucks and there is no other corporate friendly way to get it. They probably saw themselves getting destroyed in enterprise and realised it was existential to be able to compete with Anthropic on AWS.

hobofan 4 hours ago [-]
May be very dependent on country/region, but in Germany where Azure is quite prevalent in mid-market companies (due to heavy historic reliance on Microsoft in general), OpenAI + Azure seems to have worked as a great synergy. Few customers I've worked are even trying to reach for other providers, as the OpenAI models are available for them and promoted by Azure.
skybrian 16 hours ago [-]
It seems pretty clear that OpenAI renegotiated their agreement in preparation for this:

https://news.ycombinator.com/item?id=47921248

16 hours ago [-]
NikolaosC 5 hours ago [-]
ur probably right on the margin. Anthropic doesn't break it out, but enterprise spend on Bedrock is the highest quality revenue in the AI stack right now. Itzs sticky, multi-year, embedded in existing AWS commits. OpenAI was watching that compound while stuck on Azure
9 hours ago [-]
brokencode 14 hours ago [-]
Just curious, what’s wrong with Azure?
zmmmmm 14 hours ago [-]
It might be just our region, but for a long time we couldn't access current frontier models at all. Only old GPT4 level models. Meanwhile, Anthropic is rolling out access to every model within 24 hours of announcement to Bedrock.
7thpower 12 hours ago [-]
Not sure how much Azure OAI has changed, but when I last used it 2-3 years ago, it was just a farce to get you using provisioned throughput. The throughput quotas were small, the process to request more was bureaucratic, and the Azure SAs were

It was also very clear the OAI and MS teams held each other in contempt (not relevant, but was interesting and grew in the immediate aftermath of the Altman drama).

So why were we using it? OpenAI don’t really have an enterprise go to market, bedrock still relied on Claude 2, and we weren’t willing to YOLO on clickthroughs.

Once Claude 3 came out, we jumped ship. That sucked too, although I hear it’s gotten better though.

Foobar8568 9 hours ago [-]
I have been out of openai azure deployments for a whole,1year+, but we had often spikes in latency, escalated to the Head of Azure Europe, and still no official clues about them, meanwhile they were trying to get us in some kind of collaboration announcements. And it was the only reasons we had a few meetings with that guy.

So yeah Azure sucked ass and plenty of outages or latency, like 3min for first byte while usually it was max 30sec to 1min, if not even faster (memory is a bit fussy)

jojobas 9 hours ago [-]
Everything. https://isolveproblems.substack.com/p/how-microsoft-vaporize...
jasobake 18 hours ago [-]
As someone who works at big tech and spends countless hours in meetings hoping to get some small feature coordinated for deployment across two teams, I can't imagine the amount of meetings and 6-pagers that were involved in running these models on bedrock's hardware.
33MHz-i486 17 hours ago [-]
at this level they just decide and spin up a swat team to execute it in a couple weeks without politicking. the bureaucratic ways, reviews are just for the low levels, to keep them busy with feature scraps while they mostly do operations
aab99 14 hours ago [-]
yup, I think there are few public articles on aws mantle so you can look it up, but internally this is pretty common knowledge. The entire inference engine of bedrock is built and maintained by a handful of ec2 engineers (all principals and above). Judging by the commit history of the project they are able to just build independent of any of the traditional bureaucracy.
avereveard 59 minutes ago [-]
For better and worse tho.
33MHz-i486 13 hours ago [-]
the way in which Mantle was built is highlighted internally by PEs as some sort of triumph but really its a fairly tone deaf indicted of AWS’s engineering culture ... “To achieve a meaningful result in a reasonable amount of time we had to break nearly ever constraint that we force all other engineers to work under. good luck to you plebs of L6 and below”
darkwater 5 hours ago [-]
I'm going to play devil's advocate here: _what if_ most of lower level engineers are actually not able to self-organize themselves like this dedicated, I bet hand-picked group of PE can? I'm pretty sure AWS ruthless culture would gladly use less middle-managers and be swifter in time to market if it were so easy, no? What works for a single, highly-focused project (or a handful of similar situations) doesn't scale when you have to take care of bazillion customers, do boring/smaller tasks and keep the machine ticking.
jjmarr 1 hours ago [-]
[deleted]
3748595995 54 minutes ago [-]
can I suck your d ick?
spelunker 12 hours ago [-]
Check out this all of this stuff we can build with a room full of PEs and no rules!
o10449366 17 hours ago [-]
Lol, spinning up swat teams because someone high up decides "drop everything this is my pet priority now" is politicking. It looks good for the leaders, meanwhile its the engineers pulling the all nighters and dealing with having to maintain systems that are operationally compromised from day 0 because there's no proper planning/scoping involved other than "Big Man says this needs to be done in 2 weeks"
tt24 12 hours ago [-]
Okay but I’ll be able to use OpenAI models on bedrock now

Why would I care if AWS asks their engineers to work a little harder on a project

phillipcarter 16 hours ago [-]
Sure, but also...

...anyone with a brain at AWS knows that supporting OpenAI's latest models on Bedrock is simply good for AWS. That context is rather important!

ignoramous 11 hours ago [-]
> engineers pulling the all nighters

There's always some carrot with the stick, even if an imaginary one!

giancarlostoro 17 hours ago [-]
Depends on how its implemented, but Amazon already did add gpt-oss-20b so if the model is similar enough to the OSS variant of GPT, it might not have been as complicated as you might think.
londons_explore 17 hours ago [-]
I imagine there's lots of custom kernels and optimization...

Openai hasn't been publishing innovations for quite a while.

NamlchakKhandro 16 hours ago [-]
Neither has anthropic.

They're both just stealing ideas from pimono extensions

ignoramous 11 hours ago [-]
Some say they have been stealing ideas from Chaebols [0] on circular deals [1], government reciprocation, & monopolisation.

[0] https://business.columbia.edu/sites/default/files-efs/imce-u...

[1] https://www.reuters.com/business/autos-transportation/compan...

vicchenai 15 hours ago [-]
The enterprise sales motion here is interesting. A lot of regulated industries (finance, healthcare) have existing AWS contracts with data residency commitments baked in. OpenAI on Bedrock basically lets those orgs skip the separate DPA negotiation with OpenAI. Could be a bigger unlock than it looks on paper.
epistasis 17 hours ago [-]
Claude got a looooot more buy in with a lot of privacy-concerned orgs I work with because they could access it through their "trusted" intermediate Amazon. OpenAI has been banned and is not trusted. I'm not sure that I agree with these orgs' legal teams' assessments, but they definitely read the terms of service far closer than I did.

We will see if this changes the equation, but it feels like OpenAI is pretty far behind and playing catch up on all fronts. Though to be honest, "pretty far behind" is like 2-8 weeks in the AI world, so it may not matter a ton, it's mostly perception. And for me and my information bubble, perception of OpenAI is rock-bottom due to Sam Altman. From appearing unethical to appearing unhinged with demands from fabs and everything else, I'm not a fan.

fny 15 hours ago [-]
You can sign ZDR agreements with any of the major LLM providers. Using AWS alone is also not sufficient. Even though AWS is running the model, you need to contact them for proper ZDR.[0]

[0]: https://platform.claude.com/docs/en/build-with-claude/claude...

PretzelJudge 14 hours ago [-]
Helpful link. Thank you.

I think that when people are worried about ZDR, what they really worry about is data governance. From what I’ve seen there’s a general distrust of OpenAI. AWS may keep your data around (without formal ZDR) but the concern of governance (using your data to train without your consent) seems like it would be much lower, because any breach of contract at AWS would have potential to destroy trust in what’s already a massively profitable company, so the incentives just aren’t there.

I’m not claiming OpenAI is training on API data. Just that they don’t have as strong of an incentive not to as AWS.

donavanm 13 hours ago [-]
AWS took limited data retention very seriously starting around 2015. Before that it was reasonable controls and a strong culture preserving customer privacy. After 2015ish they started implementing strong controls, to where service team members cant feasibly access customer data in the service they run, and account termination starts a legit data removal process (“GDPR compliance”). They also take the terms of service and user agreement (“your data” etc) very seriously in general.
bg24 14 hours ago [-]
While Anthopic has the best model and a focussed (no disturbance, lawsuits) leadership, they got a lot of enterprise access due to AWS. It is mutual no doubt, with both sides benefitting. The culture of feedback loop of AWS customers would have helped them in getting to enterprise-ready faster. Just my hypothesis.
stingraycharles 11 hours ago [-]
But Azure is just as big, if not equally big, in Enterprise. The argument that Azure didn’t give them enough access within enterprises doesn’t make sense to me.
dannyw 5 hours ago [-]
Microsoft Azure and OpenAI have been basically hating each other since the beginning, since their incentives were completely misaligned.
consumer451 15 hours ago [-]
Legally, SLA, and data concern-wise, is this any better than OpenAI on Azure? That has been around for a while.
PretzelJudge 14 hours ago [-]
By default you can’t access the latest OpenAI models unless you request access. We requested access for a very straightforward use case and never got it. We switched to Anthropic and Bedrock for that reason.
13 hours ago [-]
UqWBcuFx6NV4r 14 hours ago [-]
You don’t have to use Azure.
outside1234 17 hours ago [-]
The thing they are really wildly behind on is a business model. They are losing wild amounts of money per customer and it is hard to see how the competitive situation is going to allow them to fix that.
echelon 17 hours ago [-]
Given the scaling hurdles Claude Code / Opus is having, those Anthropic customers might leave to Codex. I'm _this_ close.
jwilliams 17 hours ago [-]
Codex is pretty good. Its friction to switch but I think it’s sensible being across multiple AI toolchains.
try-working 14 hours ago [-]
No friction in switching coding models.
NamlchakKhandro 16 hours ago [-]
Pi mono.

Nuff said

unrelat3d 16 hours ago [-]
This is what I use now after testing others through 2025

It has the most "UNIX" feel of a simple app that you compose the just right flow from and nothing more

felixgallo 15 hours ago [-]
Thing is, if you're using Codex, you're supporting Sam Altman and the idea of Sam Altmans, in the same way that if you use X or buy a Tesla, you're supporting Elon Musk and the idea of Elon Musks. That's a pretty big tax to factor into the usage of such products. If you even got 5% better coding results, would that make up for the future they're trying to build?
sheeshkebab 12 hours ago [-]
The more Dario talks the less I want to have anything to do with his wares.
xienze 15 hours ago [-]
Dario wants to replace you with AI as well. Don't be fooled into thinking he's your friend because he said no to Trump that one time. I'll remind you that Musk used to be the left's hero not too long ago.
felixgallo 12 hours ago [-]
I'm in the "AI could be good for humanity" camp, and in this camp, we believe that Dario/Anthropic is a radically better choice going forward than the alternatives at this moment. In this camp we are not 'fooled into thinking he's our friend because he said no to Trump that one time', we are evaluating the entire set of available information and figuring that Anthropic's the best bet.

As for Musk ever being "the left"'s "hero" -- that's amazing, that's what Pauli would call 'not even wrong'.

oefrha 3 hours ago [-]
Funny of you to bring up "humanity" while singing praise of the guy who's on record A-OK with aiding mass surveillance of anywhere not the U.S., and who's happy to help kill people but just wouldn't do it fully automatically, merely because he doesn't believe the tech is there yet. All while collaborating with Palantir.

If by "the best bet" you mean slightly less shitty bet then maybe.

felixgallo 2 hours ago [-]
That’s approximately what I mean. Although in comparison to the alternatives of Altman and Musk, the distinction is stark and significant.
oefrha 2 hours ago [-]
For 96% of humanity it’s same difference.
epistasis 17 hours ago [-]
I'm getting pretty close too, but I wouldn't switch to Codex I'd switch to one of the open agents that can use any backing LLM. My reasoning is that if I'm willing to pay the cost of the small changes in usage, I might as well switch to an open source agent that I can add my own convenience features to, like remote sessions and phone-based operation.
jfkimmes 17 hours ago [-]
Codex is open source and allows any model to be configured.
epistasis 17 hours ago [-]
Many thanks for that info!
bossyTeacher 17 hours ago [-]
Why Codex when you can use something that hasn't been touched by Sam Altman? Surely, your drive to get the very best model isn't stronger than your sense of ethics?
NamlchakKhandro 16 hours ago [-]
Codex is not open source. And it's not even that extensible
milkshakes 16 hours ago [-]
https://github.com/openai/codex
16 hours ago [-]
ribosometronome 17 hours ago [-]
What would be subscription customers, no? Rather than Bedrock or per-api customers? Many of the companies running on Bedrock or by-use have per day limits above the max monthly subscription costs.
johnbarron 16 hours ago [-]
It just not about AWS being some "trusted intermediary"... it's that the model runs inside the customer own AWS account under a different contract. AWS explicitly states inputs/outputs are not shared with model providers and are not used to train base models [1]

And for OpenAI, there is a May 2025 preservation order in NYT v. OpenAI. The court is forcing OpenAI to retain ChatGPT output logs indefinitely, including chats users have deleted that would normally be purged within 30 days [2]. That makes it a non starter for HIPAA/GDPR bound orgs.

[1] https://aws.amazon.com/bedrock/faqs/

[2] https://openai.com/index/response-to-nyt-data-demands/

hn_throwaway_99 15 hours ago [-]
I'm confused, your own #2 link says that Open AI is not bound to store output logs indefinitely going forward:

> Update on October 22, 2025:

> After months of litigation, we are no longer under a legal order to retain consumer ChatGPT and API content indefinitely. Our obligations under the earlier order ended on September 26, 2025.

> We’ve returned to our standard data retention practices :

> Deleted ChatGPT conversations and Temporary Chats will be automatically deleted from our systems within 30 days (opens in a new window).

> API data will also be automatically deleted after 30 days.

TZubiri 11 hours ago [-]
It's like Coca-Cola being banned at a school, and then Pepsi getting some contracts with the cafeteria because of it.
giancarlostoro 17 hours ago [-]
They're also not focused exclusively only on building an LLM, they have video and image generation too. Anthropic has one single focus, and this is why they are usually at the very top in the SWE benchmarks.
phillipcarter 17 hours ago [-]
Isn't it the case that OpenAI and Anthropic regularly just swap for whoever is at the top of the latest benchmarks? They're also so close in scores that it's effectively a wash anyways.

What OP is referring to is Anthropic aligning with corporate terms and conditions early, positioning themselves to be effectively resold by AWS rather than requiring orgs to procure them directly. This is huge in the enterprise world because the processes to get broad approval are generally far smaller and shorter for "just another AWS service" compared to a whole new vendor.

djtriptych 15 hours ago [-]
OpenAI did teh same thing with Microsoft/Azure though.
Grimblewald 15 hours ago [-]
Isn't it an open secret that benchmarks are largly irrelevant at this point? Why else we do all have a personalized test battery for new models? That said i've stopped testing chatgpt entierly. Its still ok but is beaten by local models and it gets thrashed by non oai frontier providers. I get the history, but holding up oai outputs as equivallent is lile comparing yahoo to google post yahoo's collapse in search domains.

Oai language models are largly irrelevant at this point imo.

epistasis 17 hours ago [-]
IMHO the benchmarks aren't useful, and ranking among the frontier models is mostly noise. The extra features around the coding agent have a much bigger impact on productivity than having to provide slightly more specification and guidance to the models; a 90% success rate versus a 92% success rate on the tasks I ask it to do is far more influenced by what I say than what the model is capable of.
DrewADesign 15 hours ago [-]
Didn’t they say Sora will only be used to internally create training data? Integrated image generation seems more in the neat feature category than some fundamental advantage, but maybe someone has use cases I haven’t considered.
hn_throwaway_99 16 hours ago [-]
Open AI is killing Sora though, so it looks like they are looking at Anthropic's playbook of focusing on enterprise use cases and seeing that it's more profitable.
dannyw 5 hours ago [-]
But then they released gpt-image-2, which is clearly SOTA.
nijave 17 hours ago [-]
This would be a nice compliance win. One less sub-processor and all our data is already on AWS so less worrying about sending it off somewhere else
KaiserPro 16 hours ago [-]
Great, I can now buy openAI through AWS with an interface that is totally incompatible with all my tools (unless AWS have finally given up and just made bedrock useful by adopting openAPI finally)
ManishR 14 hours ago [-]
Product Manager for Bedrock Mantle here. We now have bedrock-mantle endpoint which supports OpenAI API compatibility with Responses and Chat Completions support. See https://docs.aws.amazon.com/bedrock/latest/userguide/endpoin...
KaiserPro 7 hours ago [-]
Oh snap, excellent!
Silfen 12 hours ago [-]
https://docs.aws.amazon.com/bedrock/latest/userguide/bedrock...
NikolaosC 5 hours ago [-]
OpenAI just gave up Azure exclusivity, killed the AGI clause, and stopped paying Microsoft revenue share to get on AWS. Anthropic figured out 18 m ago that enterprises buy from their cloud, not from the best model. OpenAI is just catching up.
quibono 5 hours ago [-]
Just waiting for Gemma 4, DeepSeek 4 now. Then the only thing I'll be able to complain about is the completely different API to interface with (unless they FINALLY move to full OpenAI support).
dear_prudence 3 hours ago [-]
Really useful for us as we heavily rely on AWS credits for our AI usage.
2001zhaozhao 17 hours ago [-]
The market might be increasingly hard on AI startups in general as enterprises adopt providers like Amazon Bedrock and refuse to sign other deals.
lwarfield 18 hours ago [-]
Well that didn't take long.
avaer 16 hours ago [-]
It probably did, but the PR stream that the public sees is a well oiled machine.

This HN post itself has 4 simultaneous announcement links; not a coincidence.

There are billions of investor money on the line if the wrong thing is said at the wrong time, it needs to be carefully crafted and staged.

brazukadev 15 hours ago [-]
I guess OpenAI's CFO didn't get the memo[0].

0. https://www.cnbc.com/video/2026/04/28/tech-shares-fall-after...

throw03172019 18 hours ago [-]
OpenAI frontier models coming to Bedrock soon?
karmasimida 18 hours ago [-]
> Starting today, @awscloud and OpenAI are bringing the latest OpenAI models to Amazon Bedrock, launching Codex on Amazon Bedrock, and launching Amazon Bedrock Managed Agents, powered by OpenAI (all in limited preview). AWS and OpenAI will continue to bring the latest advances to Amazon Bedrock—so the models and agents you build with today continue to benefit from new breakthroughs as they arrive.

https://x.com/amazon/status/2049178618639839427

dang 18 hours ago [-]
We've updated the title above to make that clearer.

Since the product doesn't seem to be available yet, and the other links are all press releases, we'll leave the interview up as the main link.

ihsw 17 hours ago [-]
[dead]
chopete3 8 hours ago [-]
This is a big news for AWS hosted products.

Microsoft Azure has been the worst interms of maintaining a highly available service and also managing predictable latency.

Their azure customer support is bad. Not ready for any real enterprise cloud offering. They behave like Comcast customer support.

It was absolutely idiotic to lock it down to Azure. It wasn't meant to be an iphone+at&t combo where the phone is an end all be all.

A cloud product depends on a lot of services and nobody would switch cloud providers for a candy.

echelon 17 hours ago [-]
This doesn't mean you have the raw model weights, right? That's still entirely hidden / opaque?

You can just run "air gapped" inference?

Is this only of interest to enterprise customers already on AWS (who want "air gapped" behavior)? Is there any other use case for this?

This will be more expensive than calling OpenAI directly, right?

kube-system 16 hours ago [-]
A lot of companies already have data processing agreements and compliance sign-off for using AWS. Many are hesitant to send their data to AI startups with an incentive to train their models and a history of being.... loose with how they intake training data. Even when they do give assurances otherwise. AWS is more trusted in this aspect.

If this ends up similar to Claude on Bedrock, it's the same price.

londons_explore 17 hours ago [-]
This is for people who don't trust openAI with their data, but do trust Amazon.

But it also is for Devs in a company who already have a blanket agreement with Amazon, but would have an uphill battle signing an agreement with openAI.

mochow13 13 hours ago [-]
OpenAI is tailgating Anthropic apparently.
Daffrin 3 hours ago [-]
[flagged]
tokenhub_dev 6 hours ago [-]
[flagged]
shevy-java 16 hours ago [-]
Now they are ruining amazon too. It's fascinating to see.

AI is kind of like the ultimate corporation drug. They are all on it. And can't get rid of it - ever again.

unreal6 13 hours ago [-]
How are they "ruining" Amazon with these launches? That is not clear to me.
redsocksfan45 4 hours ago [-]
[dead]
try-working 14 hours ago [-]
OpenAI marching towards its future as a dumb pipe.