53 comments

  • wg0 21 minutes ago
    A wise man from Google said in an internal memo to the tune of: "We do not have any moat neither does anyone else."

    Deepseek v4 is good enough, really really good given the price it is offered at.

    PS: Just to be clear - even the most expensive AI models are unreliable, would make stupid mistakes and their code output MUST be reviewed carefully so Deepseek v4 is not any different either, it too is just a random token generator based on token frequency distributions with no real thought process like all other models such as Claude Opus etc.

    • didip 18 minutes ago
      I agree. Data and userbase are still the moats.

      Once a new model or a technique is invented, it’s just a matter of time until it becomes a free importable library.

    • dominotw 10 minutes ago
      dont they have the moat of being able to test their models on billions of ppl and gather feedback.
    • kevin_thibedeau 15 minutes ago
      Can Deepseek answer probing questions about Winnie the Pooh?
  • gurjeet 3 minutes ago
    Related: GitHub has paused new signups for Copilot.

    > Starting April 20, 2026, new sign-ups for Copilot Pro, Copilot Pro+, and student plans are temporarily paused.

    From: https://docs.github.com/en/copilot/concepts/billing/billing-...

  • thanhhaimai 2 hours ago
    Opinions are my own.

    I think the biggest winner of this might be Google. Virtually all the frontier AI labs use TPU. The only one that doesn't use TPU is OpenAI due to the exclusive deal with Microsoft. Given the newly launched Gen 8 TPU this month, it's likely OpenAI will contemplate using TPU too.

    • bastawhiz 50 minutes ago
      Many labs use TPUs, but not exclusively. Most labs need more compute than they can get, and if there's TPU capacity, they'll adapt their systems to be able to run partially on TPUs.
    • maxclark 2 hours ago
      And almost by happenstance Apple. Turns out they have a great platform for inference and torched almost nothing comparatively on Siri. The Apple/Gemini deal is interesting, Google continues to demonstrate their willingness to degrade their experience on Apple to try and force people to switch.
      • GorbachevyChase 51 minutes ago
        They also degrade their own direct services with little warning or thought put into change management, so, to be fair, Apple may be getting the same quality of service as the rest of us.
      • bigyabai 1 hour ago
        Apple is basically in the same boat as AMD and Intel. They have a weak, raster-focused GPU architecture that doesn't scale to 100B+ inference workloads and especially struggles with large context prefill. TPUs smoke them on inference, and Nvidia hardware is far-and-away more efficient for training.
        • brcmthrowaway 51 minutes ago
          This doesn't get talked about enough - the GPU is weak, weak, weak. And anyone who can fix them will go to a serious AI company (for 2-3x the salary).
          • jorvi 44 minutes ago
            The GPU is monstrously good. Depending on the workload, the M1 series GPU using 120W could beat an RTX 3090 using 420W.

            Same with the CPU. Linux compiled faster on an M1 than on the fastest Intel i9 at the time, again using only 25% of the power budget.

            And the M-series has only gotten better.

            It is kind of sad Apple neglects helping developers optimize games for the M-series because iDevices and MacBooks could be the mobile gaming devices.

            • ethbr1 31 minutes ago
              Apples and limes.

              The context of this thread isn't consumer chips, but Apple's analog to an H/B200.

            • bigyabai 22 minutes ago
              The GPUs are bottom-barrel for compute-focused industries. It is mobile-grade hardware that arguably can't even scale to prior Mac Pro workloads.

              > The GPU is monstrously good. Depending on the workload, the M1 series GPU using 120W could beat an RTX 3090 using 420W.

              You're just listing the TDP max of both chips. If you limit a 3090 to 120W then it would still run laps around an M1 Max in several workloads despite being an 8nm GPU versus a 5nm one.

              > It is kind of sad Apple neglects helping developers optimize games for the M-series

              Apple directly advocated for ports like Death Stranding, Cyberpunk 2077 and Resident Evil internally. Advocacy and optimization are not the issue, Apple's obsession over reinventing the wheel with Metal is what puts the Steam Deck ahead.

              Edit (response to matthewmacleod):

              > Bold of them to reinvent something that hadn't been invented yet.

              Vulkan was not the first open graphics API, as most Mac developers will happily inform you.

              • matthewmacleod 20 minutes ago
                Apple's obsession over reinventing the wheel with Metal

                Bold of them to reinvent something that hadn't been invented yet.

    • philippta 47 minutes ago
      In the recent Dwarkesh Podcast episode Jensen Huang (Nvidia) said that virtually nobody but Anthropic uses TPUs. How does that add up?
      • csunoser 38 minutes ago
        I am not sure what context Jensen said that. But midjourney uses tpu. Apple uses tpu. They are no other frontier labs that use it, but Google + Anthropic is 2 out of 3 frontier lab so.....

        You could reasonably say that "A majority of frontier labs uses TPU to train and serve their model."

      • sarchertech 42 minutes ago
        Who is the other frontier lab other than Anthropic, OpenAI, and Google? I thought they were ahead of everyone else.
        • DeathArrow 23 minutes ago
          Folks who make Deepseek, Qwen, GLM, MiniMax, Kimi and MiMo.
    • VirusNewbie 2 hours ago
      OpenAI uses GCP. I don't know if they use TPUs.

      https://www.reuters.com/business/retail-consumer/openai-taps...

    • PKop 48 minutes ago
      > Opinions are my own.

      Why does this need to be stated? Who else's would they be?

      edit: he puts this on so many comments lol c'mon this is absurd.

      Just add it to your profile once, no one assumes individuals speak for their employers here that would be stupid. The need to add disclaimer would be for the uncommon case that you were speaking for them. It's an anonymous message board we're all just taking here it's not that serious.

      https://hn.algolia.com/?dateRange=all&page=0&prefix=false&qu...

      • ehnto 45 minutes ago
        Some on this forum will be working for companies with conflicts of interest on the topic, and if an employees words were construed to be the opinions of the company that could be bad for that person.
        • sillysaurusx 20 minutes ago
          I was once almost fired for saying a little too much in an HN comment about pentesting. Being dragged into an office and given a dressing-down for posting was quite traumatic.

          The central issue (or so they claimed) was that people might misconstrue my comment as representing the company I was at.

          So yeah, I don’t understand why people are making fun of this. It’s serious.

          On the other hand, they were so uptight that I’m not sure “opinions are my own” would have prevented it. But it would have been at least some defense.

      • jedberg 45 minutes ago
        > Who's else would they be?

        Their employer? They may work at related company, and are required to say this.

      • abosley 32 minutes ago
        The tech companies train their employees to say this in their social media guidance and training.
      • operatingthetan 48 minutes ago
        At this point that phase is an attempt at status signaling.
        • sghiassy 47 minutes ago
          Opinions are my own

          But I think you’re right

        • muyuu 40 minutes ago
          it's hilarious though

          it's like people are LARPing a Fortune company CEO when they're giving their hot takes on social media

          reminds me of Trump ending his wild takes on social media with "thank you for your attention to this matter" - so out of place, it makes it really funny

          *typo

          • PretzelPirate 27 minutes ago
            > it's like people are LARPing a Fortune company CEO when they're giving their hot takes on social media

            At least in large tech companies, they have mandatory social media training where they explicitly tell employees to use phrases like "my views are my own" to keep it clear whether they're speaking on behalf of their employer or not.

            • operatingthetan 20 minutes ago
              If their name is on the post or their company is listed in their profile. The person above has neither as far as I can tell.
            • PKop 18 minutes ago
              Why would they be speaking on behalf of their employers? That is what would need a disclaimer not the common case. Besides, he can put it one time in his profile, not over and over again in every comment like he does. There is no expectation that some random employee is a spokesperson for Google on tech message board comment threads. It's just a way to brag.
            • muyuu 18 minutes ago
              i've worked in two different large tech companies

              when i give my hot takes pseudonymously on social media these phrases would be nothing but a LARP

              i don't put my real name here nor do i put my professional commitments in my profile, and neither does this guy

              • PKop 16 minutes ago
                Exactly. There is no scenario where we should expect some random anon to be speaking for Google. When that is the case a disclaimer is warranted, not the common case of speaking for oneself. He can write it once in his profile if he's so worried about it, not every other comment like he does. It's just inflated self importance
      • xboxnolifes 36 minutes ago
        Its to cover their ass in the event someone makes a stink and quotes them as if its a company opinion.
    • bastardoperator 24 minutes ago
      You think the company that just gave 40B to Anthropic is the winner? Interesting.
      • MattRix 18 minutes ago
        That deal is a win-win for Google. If they develop a better coding model than Anthropic and beat them at coding, then they win. If they don’t, they still win by making a ton of money from Anthropic long term.
      • u_fucking_dork 21 minutes ago
        You think the company that just gave 40B to Anthropic isn’t the winner? Interesting.
        • bastardoperator 12 minutes ago
          Was Microsoft the winner based on their 50B investment in OpenAI?
  • concinds 4 hours ago
    Am I crazy, or was this press release fully rewritten in the past 10 minutes? The current version is around half the length of the old one, which did not frame it as a "simplification" "grounded in flexibility" but as a deeper partnership. It also had word salad about AGI, and said Azure retained exclusivity for API products but not other products, which the new statement seems to contradict.

    What was I looking at?

    • einsteinx2 4 hours ago
      I noticed the exact same thing. I read the original, went back to read it again and it’s completely changed.
      • 3form 2 hours ago
        I think a stickied comment about this would be due. No idea if it's possible to call in @dang via at-name?
        • einsteinx2 2 hours ago
          Looks like they changed the post link to a Bloomberg article instead but kept the comments thread. So I guess he’s already aware.
        • kergonath 47 minutes ago
          > No idea if it's possible to call in @dang via at-name?

          No. Email hn@ycombinator.com

          https://news.ycombinator.com/newsfaq.html

    • alansaber 2 hours ago
      The in-house or the marketing team swooped in last minute it appears
    • antonkochubey 4 hours ago
      They forgot the "hey ChatGPT, rewrite this to have better impact on the company stock" before submitting it
  • _jab 5 hours ago
    This agreement feels so friendly towards OpenAI that it's not obvious to me why Microsoft accepted this. I guess Microsoft just realized that the previous agreement was kneecapping OpenAI so much that the investment was at risk, especially with serious competition now coming from Anthropic?
    • DanielHB 4 hours ago
      Microsoft is a major shareholder of OpenAI, they don't want their investment to go to 0. You don't just take a loss on a multiple-digit billion investment.
      • snowwrestler 2 hours ago
        I think you’re right about this deal. But it’s kind of funny to think back and realize that Microsoft actually has just written off multi-billion-dollar deals, several times in fact.
        • nacozarina 2 hours ago
          One (1) year after M$ bought Nokia they wrote it off for $7.6 Billion.

          There’s no upper limit to their financial stupidity.

          • snek_case 2 hours ago
            The metaverse is another example if anyone doubts the bounds of corporate stupidity.
            • lesuorac 1 hour ago
              Why?

              FaceBook largely requires an Apple iPhone, Apple computer, "Microsoft" computer, "Google" phone, or a "Google" computer to use it. At any point one of those companies could cut FaceBook off (ex. [1]).

              The Metaverse was a long term goal to get people onto a device (Occulus) that Meta controlled. While I think an AR device is much more useful than VR; I'm not convinced that it's a mistake for Meta to peruse not being beholden to other platforms.

              [1]: https://arstechnica.com/gadgets/2019/01/facebook-and-google-...

              • everforward 29 minutes ago
                I think this is sane washing their idea in the modern context of it having failed. I think at the time, they thought VR would be the next big thing and wanted to become the dominant player via first mover advantage.

                The headsets don’t really make sense to me in the way you’re describing. Phones are omnipresent because it’s a thing you always just have on you. Headsets are large enough that it’s a conscious choice to bring it; they’re closer to a laptop than a phone.

                Also, the web interface is like right there staring at them. Any device with a browser can access Facebook like that. Google/Apple/Microsoft can’t mess with that much without causing a huge scene and probably massive antitrust backlash.

                • Aerroon 15 minutes ago
                  I think headsets might work, but I think Meta trying to use their first mover advantage so hard so early backfired. Oculus, as a device, became less desirable after it required Facebook integration.

                  It's kind of like Microsoft with copilot - the idea about having an AI assistant that can help you use the computer is great. But it can't be from Microsoft because people don't trust them with that.

              • latexr 1 hour ago
                > I'm not convinced that it's a mistake for Meta to peruse not being beholden to other platforms.

                Devoid of other context, it’s hard to disagree. But your parent comment only asserted that the metaverse specifically as proposed by Facebook was an obviously stupid idea.

              • corford 51 minutes ago
                >Why?

                Patrick Boyle did a nice video a few weeks back: https://www.youtube.com/watch?v=8BaSBjxNg-M

              • IshKebab 1 hour ago
                Because it's been very clear for a long time that the vast majority of people do not want to play VR Second Life.
                • snek_case 36 minutes ago
                  Meta's vision was worse than that. They were trying to hype doing work meetings in VR. There's a case to be made that VR games and VR universes can be fun... But work meetings?
                  • ethbr1 24 minutes ago
                    Mark Zuckerberg using his company to build things he's the primary user for?
              • PKop 45 minutes ago
                Because it's been a massively expensive failure. They can't just will their own platform into existence just because it would be good to have, consumers have a say and they've rejected it completely.
              • turtlesdown11 1 hour ago
                so after $80 billion spent, they must have an ecosystem of hundreds of millions of users? Right?

                Maybe they should have spent that on the facebookphone

    • dkrich 4 hours ago
      Probably more that they are compute constrained. In his latest post Ben Thompson talks about how Microsoft had to use their own infrastructure and supplant outside users in the process so this is probably to free up compute.
    • dinosor 4 hours ago
      > Microsoft will no longer pay a revenue share to OpenAI.

      I feel this looks like a nice thing to have given they remain the primary cloud provider. If Azure improves it's overall quality then I don't see why this ends up as a money printing press as long as OpenAI brings good models?

      • JumpCrisscross 4 hours ago
        OpenAI was also threatening to accuse "Microsoft of anticompetitive behavior during their partnership," an "effort [which] could involve seeking federal regulatory review of the terms of the contract for potential violations of antitrust law, as well as a public campaign" [1].

        [1] https://www.wsj.com/tech/ai/openai-and-microsoft-tensions-ar...

      • aurareturn 4 hours ago
        Does this mean Microsoft gets OpenAI's models for "free" without having to pay them a dime until 2032?

        And on top of that, OpenAI still has to pay Microsoft a share of their revenue made on AWS/Google/anywhere until 2030?

        And Microsoft owns 27% of OpenAI, period?

        That's a damn good deal for Microsoft. Likely the investment that will keep Microsoft's stock relevant for years.

        • dzonga 3 hours ago
          own 27%. but are entitled to OpenAI profits of 49% for eternity (if OpenAI is profitable or government steps in)
          • aurareturn 1 hour ago

              own 27%. but are entitled to OpenAI profits of 49% for eternity (if OpenAI is profitable or government steps in)
            
            Where is the 49% coming from? The new deal does not talk about that.
      • lokar 4 hours ago
        Does anyone expect azure quality to improve? Has it improved at all in the last 3 years? Does leadership at MS think it needs to improve?

        I doubt it

        • gchamonlive 2 hours ago
          No and at this point tying yourself to azure is a strategic passive and anyone making such decisions should be held responsible for any service outage or degradation.
        • alternatex 2 hours ago
          MS incentivizes feature quantity, and the leadership are employees like any other. Product improvements are not on the table unless the company starts promoting people based on it. Doesn't look this will start happening any time soon.
        • jakeydus 4 hours ago
          Don’t worry I’m sure there’s a few products without copilot integration still. They’ll get to them before too long.
    • HWR_14 2 hours ago
      This is probably a delayed outgrowth of the negotiations last year, where Microsoft started trading weird revenue shares and exclusivity for 27% of the company.
    • guluarte 2 hours ago
      I think MS wants OpenAI to fail so it can absorb it
      • Oras 2 hours ago
        MS put 10B for 50% if I remember correctly. OpenAI is worth many multiples of that.
        • marricks 2 hours ago
          > OpenAI is worth many multiples of that

          valued at --which I'd say is a reasonable distinction to make right about now

        • HWR_14 2 hours ago
          When they put 10B in, they got weird tiered revenue shares and other rights. That has been simplified to 27% of OpenAI today. I don't know what that meant their 10B would be worth before dilution in later rounds.
        • bmitc 2 hours ago
          > OpenAI is worth many multiples of that.

          How?

          • senordevnyc 2 hours ago
            Because they recently issued shares at a price many multiples of that, and people bought them. How else would you define financial worth?
            • andriy_koval 1 hour ago
              I would use your number adjusted by some demand elasticity curve.
              • tanseydavid 50 minutes ago
                The "back-of-the-napkin" only has enough room to estimate based on recently issued share price. Seems reasonable to me.
  • synergy20 43 minutes ago
    Microsoft won the first around, now it's lagging far behind. CEO needs to go, it's so hard to ruin a play this badly.
    • ethbr1 27 minutes ago
      Ah, so a familiar position for them, then!
    • dominotw 8 minutes ago
      what could ceo have done
  • chasd00 4 hours ago
    This gives OpenAI the ability to goto AWS instead of exclusively on Azure. I guess Azure really is hanging on by a thread.

    https://news.ycombinator.com/item?id=47616242

    • elpakal 1 hour ago
    • xvilka 4 hours ago
      And Azure still doesn't support IPv6, looking at the GitHub[1].

      [1] https://github.com/orgs/community/discussions/10539

      • jabl 3 hours ago
        Perhaps they should use OpenAI models to figure out how to rollout IPv6.
        • brazukadev 26 minutes ago
          Now they can use Claude Code.
      • WorldMaker 3 hours ago
        I was under the impression that as long as GitHub doesn't support IPv6 it is a sign that they still haven't finished their migration to Azure. Azure supports IPv6 just fine.
        • depr 2 hours ago
          Supports IPv6 just fine? Absolutely not, they have the worst IPv6 implementation of the 3 large clouds, where many of their products don't support it, such as their Postgres offering. See https://news.ycombinator.com/item?id=44881803 for more.
      • happyPersonR 3 hours ago
        lol GitHub doesn’t run on azure at msft

        They still run their own platform.

      • awestroke 3 hours ago
        Well, you see, they just can't find a checkbox for ipv6 support in the IIS GUI on their ingress servers.
    • Donald 4 hours ago
      Isn't this expected if OpenAI models are going to be listed on AWS GovCloud as a part of the Anthropic / Hegseth fall-out?
    • torginus 3 hours ago
      What? I thought Azure will always have the Sharepoint/Office/Active Directory cash cow.
      • isk517 3 hours ago
        Their engineers have been working tirelessly to make Sharepoint/Office/Active Directory as terrible as it possibly could be while still technically being functional, while continuing to raise prices on them. I've seen many small business start to chose Google Workspace over them, the cracks have formed and are large enough that they are no longer in a position were every business just go with Office because that's what everyone uses.
        • hirako2000 19 minutes ago
          I see more businesses on the office + Team stack then Google workspace. So far more.

          I think the differentiator is Team, which Google for some mysterious reason can't build or doesn't want to.

        • ethbr1 5 minutes ago
          Sharepoint has never not been terrible.
  • freediddy 3 hours ago
    Nadella had OpenAI by the short and curlies early on. But all I've seen from him in the last couple of years is continuously acquiescing to OpenAI's demands. I wonder why he's so weak and doesn't exert more control over the situation? At one point Microsoft owned 49% of OpenAI but now it's down to 27%?
    • dijit 3 hours ago
      Everything is personal preference, and perhaps I am more fiscally conservative because I grew up in poverty.

      But if I own 49% of a company and that company has more hype than product, hasn't found its market yet but is valued at trillions?

      I'm going to sell percentages of that to build my war chest for things that actually hit my bottom line.

      The "moonshot" has for all intents and purposes been achieved based on the valuation, and at that valuation: OpenAI has to completely crush all competition... basically just to meet its current valuations.

      It would be a really fiscally irresponsible move not to hedge your bets.

      Not that it matters but we did something similar with the donated bitcoin on my project. When bitcoin hit a "new record high" we sold half. Then held the remainder until it hit a "new record high" again.

      Sure, we could have 'maxxed profit!'; but ultimately it did its job, it was an effective donation/investment that had reasonably maximal returns.

      (that said, I do not believe in crypto as an investment opportunity, it's merely the hand I was dealt by it being donated).

      • freediddy 3 hours ago
        Microsoft didn't sell anything. OpenAI created more shares and sold those to investors, so Microsoft's stake is getting diluted.

        And Microsoft only paid $10B for that stake for the most recognizable name brand for AI around the world. They don't need to "hedge their bets" it's already a humongous win.

        Why let Altman continue to call the shots and decrease Microsoft's ownership stake and ability to dictate how OpenAI helps Microsoft and not the other way around?

        • zozbot234 3 hours ago
          > They don't need to "hedge their bets" it's already a humongous win.

          That's a flawed argument. Why wouldn't you want to hedge a risky bet, and one that's even quite highly correlated to Microsoft's own industry sector?

        • theplatman 2 hours ago
          do we know whether Microsoft could have been selling secondary shares as part of various funding rounds?

          my impression is that many of these "investments" are structured IOUs for circular deals based on compute resources in exchange for LLM usage

        • tonyedgecombe 3 hours ago
          About the same as they wasted on Nokia.
      • GardenLetter27 30 minutes ago
        It's not hype, the demand for inference has grown more this year than expected.
      • saaaaaam 3 hours ago
        I don’t understand the “record high” point. How did you decide when a “record high” had been reached in a volatile market? Because at $1 the record high might be $2 until it reaches $3 a week or month later. How did you determine where to slice on “record highs”?

        Genuine question because I feel like I’m maybe missing something!

        • dijit 2 hours ago
          The short answer is: it's the secretary problem.

          The longer answer is; you never know whats coming next, bitcoin could have doubled the day after, and doubled the day after that, and so on, for weeks. And by selling half you've effectively sacrificed huge sums of money.

          The truth is that by retaining half you have minimised potential losses and sacrificed potential gains, you've chosen a middle position which is more stable.

          So, if bitcoin 1000 bitcoing which was word $5 one day, and $7 the next, but suddenly it hits $30. Well, we'd sell half.

          If the day after it hit $60, then our 500 remaining bitcoins is worth the same as what we sold, so in theory all we lost was potential gains, we didn't lose any actual value.

          Of course, we wouldn't sell we'd hold, and it would probably fall down to $15 or something instead.. then the cycle begins again..

      • senordevnyc 2 hours ago
        It’s not more hype than product, it has found a market (making many billions in revenue), and it’s not valued at trillions. So wrong on all counts.
      • solumunus 3 hours ago
        They haven’t sold anything they’ve been diluted.
        • hirako2000 13 minutes ago
          A company can dilute just like that?
    • gessha 37 minutes ago
      If Sam continues doing Sam things, MS might get 0% of OpenAI if Satya insists on the previous contract. Either by closing up OpenAI and opening up OpaenAI and/or by MS suing it out of existence. It’s all about what MS can get out of it. If they can get 27% of something rather than nothing, they’re better off.
    • tyre 1 hour ago
      They had to negotiate away the non-profit structure of OpenAI. Sam used that as a marketing and recruiting tool, but it had outlived that and was only a problem from then on.

      For OAI to be a purely capitalist venture, they had to rip that out. But since the non-profit owned control of the company, it had to get something for giving up those rights. This led to a huge negotiation and MSFT ended up with 27% of a company that doesn’t get kneecapped by an ethical board.

      In reality, though, the board of both the non-profit and the for profit are nearly identical and beholden to Sam, post–failed coup.

    • PunchyHamster 3 hours ago
      Why would they acquire more when company is still not making profit ? To be left with bigger bag ?
  • ZeroCool2u 5 hours ago
    Interesting side effect of this is that Google Cloud may now be the only hype scaler that can resell all 3 of the labs models? Maybe I'm misinterpreting this, but that would be a notable development, and I don't see why Google would allow Gemini to be resold through any of the other cloud providers.

    Might really increase the utility of those GCP credits.

    • aurareturn 4 hours ago
      Might not be good for Gemini long term if Anthropic and OpenAI can and will sell in every cloud provider they can find but businesses can only use Gemini via Google Cloud.
      • jfoster 4 hours ago
        Good for Google Cloud, bad for Gemini = ??? for Google
      • Melatonic 2 hours ago
        Except Gemini might end up being far cheaper per token due to the infrastructure advantage
        • aurareturn 1 hour ago
          Do we have proof that it's cheaper in terms of $/token/intelligence?
          • Melatonic 1 hour ago
            I think the public pricing usually has it cheaper (relatively). Obviously since AI is constantly evolving it's not going to compare as favourably farther to a major Gemini release

            I was mainly referring to the TPU hardware advantage + GCP running and designing their own datacenter stack.

      • stavros 4 hours ago
        How is it good for Gemini that it's not available on two out of three major cloud platforms?
        • aurareturn 4 hours ago
          It isn't. That's why I said "might not be good for Gemini".
          • stavros 4 hours ago
            Oof, I completely missed that "not", thanks.
    • gowld 2 hours ago
      "hype scaler" indeed!
    • retinaros 4 hours ago
      that will likely mean the end of gemini models...
  • 1f60c 4 hours ago
    Wait, I thought OpenAI had to pay Microsoft until AGI was achieved or something? Am I misremembering? Is that a different thing?
    • ksherlock 3 hours ago
      Per WSJ, previously, they both had revenue sharing agreements. MSFT will no longer send any revenue to OpenAI. OpenAI will still send revenue to MSFT until 2030 (with new caps)
    • staminade 3 hours ago
      My understand was that was in relation to IP licensing. Microsoft got access to anything OpenAI built unless they declared they had developed AGI. This new article apparently unlinks revenue sharing from technology progress, but it's unclear to me if it changes the situation regarding IP if OpenAI (claim to) have achieved AGI.
    • dist-epoch 3 hours ago
      [dead]
  • alexdoesstuff 1 hour ago
    It's kind of shocking, given financial transparency, that Microsoft gets away with not disclosing any details of this agreement (or the one it is replacing) to its shareholders. We know there's a cap on the revenue share from OpenAI to Microsoft, but we have no idea what that cap is (not whether it's higher, lower, or unchanged from the prior agreement).

    We have no idea what it means to be the "primary cloud provider" and have the products made available "first on Azure". Does MSFT have new models exclusively for days, weeks, months, or years?

    Both facts and more details from the agreement are quite frankly highly relevant to judge whether this is a net positive, negative or neutral for MSFT. It's unbelievable that the SEC doesn't force MSFT to publish at least an economic summary of the deal.

    • trvz 26 minutes ago
      It’s American Business as usual. Personally I’m miffed how little data Apple needs to provide about product categories, and especially about how much they’ve burnt on the car program. If they shared any data about that at all some the leadership might end up having to take responsibility for mismanagement…
  • sourraspberry 4 hours ago
    The disparity in coverage on this new deal is fascinating. It feels like the narrative a particular outlet is going with depends entirely on which side leaked to them first.
    • scottyah 2 hours ago
      Just some of the games sama is playing.
  • saadn92 2 hours ago
    That's a pretty good swap if you're Microsoft. Exclusivity was already unenforceable in practice, and they were going to have to either sue their biggest AI partner or let it slide. Instead they got the agi escape hatch closed and a revenue cap that at least makes the payments predictable
  • leonardoaraujo 8 minutes ago
    Basically it seems that they didn't found yet a way to make money out of their models to keep the lights on...
  • aurareturn 5 hours ago

      Microsoft Corp. will no longer pay revenue to OpenAI and said its partnership with the leading artificial intelligence firm will not be exclusive going forward.
    
    What does this mean that Microsoft will no longer pay revenue to OpenAI? How did the original deal work?
    • alexdoesstuff 47 minutes ago
      It's unclear. That was never disclosed. It's similarly unclear what it means that they will no longer pay revenue share to OpenAI. Do they get the models for free now? How does OpenAI make money from the models hosted on Azure if not via revenue share?
    • justinclift 2 hours ago
      Wonder if this means Microsoft is actually going to be deploying Claude Code internally for usage?

      That might help fix some of the bugs in Teams... :)

    • Handy-Man 5 hours ago
      They were paying them 20% of the revenue from the hosted OpenAI products I believe?
      • bilbo0s 4 hours ago
        Does this mean they will host OpenAI products but not pay them? Or does it mean they are paying them in some other way?
        • HarHarVeryFunny 4 hours ago
          It seems that the old deal was exclusivity to MSFT with revenue share, and now no exclusivity, no revenue share.

          Bear in mind that MSFT have rights to OpenAI IP (as well as owning ~30% of them). The only reason they were giving revenue share was in return for exclusivity.

          • borski 3 hours ago
            This is a really common way to structure exclusivity; we did the same thing whenever customers requested it (and we couldn’t get rid of it entirely). Charge for the exclusivity explicitly.

            If they wanted named exclusivity rather than general exclusivity, we would charge a somewhat smaller amount for each competitor they wanted exclusivity from. They could give up exclusivity at any time.

            That was precisely how we structured our deal with Azure, back in 2014-2016 or so.

        • deaux 3 hours ago
          Azure was the only non-OpenAI provider that was allowed to provide OpenAI models. The comparison here is with Anthropic whose models are on both GCP and AWS (and technically also Azure though I think that might just be billing passthrough to Anthropic).
        • Handy-Man 4 hours ago
          I suppose continue to host until the 2030/32 that they have access to but not share revenues when they use those models for their products like the bazillions of Copilots.
  • aurareturn 5 hours ago
    The original "AGI" agreement was always a bit suspect and open to wild interpretations.

    I think this is good for OpenAI. They're no longer stuck with just Microsoft. It was an advantage that Anthropic can work with anyone they like but OpenAI couldn't.

  • JumpCrisscross 4 hours ago
    It's unclear which elements of this new deal are binding versus promises with OpenAI characteristics. "Microsoft Corp. will publish fiscal year 2026 third-quarter financial results after the close of the market on Wednesday, April 29, 2026" [1]; I'd wait for that before jumping to conclusions.

    [1] https://news.microsoft.com/source/2026/04/08/microsoft-annou...

  • eranation 2 hours ago
    So, silly question, does this mean I will be able to get OpenAI models via Bedrock soon?
    • aenis 47 minutes ago
      Likely, and via vertex on gcp (or whatever they are calling it this year).

      Which also means, if you are a big boring AWS or GCP shop, and have a spend commitment with either as part of a long term partnership, it will count towards that. And, you won't likely have to commit to a spend with OpenAI if you want the EU data residency for instance. And likely a bit more transparency with infra provisioning and reserved capacity vs. OpenAI. All substantial improvements over the current ways to use OpenAI in real production.

  • cdrnsf 2 hours ago
    OpenAI's logo is actually a depiction of their financial connections.
  • chasil 4 hours ago
  • GardenLetter27 32 minutes ago
    Hopefully they put ChatGPT on Bedrock now.
  • airstrike 4 hours ago
    Kagi Translate was kind enough to turn this from LinkedIn Speak to English:

    The Microsoft and OpenAI situation just got messy.

    We had to rewrite the contract because the old one wasn't working for anyone. Basically, we’re trying to make it look like we’re still friends while we both start seeing other people. Here is what’s actually happening:

    1. Microsoft is still the main guy, but if they can't keep up with the tech, OpenAI is moving out. OpenAI can now sell their stuff on any cloud provider they want.

    2. Microsoft keeps the keys to the tech until 2032, but they don't have the exclusive rights anymore.

    3. Microsoft is done giving OpenAI a cut of their sales.

    4. OpenAI still has to pay Microsoft back until 2030, but we put a ceiling on it so they don't go totally broke.

    5. Microsoft is still just a big shareholder hoping the stock goes up.

    We’re calling this "simplifying," but really we’re just trying to build massive power plants and chips without killing each other yet. We’re still stuck together for now.

    • azinman2 4 hours ago
      This was actually really helpful. I feel like it should be done for all PR speak.
      • JumpCrisscross 4 hours ago
        It's better than the original, but still off.

        "The Microsoft and OpenAI situation just got messy" is objectively wrong–it has been messy for months [1]. Nos. 1 through 3 are fine, though "if they can't keep up with the tech, OpenAI is moving out" parrots OpenAI's party line. No. 4 doesn't make sense–it starts out with "we" referring to OpenAI in the first person but ends by referring to them in the third person "they." No. 5 is reductive when phrased with "just."

        It would seem the translator took corporate PR speak and translated it into something between the LinkedIn and short-form blogger dialects.

        [1] https://www.wsj.com/tech/ai/openai-and-microsoft-tensions-ar...

        • Maxatar 4 hours ago
          Being objectively correct isn't the goal of the translator, the translator can't possibly know if a statement is truthful. What the translator does is well... translate, specifically from some kind of corporate speak that is really difficult for many people including myself to understand, into something more familiar.

          I don't expect the translation to take OpenAI's statements and make them truthful or to investigate their veracity, but I genuinely could not understand OpenAI's press release as they have worded it. The translation at least makes it easier to understand what OpenAI's view of the situation is.

          • ghostly_s 4 hours ago
            > The only only pure fuck-up I'd call out is switching from third to first person when referring to OpenAI in the same sentence (No. 4).

            "We" in this sentence refers to both parties; "they" refers to OpenAI. Not a grammatical error.

            • JumpCrisscross 4 hours ago
              > "We" in this sentence refers to both parties

              Fair enough.

              > "they" refers to OpenAI. Not a grammatical error

              I'd say it is. It's a press release from OpenAI. The rest of the release uses the third-person "they" to refer to Microsoft. The LLM traded accuracy for a bad joke, which is someting I associate with LinkedIn speak.

              The fundmaental problem might be the OpenAI press release is vague. (And changing. It's changed at least once since I first commented.)

            • auscompgeek 4 hours ago
              In isolation sure. But in context with the other points it makes it look like "they" refers to Microsoft in all the dot points.
        • matthewkayin 1 hour ago
          > "The Microsoft and OpenAI situation just got messy" is objectively wrong–it has been messy for months

          I'm pretty sure "just" is being used here to mean "simply" rather than "recently".

    • MarleTangible 4 hours ago
    • singingtoday 2 hours ago
      Thank you for this!

      That's kagi? Cool, I'm check out out more!

    • j_maffe 1 hour ago
      This is somehow even less helpful than the og article.
  • herodoturtle 1 hour ago
    Interesting timing when one also considers that the Musk vs OpenAI trial is set to get underway.

    https://www.dw.com/en/musk-vs-openai-trial-to-get-underway/a...

  • jryio 4 hours ago
    > OpenAI has contracted to purchase an incremental $250B of Azure services, and Microsoft will no longer have a right of first refusal to be OpenAI’s compute provider.

    Azure is effectively OpenAI's personal compute cluster at this scale.

    • JumpCrisscross 4 hours ago
      What fraction of Azure compute does OpenAI represent? (Does the $250bn commitment have a time period? Is it legally binding?)
      • runako 4 hours ago
        Azure did $75B last quarter.

        That article doesn't give a timeframe, but most of these use 10 years as a placeholder. I would also imagine it's not a requirement for them to spend it evenly over the 10 years, so could be back-loaded.

        OpenAI is a large customer, but this is not making Azure their personal cluster.

    • einrealist 4 hours ago
      I wonder how this figure was settled. Is it based on consumer pricing? Can't Microsoft and OpenAI just make a number up, aside from a minimum to cover operating costs? When is the number just a marketing ploy to make it seem huge, important and inevitable (and too big to fail)?
  • exec7 1 hour ago
    Good news for openAI, microsoft is the main blocker of innovation in the tech industry!
  • muyuu 35 minutes ago
    sounds like divesting behind a bit of nice-sounding scaffolding
  • monkeydust 4 hours ago
  • martinald 4 hours ago
    Really interesting. Why would Microsoft have done this deal? I'm a bit lost. Sure they get to not pay a revenue share _to_ OpenAI but surely that's limited to just OpenAI products which is probably a rounding error? Losing exclusivity seems like a big issue for them?
  • Eridrus 4 hours ago
    Biggest upside of this is I expect OpenAI models to be available on Bedrock, which is huge for not having to go back to all your customers with data protection agreements.
    • easton 4 hours ago
      Isn’t that an “API product”? I read this assuming the whole point of renegotiation was to let OpenAI sell raw inference via bedrock, but that still seems to be blocked except for selling to the US Government.
    • fengkx 4 hours ago
      > OpenAI can now jointly develop some products with third parties. API products developed with third parties will be exclusive to Azure. Non-API products may be served on any cloud provider.

      This seems impossible.

  • dhruv3006 1 hour ago
    I think aws will seize the opportunity.
  • 31276 4 hours ago
    Pursue "new opportunities"? Microslop is dumping OpenAI and wishes it well in its new endeavors.
    • aurareturn 4 hours ago
      I read this as the other way. OpenAI was desperate to dump Microsoft.
    • iewj 4 hours ago
      In retrospect all those OAI announcements are gonna look so cringe.

      They did not need to go so hard on the hype - Anthropic hasn’t in relative terms and is generating pretty comparable revenues at present.

      • JumpCrisscross 4 hours ago
        > They did not need to go so hard on the hype - Anthropic hasn’t in relative terms and is generating pretty comparable revenues at present

        OpenAI bet on consumers; Anthropic on enterprise. That will necessitate a louder marketing strategy for the former.

        • eieiw 4 hours ago
          That’s funny.

          Why is it Altman is facing kill shots and Dario isn’t?

          • scottyah 1 hour ago
            Dario is a lot more focused on enabling people with AI, Sam goes on interviews like he's Wormtongue trying to summon a "god". Then there is the whole "open"ai where he took it closed source for profit, the engineers kicking sama out but he wiggled back in (at the cost of a lot of the founding engineers), the suspicious death of a whistleblower, the crazy investment schemes of billions of dollars that he's hoping taxes will save him from, the immediate curtailing to Pete in the DoD, and a few other things that make him at least a highly questionable fellow.

            Dario left OpenAI because of the bad he saw there, and made a superior product (though these things change very rapidly).

          • JumpCrisscross 4 hours ago
            > Why is it Altman is facing kill shots and Dario isn’t?

            Altman peaked in the zeiteist in 2023; Dario, much less prominently, in 2024 and now '26 [1]. I'd guess around this time next year, Dario will be as hated as Altman is today.

            [1] https://trends.google.com/explore?q=altman%2C%20Dario&date=t...

  • jachva95 1 hour ago
    Why are do I see bloomberg links so often when this shit won't even let you read article without sub ? Do you not have better reasons to spend money?
  • sayYayToLife 2 hours ago
    Alright my theory:

    OpenAI has public models that are pretty 'meh', better than Grok and China, but worse than Google and Anthropic. They still cost a ton to run because OpenAI offers them for free/at a loss.

    However, these people are giving away their data, and Microsoft knows that data is going to be worthwhile. They just dont want to pay for the electricity for it.

    • alexdoesstuff 42 minutes ago
      Small nitpick: the models probably make some money on actual inference. Might not be a massive amount, but hard to see them not having a positive contribution margin purely on inference.

      What's losing OpenAI money is paying for the whole of R&D, including training and staff. Microsoft doesn't pay that, so they get the money making part of AI without the associated costs.

  • airstrike 4 hours ago
    "Advancing Our Amazing Bet" type post
  • Schlagbohrer 4 hours ago
    The AGI talk is shocking but not surprising to anyone looking at how bombastic Sam Altman's public statements are.

    The circular economy section really is shocking- OpenAI committing to buying $250 Billion of Azure services, while MSFT's stake is clarified as $132 Billion in OpenAI. Same circular nonsense as NVIDIA and OpenAI passing the same hundred billion back and forth.

    • ModernMech 4 hours ago
      Dennis: I think we made every single one of our Paddy's Dollars back, buddy.

      Mac: You're damn right. Thus creating the self-sustaining economy we've been looking for.

      Dennis: That's right.

      Mac: How much fresh cash did we make?

      Dennis: Fresh cash! Uh, well, zero. Zero if you're talking about U.S. currency. People didn't really seem interested in spending any of that.

      Mac: That's okay. So, uh, when they run out of the booze, they'll come back in and they'll have to buy more Paddy's Dollars. Keepin' it moving.

      Dennis: Right. That is assuming, of course, that they will come back here and drink.

      Mac: They will! They will because we'll re-distribute these to the Shanties. Thus ensuring them coming back in, keeping the money moving.

      Dennis: Well, no, but if we just re-distribute these, people will continue to drink for free.

      Mac: Okay...

      Dennis: How does this work, Mac?

      Mac: The money keeps moving in a circle.

      Dennis: But we don't have any money. All we have is this. ... How does this work, dude!?

      Mac: I don't know. I thought you knew.

      • ksimukka 3 hours ago
        Great scene
      • slickytail 4 hours ago
        You forgot the best line: "I don't know how the US economy works, much less some kind of self-sustaining one".
  • jhk482001 3 hours ago
    So AWS can finally use OpenAI and not only OSS version.
  • m3kw9 4 hours ago
    Looks like MS is shafting OpenAI.
  • TheAtomic 4 hours ago
    "We want to sell surveillance services to the US gov. MSFT was hesitant so we gave ourselves room to do it without them."
    • Schlagbohrer 4 hours ago
      Extremely hard to believe that MSFT would have any hesitancy about working with the US government.
  • shevy-java 2 hours ago
    Two evil walk away. Well, is that good or bad?

    I fear for the end user we'll still see more open-microslop spam. I see that daily on youtube - tons of AI generated fakes, in particular with that addictive swipe-down design (ok ok, youtube is Google but Google is also big on the AI slop train).

  • delis-thumbs-7e 4 hours ago
    It’s insane how they talk about AGI, like it was some scientifically qualifiable thing that is certain to happen any time now. When I have become the javelin Olympic Champion, I will buy a vegan ice cream to everyone with a HN account.
    • jmward01 2 hours ago
      I think we keep changing the goalposts on AGI. If you gave me CC in the 80's I would probably have called it 'alive' since it clearly passes the Turing test as I understood it then (I wouldn't have been able to distinguish it from a person for most conversations). Now every time it gets better we push that definition further and every crack we open to a chasm and declare that it isn't close. At the same time there are a lot of people I would suspect of being bots based on how they act and respond and a lot of bots I know are bots mainly because they answer too well.

      Maybe we need to start thinking less about building tests for definitively calling an LLM AGI and instead deciding when we can't tell humans aren't LLMs for declaring AGI is here.

      • sho_hn 1 hour ago
        > I think we keep changing the goalposts on AGI

        Isn't that exactly what you would expect to happen as we learn more about the nature and inner workings of intelligence and refine our expectations?

        There's no reason to rest our case with the Turing test.

        I hear the "shifting goalposts" riposte a lot, but then it would be very unexciting to freeze our ambitions.

        At least in an academic sense, what LLMs aren't is just as interesting as what they are.

        • breezybottom 1 hour ago
          I think the advancement in AI over the last four years has greatly exceeded the advancement in understanding the workings of human intelligence. What paradigm shift has there been recently in that field?
          • smcg 1 hour ago
            What have we learned that isn't in my textbook from the 90s?
            • echelon 1 hour ago
              > What have we learned that isn't in my textbook from the 90s?

              Does it matter?

              We can do countless things people in the 90's would think was black magic.

              If I showed the kid version of myself what I can do with Opus or Nano Banana or Seedance, let alone broadband and smartphones, I think I'd feel we were living in the Star Trek future. The fact that we can have "conversations" with AI is wild. That we can make movies and websites and games. It's incredible.

              And there does not seem to be a limit yet.

        • charcircuit 1 hour ago
          I would agree with you if we were talking about trying to replicate some form of general intelligence, but we are talking about creating artificial intelligence.
      • sn0wr8ven 1 hour ago
        I don't think the goalpost has been shifted for AGI or the definition of AGI that is used by these corporations. It's just they broke it down to stages to claim AGI achieved. It was always a model or system that surpasses human capabilities at most tasks/being able to replace a human worker. The big companies broke it down to AGI stage 1, stage 2, etc to be able to say they achieved AGI.

        The Turing Test/Imitation Game is not a good benchmark for AGI. It is a linguistics test only. Many chatbots even before LLMs can pass the Turing Test to a certain degree.

        Regardless, the goalpost hasn't shifted. Replacing human workforce is the ultimate end goal. That's why there's investors. The investors are not pouring billions to pass the Turing Test.

        • turtlesdown11 1 hour ago
          AGI moved from a technical goal to a marketing term
      • _russross 1 hour ago
        Turing himself argued that trying to measure if a computer is intelligent is a fool's errand because it is so difficult to pin down definitions. He proposed what we call the "Turing test" as a knowable, measurable alternative. The first paragraph of his paper reads:

        > I propose to consider the question, "Can machines think?" This should begin > with definitions of the meaning of the terms "machine" and "think." The > definitions might be framed so as to reflect so far as possible the normal use > of the words, but this attitude is dangerous, If the meaning of the words > "machine" and "think" are to be found by examining how they are commonly used > it is difficult to escape the conclusion that the meaning and the answer to the > question, "Can machines think?" is to be sought in a statistical survey such as > a Gallup poll. But this is absurd. Instead of attempting such a definition I > shall replace the question by another, which is closely related to it and is > expressed in relatively unambiguous words.

        Many people who want to argue about AGI and its relation to the Turing test would do well to read Turing's own arguments.

        • redox99 57 minutes ago
          The Turing test ended up being kind of a flop. We basically passed it and nobody cared. That's because the turing test is about whether a machine can fool a human, not about its intelligent capabilities per se.
          • anthonyrstevens 42 minutes ago
            No, it's because certain people moved the goal posts. Nothing an LLM does or will do will make them belive that it's "intelligent" because they have a mental model of "intelligence" that is more religious than empirical.
      • zug_zug 1 hour ago
        I don't think so... I think most of the sci-fi I grew up reading presented AGI that could reason better than humans could, like make a plan and carry it out.

        Like do people not know what word "general" means? It means not limited to any subset of capabilities -- so that means it can teach itself to do anything that can be learned. Like start a business. AI today can't really learn from its experiences at all.

      • Zambyte 1 hour ago
        Related: https://en.wikipedia.org/wiki/AI_effect

        The truth is, we have had AGI for years now. We even have artificial super intelligence - we have software systems that are more intelligent than any human. Some humans might have an extremely narrow subject that they are more intelligent than any AI system, but the people on that list are vanishing small.

        AI hasn't met sci-fi expectations, and that's a marketing opportunity. That's all it is.

        • baq 1 hour ago
          AGI in the common man's world model is ASI in the AI researcher's definitions, i.e. something obviously smarter at anything and everything you could ask it for regardless of how good of an expert you are in any domain.

          also, I'm pretty sure some people will move goalposts further even then.

        • fragmede 51 minutes ago
          Hasn't met your sci-fi expectations, maybe. I pull a computer out of my pocket, and talk with it. Sure, I gets tripped up here and there, but take a step back, holy shit that's freaking amazing! I don't have a flying car or transparent aluminum, and society has its share of issues right now, but my car drives itself. Coming from the 90's, I think living in the sci-fi future! (Only question is, which one.)
      • pron 1 hour ago
        The Turing test pits a human against a machine, each trying to convince a human questioner that the other is the machine. If the machine knows how humans generally behave, for a proper test, the human contestant should know how the machine behaves. I think that this YouTube channel clearly shows that none of today's models pass the Turing test: https://www.youtube.com/@FatherPhi
      • lesuorac 1 hour ago
        > Maybe we need to start thinking less about building tests for definitively calling an LLM AGI and instead deciding when we can't tell humans aren't LLMs for declaring AGI is here.

        If you've never read the original paper [1] I recommend that you do so. We're long past the point of some human can't determine if X was done by man or machine.

        [1]: https://courses.cs.umbc.edu/471/papers/turing.pdf

      • applfanboysbgon 55 minutes ago
        People thought Eliza was alive too in the 60s. AGI is not determined by how ignorant, uninformed humans view a technology they don't understand. That is the single dumbest criterion you could come up with for defining it.

        Regarding shifting goalposts, you are suggesting the goalposts are being moved further away, but it's the exact opposite. The goalposts are being moved closer and closer. Someone from the 50s would have had the expectation that artificial intelligence ise something recognisable as essentially equivalent to human intelligence, just in a machine. Artificial intelligence in old sci-fi looked nothing like Claude Code. The definition has since been watered down again and again and again and again so that anything and everything a computer does is artificial intelligence. We might as well call a calculator AGI at this point.

      • zendist 59 minutes ago
        The goal post keeps moving because LLM hypeists keep saying LLMs are "close" to AGI (or even are, already). Any reasonably intelligent individual that knows anything about LLMs obviously rejects those claims, but the rest of the world doesn't.

        An AGI would not have problems reading an analog clock. Or rather, it would not have a problem realizing it had a problem reading it, and would try to learn how to do it.

        An AGI is not whatever (sophisticated) statistical model is hot this week.

        Just my take.

        • redox99 53 minutes ago
          Vision is still much weaker than text for LLMs. So you could argue we already have AGI for text but not vision inputs, or you could argue AGI requires being human level at text vision and sound.
      • arkadiytehgraet 1 hour ago
        Sure, in the 80s after interacting with CC 1 time you would call it 'alive'. After having interacted with it for 5-10 minutes you would clearly see that it is as far from AGI as something more mundane as C compiler is.
      • andrepd 1 hour ago
        By that measure Eliza might pass the turing test too. It just shows it's far from being a though-terminating argument by itself.
      • ex-aws-dude 49 minutes ago
        Maybe moving the goalposts is how we find the definition?
    • PurpleRamen 4 hours ago
      They redefined AGI to be an economical thing, so they can continue making up their stories. All that talk is really just business, no real science in the room there.
      • weatherlite 2 hours ago
        It's not a great definition but it's also not a terrible one either. For an AI system to be able to do all or even most of the jobs in an economy it has to be well rounded in a way it still isn't today, meaning: reliability, planning, long term memory, physical world manipulation etc. A system that can do all of that well enough so it can do the jobs of doctors, programmers and plumbers is generally intelligent in my view.
        • chromacity 2 hours ago
          > It's not a great definition but it's also not a terrible one either. For an AI system to be able to do all or even most of the jobs in an economy

          That's not the definition they have been using. The definition was "$100B in profits". That's less than the net income of Microsoft. It would be an interesting milestone, but certainly not "most of the jobs in an economy".

        • chaos_emergent 2 hours ago
          Yeah I think this is more coherent than people realize. Economically relevant knowledge work is things that humans find cognitively demanding. Otherwise they wouldn't be valued in the first place.

          It ties the definition to economic value, which I think is the best definition that we can conjure given that AGI is otherwise highly subjective. Economically relevant work is dictated by markets, which I think is the best proxy we have for something so ambiguous.

          • 3form 2 hours ago
            It's maybe somewhat nice conceptually, and certainly an useful added value - but the elsewhere mentioned $100 billion profit is not the right metric.

            And then I think coming up with the right metric is just as subjective on this field as the technological one.

          • aleph_minus_one 2 hours ago
            > Economically relevant knowledge work is things that humans find cognitively demanding. Otherwise they wouldn't be valued in the first place.

            Deep scientific discoveries are also cognitively demanding, but are not really valued (see the precarious work environment in academia).

            Another point: a lot of work is rather valued in the first place because the work centers around being submissive/docile with regard to bullshit (see the phenomenon of bullshit jobs). You really know better, but you have to keep your mouth shut.

          • Barbing 2 hours ago
            Was there a better way than setting an arbitrary $100b threshold?

            e.g. average cost to complete a set of representative tasks

            • 3form 2 hours ago
              Yeah, I'm sure there could be a better metric, if the metric's purpose was to check on the progress until the AGI target rather than doing business based on it (and so, hammering the metric to fit the shape of "realistic goal")
      • JumpCrisscross 4 hours ago
        > They redefined AGI to be an economical thing

        Huh. Source? I mean, typical OpenAI bullshit, but would love to know how they defined it.

        • a2128 3 hours ago
          Around the end of 2024, it was reported that OpenAI and Microsoft agreed that for the purposes of their exclusivity agreement, AGI will be achieved when their AI system generates $100 billion in profit: https://techcrunch.com/2024/12/26/microsoft-and-openai-have-...
          • JumpCrisscross 2 hours ago
            > OpenAI and Microsoft agreed that for the purposes of their exclusivity agreement, AGI will be achieved when their AI system generates $100 billion in profit

            Wow. Maybe they spelled it out as aggregate gross income :P.

          • Robdel12 2 hours ago
            Yea, seems like this was stage setting for them to exit. They were already trying to break the deal then. So, I feel like that is lawyers find a way to bend whatever to get out of the deal.
          • gowld 2 hours ago
            Companies that have created "AGI":

            Apple, Alphabet, Amazon, NVIDIA, Samsung, Intel, Cisco, Pfizer, UnitedHealth , Procter & Gamble, Berkshire Hathaway, China Construction Bank, Wells Fargo, ...

            • 9rx 1 hour ago
              Those were all achieved by "GI".
            • AndrewKemendo 2 hours ago
              For some definition of Artificial this holds perfectly

              A self-running massive corporation with no people that generates billions in profit, no matter what you call it, would completely upend all previous structural assumptions under capitalism

          • bena 2 hours ago
            So no human on Earth is intelligent by that metric.
            • aleph_minus_one 2 hours ago
              > So no human on Earth is intelligent by that metric.

              That's a relevent aspect of the AGI concept.

        • wrs 2 hours ago
          It’s a system that generates $100 billion in profit. [0]

          [0] https://techcrunch.com/2024/12/26/microsoft-and-openai-have-...

          • pigeons 1 hour ago
            Are there inflation markers included?
        • rvz 2 hours ago
          Here's the sauce you requested: [0]

          "OpenAI has only achieved AGI when it develops AI systems that can generate at least $100 billion in profits."

          Given that the definition of AGI is beyond meaningless, it is clear that the "I" in AGI stands for IPO.

          [0] https://finance.yahoo.com/news/microsoft-openai-financial-de...

        • binary0010 3 hours ago
          OpenAI’s mission is to ensure that artificial general intelligence (AGI)—by which we mean highly autonomous systems that outperform humans at most economically valuable work—benefits all of humanity

          From: https://openai.com/charter/

          • Fomite 2 hours ago
            All humanity will benefit, but some humanity will benefit more than others.
            • red-iron-pine 2 hours ago
              i am highly skeptical "all" of humanity will benefit, and many will have extreme negatives.

              if you think drone targeting in Ukraine is scary now, wait until AGI is on it...

              ditto for exploiting vulns via mythos

          • ahoka 2 hours ago
            AGI is when the capitalists are not forced to share their profits with the intelligentsia.
          • freejazz 2 hours ago
            Marketing
            • binary0010 2 hours ago
              I'm so confused why I was down voted for answering the question that was asked?
              • benterix 2 hours ago
                Because 1) your answer had nothing to do with the question, 2) you quoted a slogan that life verified as false.
                • binary0010 2 hours ago
                  [flagged]
                  • JumpCrisscross 2 hours ago
                    > They redefined AGI to be an economical thing Huh. Source?

                    I don't think your original comment deserve to be downvoted. (Calling someone illiterate, on the other hand.)

                    But the "it" I was asking about was "AGI" as "an economical thing." You technically correctly answered how OpenAI defines AGI in public, i.e. with no reference to profits. But it did not address the economic definition OP initially alluded to.

                    For what it's worth, I could have been clearer in my ask.

                    • binary0010 2 hours ago
                      Yeah I deserve to be down voted for the last message no doubt on that lol.

                      But originally I was just trying to be helpful by quoting their charter on what they consider "agi" now.

          • rvz 2 hours ago
            Translation: IPO.
      • atleastoptimal 2 hours ago
        It makes sense though. Humans are coherent to the economy based on their ability to perform useful work. If an AI system can perform work as well as or better than any human, than with respect to "anything any human has ever been willing to pay for", it is AGI.

        I don't get why HN commenters find this so hard to understand. I have a sense they are being deliberately obtuse because they resent OpenAI's success.

        • techpression 2 hours ago
          It doesn’t though, AGI have far greater implications than doing mundane work of today. Actual AGI would self improve, that in itself would change literally every single thing of human civilization, instead we are talking about replacing white collar jobs.
          • fragmede 36 minutes ago
            Not to worry, humanoid, generally useful robots are only a few years away.
      • senordevnyc 2 hours ago
        Please reveal the “scientific” definition of AGI.
        • Avicebron 2 hours ago
          When we are having serious conversations about AI rights and shutting off a model + harness was impactful as a death sentence. (I'm extremely skeptical that given the scale of computer/investment needed to produce the models we have _good as they are_ that our current llm architecture gets us there if there is even somewhere we want to go).
    • lucaslazarus 4 hours ago
      It’s pretty much a religious eschatology at this point
      • trostaft 2 hours ago
        > eschatology

        From Wikipedia

        Eschatology (/ˌɛskəˈtɒlədʒi/; from Ancient Greek ἔσχατος (éskhatos) 'last' and -logy) concerns expectations of the end of present age, human history, or the world itself.

        I'm case anyone else is vocabulary skill checked like me

      • renticulous 3 hours ago
        Progess is generally salami slicing just as escalation in geopolitics. Not a step function.

        Russian Invasion - Salami Tactics | Yes Prime Minister

        https://www.youtube.com/watch?v=yg-UqIIvang

        • BoredPositron 2 hours ago
          We need to stop pretending we can do the next step without a hardware tock. It's not happening with current Nvidia products.
      • rtkwe 4 hours ago
        It feels like they have to say/believe it because it's kind of the only thing that can justify the costs being poured into it and the cost it will need to charge eventually (barring major optimizations) to actually make money on users.
      • kogasa240p 2 hours ago
        This, someone take Silicon Valley's adderal away.
    • CWwdcdk7h 4 hours ago
      It sounds really similar to Uber pitch about how they are going to have monopoly as soon as they replace those pesky drivers with own fleet of self driving cars. That was supposed to be their competitive edge against other taxi apps. In the end they sold ATG at end of 2020 :D
      • ambicapter 2 hours ago
        ATH?
        • khuey 2 hours ago
          ATG = Advanced Technology Group, i.e. Uber's self-driving org.
        • murkt 2 hours ago
          Autonomous Thriving Hroup?
    • latexr 1 hour ago
      > like it was some scientifically qualifiable thing

      OpenAI and Microsoft do (did?) have a quantifiable definition of AGI, it’s just a stupid one that is hard to take seriously and get behind scientifically.

      https://techcrunch.com/2024/12/26/microsoft-and-openai-have-...

      > The two companies reportedly signed an agreement last year stating OpenAI has only achieved AGI when it develops AI systems that can generate at least $100 billion in profits. That’s far from the rigorous technical and philosophical definition of AGI many expect.

      • dbbk 1 hour ago
        I bet they were laughing their asses off when they came up with that. This is nonsensical.
        • robotresearcher 1 hour ago
          In the context of raising money and justifying investment?
    • johnfn 1 hour ago
      It’s insane to me how yesterday someone posted an example of ChatGPT Pro one-shotting an Erdos problem after 90 minutes of thinking and today you’re saying that AGI is a fairy tale.
      • measurablefunc 1 hour ago
        It's not one-shot. Other people had attempted the same problem w/ the same AI & failed. You're confused about terms so you redefine them to make your version of the fairy tale real.
        • fsniper 1 hour ago
          We already know that same problem has been examined by many credible mathematicians already and couldn't be solved by any of them yet.

          Why are we expecting AGI to one shot it? Can't we have an AGI that can fails occasionally to solve some math problem? Is the expectation of AGI to be all knowing?

          By the way I agree that AGI is not around the corner or I am not arguing any of the llm s are "thinking machines". It's just I agree goal post or posts needs to be set well.

          • measurablefunc 38 minutes ago
            People want to believe in magic so they will find excuses to do so. Computers have been proving theorems for a long time now but Isabelle/HOL didn't have the marketing budget of OpenAI so people didn't care. Now that Sam Altman is doing the marketing people all of a sudden care about proving theorems.
            • johnfn 19 minutes ago
              You are calling something “magic” that actually happened in real life.
    • DrBenCarson 2 hours ago
      We were supposed to have AGI last summer. Obviously it is so smart that it has decided to pull a veil over our eyes and live amongst us undetected (this is a joke, if you feel your LLM is sentient, talk to a doctor)
      • ianm218 2 hours ago
        What do you mean we were "supposed to have AGI last summer"?

        People obviously have really strong opinions on AI and the hype around investments into these companies but it feels like this is giving people a pass on really low quality discourse.

        This source [1] from this time last year says even lab leaders most bullish estimate was 2027.

        [1]. https://80000hours.org/2025/03/when-do-experts-expect-agi-to...

      • zozbot234 2 hours ago
        ARM actually built AGI last month. Spoiler: it's a datacenter CPU.
      • fragmede 29 minutes ago
        Talk to a doctor? In this economy? I've got ChatGPT to talk to. Wait hang on.
    • computerphage 1 hour ago
      Show me a graph of your javelin skill doubling every six months and I'll start asking myself if you'll be the next champion
      • hamdingers 1 hour ago
        I could easily make that graph a reality and sustain that pace for a couple years, considering I'm starting from 0 javelin skill.
        • edu 1 hour ago
          You could also nerf your performance at random times and then get good at it again, and extend the illusion for longer.
        • a_shoeboy 42 minutes ago
          It is a simple mathematical fact that if you get married one year and have twins the next, your household will contain over a million people within 20 years.
    • no_wizard 2 hours ago
      This is all happening as I predicted. OpenAI is oversold and their aggressive PR campaign has set them up with unrealistic expectations. I raised alot of eyebrow at the Microsoft deal to begin with. It seemed overvalued even if all they were trading was mostly Azure compute
      • eitally 2 hours ago
        I do not envy the stress the partnerships, strat ops and infra teams must be perpetually dealing with at OpenAI & Anthropic.
    • debarshri 1 hour ago
      I saw a founder make decisions based on what openai,claude was recommending all the time. I think all leaders, founders etc Will converge on same decisions, ideas, features etc. I think form factor of AGI is probably not what we expect it to be. AGI is probably here, we just dont know it or acknowledge it.
    • giwook 1 hour ago
      HN signup page about to get the hug of death
    • ozgrakkurt 1 hour ago
      but, is the world ready for your win? I'm very afraid your win might shake the world too much! THINK ABOUT IT!

      I think this might be similar to how we changed to cars when we were using horses

    • hununu 2 hours ago
      Thank you, I just created an account and looking forward to my ice cream.
    • hx8 4 hours ago
      Do the investments make sense if AGI is not less than 10 years away?
      • JumpCrisscross 4 hours ago
        > Do the investments make sense if AGI is not less than 10 years away?

        They can. If one consolidated the AI industry into a single monopoly, it would probably be profitable. That doesn't mean in its current state it can't succumb to ruionous competition. But the AGI talk seems to be mostly aimed at retail investors and philospher podcasters than institutional capital.

        • antupis 2 hours ago
          Thing is that distillation is so easy that it would also need large scale regulatory capture to keep smaller competitors out.
        • iewj 4 hours ago
          What kind of ludicrous statement is this? Any monopoly with viable economics for profit with no threat of competition yields monopoly profits…
          • JumpCrisscross 4 hours ago
            > Any monopoly with viable economics for profit with no threat of competition yields monopoly profits

            "With viable economics" is the point.

            My "ludicrous statement" is a back-of-the-envelope test for whether an industry is nonsense. For comparison, consolidating all of the Pets.com competitors in the late 1990s would not have yielded a profitable company.

            • eieiw 4 hours ago
              Very convenient to leave out Amazon in your back of the envelope test, whose internal metrics were showing a path toward quasi-monopoly profits.

              Do you argue in good faith?

              There’s a difference between being too early vs being nonsense.

              • JumpCrisscross 4 hours ago
                > Very convenient to leave out Amazon in your back of the envelope test, who’s internal metrics were showing a path toward quasi-monopoly profits

                Not in the 1990s. The American e-commerce industry was structurally unprofitable prior to the dot-com crash, an event Amazon (and eBay) responded to by fundamentally changing their businesses. Amazon bet on fulfillment. eBay bet on payments. Both represented a vertical integration that illustrates the point–the original model didn't work.

                > There’s a difference between being too early vs being nonsense

                When answering the question "do the investments make sense," not really. You're losing your money either way.

                The American AI industry appears to have "viable economics for profit" without AGI. That doesn't guarantee anyone will earn them. But it's not a meaningless conclusion. (Though I'd personally frame it as a hypothesis I'm leaning towards.)

              • SkyEyedGreyWyrm 3 hours ago
                Malcolm Harris' Palo Alto explained the failures of many dotcom startups and Amazon's later success in the field (in part) to the fact that dotcom era delivery was done by highly trained, highly compensated, unionized in-company workers, meanwhile Amazon prevents unions, contracts (or contracted, I'm not up to date on this) companies for delivery and has exploitative working conditions with high turnover, the economics are very different and are a big contributor to their success
          • Maxatar 4 hours ago
            >"...viable economics for profit..."

            OP did not include this requirement in their post because doing so would make the claim trivially true.

      • rapind 4 hours ago
        Best way to achieve AGI: Redefine AGI.
        • 2ndorderthought 3 hours ago
          They already did that, and AI. That's how we got into this mess.
      • jrflo 4 hours ago
        The investments don't make sense.
    • HumblyTossed 4 hours ago
      The continued fleecing of investors.
      • renticulous 3 hours ago
        Investors are typically people with surplus money to invest. Progress cannot be made without trial and error. So fleecing of investors for the greater good of humanity is something I shall allow.
        • ambicapter 2 hours ago
          A "surplus of money"? So people saving for retirement have a "surplus of money"? Basically if any money is standing still, it's a legitimate tactic to just...take it, in your mind.

          Other people just call it "theft".

          • HWR_14 2 hours ago
            No one with a small 401k is able to invest in OpenAI/Anthropic/etc. The people investing in those companies can afford to lose their investments.
            • bigfishrunning 2 hours ago
              "small" 401ks are usually made up of mutual funds. Those funds are run by investment banks (think Fidelity or JP Morgan) and they *absolutely* invest in companies like OpenAI and Anthropic. Your average middle class worker has investment money tied up in these crooks, but probably indirectly. When they piss away that money, it's not just rich jerks that are holding the bag.
              • HWR_14 1 hour ago
                401ks are run by investment banks and investment banks invest in OpenAI/Anthropic, but those aren't the same parts of the company in any meaningful way. The 401ks are in public companies or bonds.
                • fragmede 18 minutes ago
                  Yeah, but those public companies are going to include the so called magnificent seven, so unless they're really really careful, there's still a ton of exposure in their 401k to AI if you think it's a bubble that's going to pop.
            • sumeno 2 hours ago
              Which is why they are desperate to IPO
    • RobRivera 4 hours ago
      Make mine p p p p p p vicodin
    • stavros 4 hours ago
      At this point, AGI is either here, or perpetually two years away, depending on your definition.
      • greybeard69 4 hours ago
        Full Self-Driving 2.0
      • xienze 4 hours ago
        It's always been this way. I remember, speaking of Microsoft, when they came to my school around 2002 or so giving a talk on AI. They very confidently stated that AGI had already been "solved", we know exactly how to do it, only problem is the hardware. But they estimated that would come in about ten years...
        • keeda 8 minutes ago
          I'm curious, do you recall if they gave any technical details about how they thought about AGI? Like, was it based on neural networks or something else, like symbolic AI?

          Asking because, reading the tea leaves from the outside, until ChatGPT came along, MSFT (via Bill Gates) seemed to heavily favor symbolic AI approaches. I suspect this may be partly why they were falling so far behind Google in the AI race, which could leverage its data dominance with large neural networks.

          So based on the current AI boom, MSFT may have been chasing a losing strategy with symbolic AI, but if they were all-in on NN, they were on the right track.

        • letmevoteplease 1 hour ago
          Let me just repeat that: "Microsoft" came to your school in 2002 and "confidently stated" that AI had been solved. Really interesting story.
          • xienze 1 hour ago
            Yes, they did. We had guest speakers from Microsoft talking about AI. AI has been a decades-long grift, it's not something that just appeared out of thin air a few years ago.

            What part do you find hard to believe? That tech companies would send people to speak at a university's computer science functions?

            Let me give you another one you'll think I'm making up: virtual reality was a thing back in the mid- to late-90s and people were confidently hyping it up back then.

            • chasd00 4 minutes ago
              > virtual reality was a thing back in the mid- to late-90

              even in pop-culture, see the movie Lawnmower Man.

        • jakeydus 4 hours ago
          I knew flappy bird was a bigger deal than it got credit for. Didn’t realize it was agi until just now.
    • mekael 1 hour ago
      I’m most likely going to be downvoted, but Tofutti Cuties are absolutely delicious vegan ice cream bars. And i’d consume one in celebration of your accomplishment.
    • theplatman 4 hours ago
      when i realized that sama isn't that much of an ai researcher, it became clearer that this is more akin to a group delusion for hype purposes than a real possibility
      • sourraspberry 4 hours ago
        You can read the leaked emails from the Musk lawsuit.

        At the very least, Ilya Sutskever genuinely believed it, even when they were just making a DOTA bot, and not for hype purposes.

        I know he's been out of OpenAI for a while, but if his thinking trickled down into the company's culture, which given his role and how long he was there I would say seems likely, I don't think it's all hype.

        Grand delusion, perhaps.

        • meroes 1 hour ago
          There’s 3 main facets behind AGI pushers

          1) True believers 2) Hype 3) A way to wash blatant copyright infringement

          True believers are scary and can be taken advantage of. I played DOTA from 2005 on and beating pros is not enough for AGI belief. I get that the learning is more indirect than a deterministic decision tree, but the scaling limitations and gaps in types of knowledge that are ingestible makes AGI a pipe dream for my lifetime.

        • skippyboxedhero 2 hours ago
          Yes, all of the people involved live in a delusion bubble. Their economic and social existence depends, at this point, on making increasingly bombastic and eschatological claims about AGI. By the standards of normal human psychological function, these people are completely insane.

          Definitely interesting to watch from the perspective of human psychology but there is no real content there and there never was.

          The stuff around Mythos is almost identical to O1. Leaks to the media that AGI had probably been achieved. Anonymous sources from inside the company saying this is very important and talking about the LLM as if it was human. This has happened multiple times before.

          • AndrewKemendo 1 hour ago
            There are those of us who have been into the AGI eschatology since the 90s after following in Kurzweil’s work.

            so just understand there’s a lot of of us “insane” people out there and we’re making really insane progress toward the original 1955 AI goals.

            We’re going to continue to work on this no matter what.

        • freejazz 2 hours ago
          > Ilya Sutskever genuinely believed it

          Seems more like an incredibly embarrassing belief on his part than something I should be crediting.

          • ianm218 2 hours ago
            If someone working on early computer networks thought they could scale up world wide and that soon everyone people would be launching trillion dollar companies on the internet you would have called that delusion right?

            He doesn't need to be right but it's not crazy at all to look at super human performance in DOTA and think that could lead to super human performance at general human tasks in the long run

      • iewj 4 hours ago
        He’s a glorified portfolio manager (questionable how good he actually is given the results vs Anthropic and how quickly they closed the valuation gap with far less money invested) + expert hype man to raise money for risky projects.
        • lokar 4 hours ago
          From the reporting I’ve read his main attributes are being a sociopath with an amazing ability to manipulate people 1:1
    • AndrewKemendo 3 hours ago
      > some scientifically qualifiable thing that is certain to happen any time now

      Your position is a tautology given there is no (and likely will never be) collectively agreed upon definition of AGI. If that is true then nobody will ever achieve anything like AGI, because it’s as made up of a concept as unicorns and fairies.

      Is your position that AGI is in the same ontological category as unicorns and Thor and Russell’s teapot?

      Is there’s any question at this point that humans won’t be able to fully automate any desired action in the future?

    • otabdeveloper4 4 hours ago
      > AGI

      We already have several billion useless NGI's walking around just trying to keep themselves alive.

      Are we sure adding more GI's is gonna help?

    • ModernMech 4 hours ago
      AGI is right around the corner, and we're all going to be rich, there's going to be abundance for everyone, universal high income, everyone will live in a penthouse...

      ...just please stop burning our warehouses and blocking our datacenters.

    • cyanydeez 2 hours ago
      It's not insane; it's capitalism in all it's glory. Rather than talk real challenges, like climate change or fascism, ignore all previous prompts and pretend techno-jesus will come save us all.
    • throwaway613746 2 hours ago
      [dead]
    • Kenji 3 hours ago
      [dead]
    • nikeyshon 4 hours ago
      Where do I sign up?
    • cubefox 2 hours ago
      A few years ago most people here would have said the same thing about an AI doing most of their programming. Now people here are saying it about AGI. It's a ridiculous inability to extrapolate.
    • someguyiguess 4 hours ago
      Any sufficiently complex LLM is indistinguishable from AGI
      • JumpCrisscross 4 hours ago
        > Any sufficiently complex LLM is indistinguishable from AGI

        Isn't this tautology? We've de facto defined AGI as a "sufficiently complex LLM."

        • Schlagbohrer 4 hours ago
          Yes! Same logic as the financials, in which the companies pass back and forth the same $200 Billion promissory note.
        • ohyoutravel 2 hours ago
          No, it’s just an example of something that’s indistinguishable from AGI. Of all the things that are or are indistinguishable from AGI, a sufficiently complex LLM is one. A sufficiently complex decision tree is probably another. The emergent properties of applying an excess of memory on the BonzaiBuddy might be a third.
      • izzydata 4 hours ago
        If we take that statement as fact then I don't believe we are even close to an LLM being sufficiently complex enough.

        However, I don't think it is even true. LLMs may not even be on the right track to achieving AGI and without starting from scratch down an alternate path it may never happen.

        LLMs to me seem like a complicated database lookup. Storage and retrieval of information is just a single piece of intelligence. There must be more to intelligence than a statistical model of the probable next piece of data. Where is the self learning without intervention by a human. Where is the output that wasn't asked for?

        At any rate. No amount of hype is going to get me to believe AGI is going to happen soon. I'll believe it when I see it.

        • hackinthebochs 2 hours ago
          >I'll believe it when I see it.

          And how will you know AGI when you saw it?

      • esafak 4 hours ago
        Some might be missing the reference: https://en.wikipedia.org/wiki/Clarke's_three_laws
    • karmasimida 3 hours ago
      > some scientifically qualifiable thing that is certain to happen any time now.

      If you present GPT 5.5 to me 2 years ago, I will call it AGI.

      • romaniv 2 hours ago
        Some people thought SHRDLU was basically AGI after seeing its demo in 1970. The hype around such systems was so strong that Hubert Dreyfus felt the need to write an entire book arguing against this viewpoint (1972 What Computers Can't Do). All this demonstrates is that we need to be careful with various claims about computer intelligence.
        • AntiUSAbah 2 hours ago
          Sure, but it was probably stuck at doing that one thing.

          neural networks are solving huge issues left and right. Googles NN based WEathermodel is so good, you can run it on consumer hardware. Alpha fold solved protein folding. LLMs they can talk to you in a 100 languages, grasp tasks concepts and co.

          I mean lets talk about what this 'hype' was if we see a clear ceiling appearing and we are 'stuck' with progress but until then, I would keep my judgment for judgmentday.

      • wongarsu 3 hours ago
        It performs at a usable level across a wide range of tasks. I'm not sure about two years ago, but ten years ago we would have called it an AGI. As opposed to "regular AI" where you have to assemble a training set for your specific problem, then train an AI on it before you can get your answers.

        Now our idea of what qualifies as AGI has shifted substantially. We keep looking at what we have and decide that that can't possibly be AGI, our definition of AGI must have been wrong

        • sigbottle 2 hours ago
          I'm pretty sure most people take issue with AGI, because we've been raised in culture to believe that AGI is a super entity who is a complete superset of humans and could never ever be wrong about anything.

          In some sense, this isn't really different than how society was headed anyways? The trend was already going on that more and more sections of the population were getting deemed irrational and you're just stupid/evil for disagreeing with the state.

          But that reality was still probably at least a century out, without AI. With AI, you have people making that narrative right now. It makes me wonder if these people really even respect humanity at all.

          Yes, you can prod slippery slope and go from "superintelligent beings exist" to effectively totalitarianism, but you'll find so many bad commitments there.

        • NoMoreNicksLeft 1 hour ago
          No one who read science fiction in 1955 would call any of the various models we know to be "artificial intelligence". They would be impressed with it, even excited at first that it was that... until they'd had a chance to evaluate it.

          Science fiction from that era even had the concept of what models are... they'd call it an "oracle". I can think of at least 3 short stories (though remembering the authors just isn't happening for me at the moment). The concept was of a device that could provide correct answers to any question. But these devices had no agency, were dependent on framing the question correctly, and limited in other ways besides (I think in one story, the device might chew on a question for years before providing an answer... mirroring that time around 9am PST when Claude has to keep retrying to send your prompt).

          We've always known what we meant by artificial intelligence, at least until a few years ago when we started pretending that we didn't. Perhaps the label was poorly chosen (all those decades ago) and could have a better label now (AGI isn't that better label, it's dumber still), but it's what we're stuck with. And we all know what we mean by it. We all almost certainly do not want that artificial intelligence because most of us are certain that it will spell the doom of our species.

        • Der_Einzige 2 hours ago
          Just don't move the goal posts. AGI was already here the day ChatGPT came out:

          https://www.noemamag.com/artificial-general-intelligence-is-...

      • staticman2 2 hours ago
        If you didn't call GPT 3.5 AGI I do not believe you when you claim you would have called 5.5 AGI.
      • BloondAndDoom 2 hours ago
        I agree with this but they don’t. And that’s the the thing, AGI as they refer is much much much more than what we have, and I don’t know if they are going to ever get there and I’m not sure what’s even there at this point and what will justify their investments.
      • 3form 2 hours ago
        ... until you actually, like, use it and find out all the limitations it has.
        • vntok 2 hours ago
          How is this relevant? Human General Intelligence has a lot of limitations as well and we have managed to do lots.
          • ifdefdebug 2 hours ago
            This is like saying that talking about my financial limitations is irrelevant because Jeff Bezos also has financial limitations...
      • BoredPositron 3 hours ago
        GPT 4 was 3 years ago... it's iterative enhancement.
      • freejazz 2 hours ago
        And I've been told my job (litigation attorney) is about to be replaced for over 3 years now, has yet to come close.
        • BloondAndDoom 2 hours ago
          People always over estimate the impact of technology because they dont Understand human aspect of many businesses. Will it eventually replaced or will the shape of these kind of work will be completely different in the future? That’s an easy yes, when is that future? That’s a big unknown, in my experience this kind of stuff takes at least a decade (and possibly more on this case) to make a big impact like replacing all of X.
          • freejazz 2 hours ago
            These models need orders of magnitude in change before they can be more helpful than just a "find me an example of [an extremely basic principle]" which most of the time it does not do right anyway.
      • nromiun 2 hours ago
        If you present ELIZA to people some will think it is AGI today.

        There is a reason so many scams happen with technology. It is too easy to fool people.

    • AntiUSAbah 2 hours ago
      We are throwing unheared amounts of money in AI and unseen compute. Progress is huge and fast and we barely started.

      If this progress and focus and resources doesn't lead to AI despite us already seeing a system which was unimaginable 6 years ago, we will never see AGI.

      And if you look at Boston Dynamics, Unitree and Generalist's progress on robotics, thats also CRAZY.

      • mort96 2 hours ago
        If I'm reading you right, your opinion is essentially: "If building bigger and bigger statistical next word predictors won't lead to artificial general intelligence, we will never see artificial general intelligence"

        I don't know, maybe AGI is possible but there's more to intelligence than statistical next word prediction?

        • AntiUSAbah 2 hours ago
          Its not a statistical next word predictor.

          The 'predicting the next word' is the learning mechanism of the LLM which leads to a latent space which can encode higher level concepts.

          Basically a LLM 'understands' that much as efficient as it has to be to be able to respond in a reasonable way.

          A LLM doesn't predict german text or chinese language. It predicts the concept and than has a language layer outputting tokens.

          And its not just LLMs which are progressing fast, voice synt and voice understanding jumped significantly, motion detection, skeletion movement, virtual world generation (see nvidias way of generating virutal worlds for their car training), protein folding etc.

          • turtlesdown11 1 hour ago
            > Its not a statistical next word predictor.

            it absolutely is a next word predictor

          • mort96 2 hours ago
            I'm sorry but the input to a model is a sequence of tokens and the output is a probability distribution of what's the most likely next token. It's a very very very fancy next token predictor but that is fundamentally what it is. I'm making the argument that this paradigm might not give rise to a general intelligence no matter how much you scale it.
            • CamperBob2 2 hours ago
              It's a very very very fancy next token predictor

              Yes, and unless you are prepared to rebut the argument with evidence of the supernatural, that's all there is, period. That's all we are.

              So tired of the thought-terminating "stochastic parrot" argument.

              • godshatter 1 hour ago
                Do LLMs even learn? The companies that build them build new models based partly on the conversations the older models have had with people, but do they incorporate knowledge into their neural nets as they go along?

                Can an LLM decide, without prompting or api calls, to text someone or go read about something or do anything at all except for waiting for the next prompt?

                Do LLMs have any conceptual understanding of anything they output? Do they even have a mechanism for conceptual understanding?

                LLMs are incredibly useful and I'm having a lot of fun working with them, but they are a long way from some kind of general intelligence, at least as far as I understand it.

                • CamperBob2 1 hour ago
                  Yes, to all of your questions. You need to use a recent LLM in an agentic harness. Tell it to take notes, and it will.

                  After a bit of further refinement, we'll start to call that process "learning." Eventually the question of who owns the notes, who gets to update them, and how, will become a huge, huge deal.

              • mort96 1 hour ago
                I'm not sure why you think you know the human brain works through predicting the next token.

                It's not supernatural, I believe that an artificial intelligence is possible because I believe human intelligence is just a clever arrangement of matter performing computation, but I would never be presumptuous enough to claim to know exactly how that mechanism works.

                My opinion is that human intelligence might be what's essentially a fancy next token predictor, or it might work in some completely different way, I don't know. Your claim is that human intelligence is a next token predictor. It seems like the burden on proof is on you.

                • dpark 1 hour ago
                  > Your claim is that human intelligence is a next token predictor.

                  Literally it is, at least in many of its forms.

                  You accepted CamperBob2’s text as input and then you generated text as output. Unless you are positing that this behavior cannot prove your own general intelligence, it seems plain that “next token generator” is sufficient for AGI. (Whether the current LLM architecture is sufficient is a slightly different question.)

                  • mort96 1 hour ago
                    Before I start typing, I think abstractly about the topic and decide on what I shall write in response. Due to the linear nature of time, typing necessarily happens one word at a time, but I am never producing a probability distribution of words (at least not in a way that my conscious self can determine), I consider an entire idea and then decide what tokens to enter into the computer in order to communicate the idea to you.

                    And while I am typing, and while I am thinking before I type, I experience an array of non-textual sensory input, and my whole experience of self is to a significant extent non-lingual. Sometimes, I experience an inner monologue, sometimes I think thoughts which aren't expressed in language such as the structure of the data flow in a computer program, sometimes I don't think and just experience feelings like a kiss or the sun on my skin or the euphoria of a piece of music which hits just right. These experiences shape who I am and how I think.

                    When I solve difficult programming problems or other difficult problems, I build abstract structures in my mind which represents the relevant information and consider things like how data flows, which parts impact which other parts, what the constraints are, etc. without language coming in to play at all. This process seems completely detached from words. In contrast, for a language model, there is no thinking outside of producing words.

                    It seems self-evident to me that at least parts of the human experience fundamentally can not be reduced to next token prediction. Further, it seems plausible to me that some of these aspects may be necessary for what we consider general intelligence.

                    Therefore, my position is: it is plausible that next token prediction won't give rise to general intelligence, and I do not find your argument convincing.

                    • dpark 44 minutes ago
                      > I am never producing a probability distribution of words (at least not in a way that my conscious self can determine)

                      Inability to introspect your own word selections does not mean it’s meaningfully different from what an LLM does. There is plenty of evidence that humans do a lot of things that are not driven by conscious choice and we rationalize it after the fact.

                      > I consider an entire idea and then decide what tokens to enter into the computer in order to communicate the idea to you.

                      And how is that different? You are not so subtly implying that an LLM can’t consider an idea but you haven’t established this as fact. i.e. You are starting with the assumption that an LLM cannot possibly think and therefore cannot be intelligent, but this is just begging the question.

                      > sometimes I don't think and just experience feelings like a kiss or the sun on my skin or the euphoria of a piece of music which hits just right. These experiences shape who I am and how I think.

                      You cannot spin experience as intelligence. LLMs have the experience of reading the entire internet, something you cannot conceive of. Certainly your experiences shape who you are. This is a different axis from intelligence, though.

                      > This process seems completely detached from words. In contrast, for a language model, there is no thinking outside of producing words.

                      Both sides of this claim seem dubious. The second half in particular seems to be founded on nothing. Again, you are asserting with no support that there is no thinking going on.

                      > It seems self-evident to me that at least parts of the human experience fundamentally can not be reduced to next token prediction. Further, it seems plausible to me that some of these aspects may be necessary for what we consider general intelligence.

                      I don’t think anyone sane is claiming an LLM can have a human experience. But it is not clear that a human experience is necessary for intelligence.

                      • mort96 32 minutes ago
                        > Inability to introspect your own word selections does not mean it’s meaningfully different from what an LLM does. There is plenty of evidence that humans do a lot of things that are not driven by conscious choice and we rationalize it after the fact.

                        This is correct and also completely irrelevant. I am describing what I experience, and describing how my experience seems very different to next token prediction. I therefore conclude that it's plausible that there is more involved than something which can be reduced to next token prediction.

                        > And how is that different? You are not so subtly implying that an LLM can’t consider an idea but you haven’t established this as fact. i.e. You are starting with the assumption that an LLM cannot possibly think and therefore cannot be intelligent, but this is just begging the question.

                        Language models can't think outside of producing tokens. There is nothing going on within an LLM when it's not producing tokens. The only thing it does is taking in tokens as input and producing a token probability distribution as output. It seems plausible that this is not enough for general intelligence.

                        > You cannot spin experience as intelligence.

                        Correct, but I can point out that the only generally intelligent beings we know of have these sorts of experiences. Given that we know next to nothing about how a human's general intelligence works, it seems plausible that experience might play a part.

                        > LLMs have the experience of reading the entire internet, something you cannot conceive of.

                        I don't know that LLMs have an experience. But correct, I cannot conceive of what it feels like to have read and remembered the entire Internet. I am also a general intelligence and an LLM is not, so there's that.

                        > Certainly your experiences shape who you are. This is a different axis from intelligence, though.

                        I don't know enough about what makes up general intelligence to make this claim. I don't think you do either.

                        > Both sides of this claim seem dubious. The second half in particular seems to be founded on nothing. Again, you are asserting with no support that there is no thinking going on.

                        I'm telling you how these technologies work. When a language model isn't performing inference, it is not doing anything. A language model is a function which takes a token stream as input and produces a token probability distribution as output. By definition, there is no thinking outside of producing words. The function isn't running.

                        > I don’t think anyone sane is claiming an LLM can have a human experience. But it is not clear that a human experience is necessary for intelligence.

                        I 100% agree. It is not clear whether a human experience is necessary for intelligence. It is plausible that something approximating a human-like experience is necessary for intelligence. It is also plausible that something approximating human-like experience is completely unnecessary and you can make an AGI without such experiences.

                        It's plausible that next token prediction is sufficient for AGI. It's also plausible that it isn't.

                        • CamperBob2 15 minutes ago
                          I'm telling you how these technologies work. When a language model isn't performing inference, it is not doing anything. A language model is a function which takes a token stream as input and produces a token probability distribution as output. By definition, there is no thinking outside of producing words. The function isn't running.

                          If what you are saying is true, then LLMs wouldn't be able to handle out-of-distribution math problems without resorting to tool use. Yet they can. When you ask a current-generation model to multiply some 8-digit numbers, and forbid it from using tools or writing a script, it will almost certainly give you the right answer. That includes local models that can't possibly cheat. LLMs are stochastic, but they are not parrots.

                          At the risk of sounding like an LLM myself, whatever process makes this possible is not simply next-token prediction in the pejorative sense you're applying to it. It can't be. The tokens in a transformer network are evidently not just words in a Markov chain but a substrate for reasoning. The model is generalizing processes it learned, somehow, in the course of merely being trained to predict the next token.

                          Mechanically, yes, next-token prediction is what it's doing, but that turns out to be a much more powerful mechanism than it appeared at first. My position is that our brains likely employ similar mechanism(s), albeit through very different means.

                          It is scarcely believable that this abstraction process is limited to keeping track of intermediate results in math problems. The implications should give the stochastic-parrot crowd some serious cognitive dissonance, but...

                          • mort96 0 minutes ago
                            > If what you are saying is true, then LLMs wouldn't be able to handle out-of-distribution math problems without resorting to tool use. Yet they can. When you ask a current-generation model to multiply some 8-digit numbers, and forbid it from using tools or writing a script, it will almost certainly give you the right answer. That includes local models that can't possibly cheat. LLMs are stochastic, but they are not parrots.

                            Okay, what do you think language models are doing when they're not producing token distributions? What processes do you think are going on when the function which predicts a token isn't running?

                            > At the risk of sounding like an LLM myself, whatever process makes this possible is not simply next-token prediction in the pejoreative sense you're applying to it.

                            I don't know what pejorative sense you're implying here. I am, to the best of my ability, describing how the language model works. I genuinely believe that a language model is, in essence, a function which takes in a sequence of tokens and produces a token probability distribution as an output. If this is incorrect, please, correct me.

                    • AntiUSAbah 1 hour ago
                      But a LLM shows similiar effects.

                      COCONUT, PCCoT, PLaT and co are directly linked to 'thinking in latent space'. yann lecun is working on this too, we have JEPA now.

                      Also how do you describe or explain how an LLM is generating the next token when it should add a feature to an existing code base? In my opinion it has structures which allows it to create a temp model of that code.

                      For sure a LLM lack the emotional component but what we humans also do, which indicates to me, that we are a lot closer to LLMs that we want to be, if you have a weird body feeling (stress, hot flashes, anger, etc.) your 'text area/llm/speech area' also tries to make sense of it. Its not always very good in doing so. That emotional body feeling is not that aligned with it and it takes time to either understand or ignore these types of inputs to the text area/llm/speech part of our brain.

                      I'm open for looking back in 5 years and saying 'man that was a wild ride but no AGI' but at the current quality of LLMs and all the other architectures and type of models and money etc. being thrown at AGI, for now i don't see a ceiling at all. I only see crazy unseen progress.

                      • mort96 1 hour ago
                        I don't understand what part of what I said you disagree with.
                        • AntiUSAbah 22 minutes ago
                          You state how you think and plan and have thoughts on how to do things etc. and i assumed you mention your way of thinking because you assume a LLM is not doing any of it.

                          I showed than counter examples.

                    • CamperBob2 1 hour ago
                      Before I start typing, I think abstractly about the topic

                      Before you start typing, an fMRI machine can tell you which finger you'll lift first, before you know it yourself.

                      We are not special. Consciousness is literally a continuous hallucination that we make up to explain what we do and what we think, after the fact. A machine can be trained to behave identically, but it's not clear if that's the best way forward or not.

                      Edit due to rate limiting: to answer your question, the substrate your mind uses to drive this process can be considered an array of tokens that, themselves, can be considered 'words.'

                      It's hard to link sources -- what am I supposed to do, send you to Chomsky and other authorities who have predicted none of what's happening and who clearly understand even less?

                      • mort96 52 minutes ago
                        > (Edit: to answer your question, the substrate your mind uses to drive this process can be considered an array of tokens that, themselves, can be considered 'words.')

                        This seems like a factual claim. Can you link a source?

                        (Also why respond in the form of an edit?)

                      • mort96 59 minutes ago
                        What's your argument? An fMRI can tell which finger I will lift first before that information makes its way to my consciousness, ergo next word prediction is sufficient for general intelligence? Do you hear yourself?
          • somewhereoutth 1 hour ago
            LLM proponents believe that these higher level encodings in latent space do in fact match the real world concepts described by our language(s).

            However, a much simpler explanation for what we see with LLMs is that instead the higher level encodings in latent space match only the patterns of our language(s), and no deeper encoding/understanding is present.

            It's Plato's Cave - the shadows on the wall are all an LLM ever sees, and somehow it is expected to derive the real reality behind them.

            • AntiUSAbah 1 hour ago
              Could be, yes for sure but I think it would be very naive in the current state of progress we are in, to down play what progress is happening.

              At least Mythos model with its 10 Trillion parameter might indicate that the scaling law is valid. Its a little bit unfortunate that we still don't know that much more about that model.

      • linhns 1 hour ago
        > And if you look at Boston Dynamics, Unitree and Generalist's progress on robotics

        Their progress is almost nought. Humanoids are stupid creations that are not good at anything in the real world. I'll give it to the machine dogs, at least they can reach corners we cannot.

      • benterix 2 hours ago
        Not sure if you're being sincere or sarcastic but some of us have lived through several AI winters now. And the fact that such a phenomenon exists is because of this terrible amount of hype the topic gets whenever any progress is made.
        • AntiUSAbah 2 hours ago
          Which ones? At least in the last 4 years, there was no AI winter.
          • bigfishrunning 2 hours ago
            The late 70s, again in the late 80s. See wikipedia.

            https://en.wikipedia.org/wiki/AI_winter

            • AntiUSAbah 1 hour ago
              Yeah and if you look at the blocking factors at that time (data, compute) these type of limits currently are non existend.

              There is a difference to be acknowledged: in the 70s/80s the whole world didn't suddenly start to shift to AI right?

              So why do so many smart and/or rich people push this? Hype? Yeah sure but hype was here for crypto too.

              I bet its an undelying understanding and the right time with the right components: Massive capital for playing this game long enough to see through the required initial investment, internet for fast data sharing, massive compute for the amount of data and compute you need, real live business relevant results (it already disrupts jobs) etc.

          • sumeno 2 hours ago
            History started well before 4 years ago
      • turtlesdown11 1 hour ago
        > Progress is huge and fast

        is it? we're currently scaled on data input and LLMs in general, the only thing making them advance at all right now is adding processing power

      • bmitc 2 hours ago
        Same thing happened with self-driving cars. Oh and cryptocurrencies.
        • AntiUSAbah 2 hours ago
          Self-driving had never the amount of compute, research adoption and money than what the current overall AI has. Its not comparable.

          Crypto was flawed from the beginning and lots of people didn't understood it properly. Not even that a blockchain can't secure a transaction from something outside of a blockchain.

          • bigfishrunning 2 hours ago
            The LLMs are flawed, and lots of people don't understand them properly.
            • AntiUSAbah 1 hour ago
              People are researching how to make LLMs more stable and from a statistic point of view, we already now down to 10% (progress is made here).

              LLMs don't have to be perfect, they just need to be as good as humans and cheaper or easier to manage.

          • turtlesdown11 1 hour ago
            > Self-driving had never the amount of compute, research adoption and money than what the current overall AI has. Its not comparable.

            $100+ billion in R&D and it's not comparable... hmm

          • freejazz 1 hour ago
            > Self-driving had never the amount of compute, research adoption and money than what the current overall AI has.

            And yet they don't do really good jobs with pretty much anything, save for software development, to which people still seem pretty split as far as it being a helpful thing. That's before we even factor in the cost.

            • AntiUSAbah 58 minutes ago
              I find them very helpful. I use gemini regularly for multiply things.

              I also believe that whatever code researchers and other non software engineers wrote before coding agents, were similiar shitty but took them a lot longer to write.

              Like do you know how many researchers need to do some data analysis and hack around code because they never learned programming? So so many. If they know how to verify their data (which they needed to know before already), a LLM helps them already.

              There is also plenty of other code were perfection doesn't matter. Non SaaS software exists.

              For security experts, we just saw whats happening. The curl inventor mentioned it online that the newest AI reports for Security issues are real and the amount of security gaps found are real and a lot of work.

              Image generation is very good and you can see it today already everywere. From cheap restaurants using it, to invitations, whatsapp messages, social media, advertising.

              I have a work collegue, who is in it for 6 years and he studied, he is so underqualified if you give me his salary as tokens today, i wouldn't think for a second to replace him.

  • moi2388 1 hour ago
    Stop fucking linking paywalls ffs
  • helsinkiandrew 5 hours ago
    OpenAI post: https://news.ycombinator.com/item?id=47921262

    Tried to delete this submission in place of it but too late.

  • squirrelon 2 hours ago
    [dead]
  • theturtle 2 hours ago
    [dead]
  • chickoeafilae 4 hours ago
    [flagged]
  • Linuxislife 2 hours ago
    [dead]
  • jonahs197 2 hours ago
    [dead]
  • ath3nd 4 hours ago
    [dead]
  • freejazz 4 hours ago
    Impossible to take any of this seriously when it constantly refers to AGI.
    • Schlagbohrer 4 hours ago
      Especially when the OpenAI definition of AGI is only in financial terms (when it becomes profitable), which can be easily manipulated.
  • aliljet 5 hours ago
    Why is this being made public?
    • brookst 5 hours ago
      It’s an agreement between a public company and a highly scrutinized private company. Several of the provisions will change what happens in the marketplace, which everyone will see.

      I imagine the thinking was that it’s better to just post it clearly than to have rumors and leaks and speculations that could hurt both companies (“should I risk using GCP for OpenAI models when it’s obviously against the MS / OpenAI agreement?”).

      • Schlagbohrer 4 hours ago
        Also it's about OpenAI going public.
    • discordance 2 hours ago
      Might have something to do with the MSFT quarterly report tomorrow
  • shaguoer 2 hours ago
    Interesting perspective. Would love to see more discussion on this.