sert_121 4 hours ago

It's interesting to see how chatgpt is becoming more and more of a starting point of the web exploration, at which they're like, why even bother searching at this point, we'll just have default workflows for maps, buy (integration of stripe already marks it), booking airlines etc, which covers so much basic stuff people would do anyways.

The biggest bottleneck for this for the past two years imo wasn't the models, but the engineering and infra around it, and the willingness of companies to work with openaio directly. Now that they've grown and have a decent userbase, companies are much more willing to pay/or involve themselves in these efforts.

This has eventual implications outside user-heavy internet use (once we see more things built on the SDK), where we're gonna see a fork in the web traffic of human centric workflows through chat, and an seo-filled, chat/agent-optimized web that is only catered to agents. (crossposted)

  • ukFxqnLa2sBSBf6 2 hours ago

    I’m not sure how many people there are like me outside of this website but there’s not a single bone in my body that wants to use AI for these things.

    Buying plane tickets for example. It’s not even that I don’t trust the AI or that I’m afraid it might make a mistake. I just inherently want to feel like I’m in control of these processes.

    It’s the same reason I’m more afraid of flying than driving despite flying being a way safer mode of travel. When I’m flying I don’t feel like I’m in control.

    • tokioyoyo 2 hours ago

      I have very normie and not-so-normie friends that ask ChatGPT almost anything. My parents consistently make use of it, and they're almost 70, and not that tech-literate. There was a fun release from Anthropic about the type of queries that they're receiving, and code-gen is minority. I think we're, once again, not the average user.

      • OccamsMirror an hour ago

        I wonder how many of those “average users” will actually happily pay what the true cost is though? Are they really getting perceivable value for it or is it just more convenient than present day Google.

    • jwpapi 2 hours ago

      I would argue a website made to buy you tickets (skyscanner f.e.) is always gonna be a better interface than chat.

      Right now I cant imagine an AI (esp. chat) being more convenient for me than skyscanner or Google Hotels, but maybe I’m missing the imagination.

    • taurath 2 hours ago

      The majority of americans are more concerned with AI. Only like 22% are optimistic. And why would they be optimistic that it'll result in a better life for them

    • freakynit 2 hours ago

      Same here. Neither do I trust these tools to be working accurately, nor do I have the patience to wait for them to complete the given task when I ca do that manually 10x faster already.

  • heavyset_go 3 hours ago

    Just like I won't trust voice assistants to make purchases for me, or make decisions that actually matter, there is no way in hell that I'm letting an LLM be able to charge my credit card, let alone book flights for me lol

    • ultrarunner 2 hours ago

      Think of how much work it'll create to correct these charges. I bet there's application for understanding and issuing refunds & corrections…

      I suspect our future is going to be a lot more frustrating, both from AI screwups and the atrophied skills of humans

      • heavyset_go 23 minutes ago

        You'll be dealing with AI agents all the way down.

        A decade ago, I used one of the hotel aggregator sites to reserve rooms for vacation, and as I call the hotel to double check something on my way to the airport, I find out that I don't actually have a reservation and my room is already occupied. They couldn't do anything about it, as it was the 3rd party aggregator's mistake.

        Just getting the aggregator to admit, that no, even though their system says I have a reservation, the hotel confirmed it didn't exist took over an hour. I had to go through several layers of customer service, and I suspect different call centers, until someone called the hotel themselves and issued a refund.

        It was miserable and stressful to do from the airport, I would have lost my mind if I had to deal with chatbots for what was already a terrible experience with an automated purchase.

  • totallymike 3 hours ago

    I dismay at the possibility of this happening. What’s the point of an internet at all if one company controls, filters, and governs our entire usage of it?

    I understand an argument can be made that google is doing similar, but at least you can still search and end up on an actual site, rather than just play telephone via chatgpt. This concept is horrifying for so many reasons.

    • sert_121 2 hours ago

      I agree with the fact that a monopolized web is not friendlier to anyone. But seeing the trajectories of tech companies in the past decade, the unfortunate north star is distribution and the relentless pursuit of it.

      Even in that dire circumstance, I wish that the web versions keep up/are maintained, instead of being slowly deprecated, which happened for a lot of mobile-native versions of applications.

  • spullara 3 hours ago

    OpenAI has had this opportunity to do this since their meteoric adoption and fumbled it with plugins and then GPTs. Ironically, Anthropic's MCP could be just the ingredient needed to capture this position.

  • nutanc 2 hours ago

    Why would you go inside a chat box and try to force fit applications and show the applications in weird ways and then finally link out to the actual application instead of just putting a chat box inside the application which is the accepted way.

    • wes-k 2 hours ago

      If I had a human assistant, I'd ask them to book my flight. The chat box is your window to your AI assistant. Maybe this new assistant hasn't earned your trust yet, but it makes sense that trust-aside, you'd ask your assistant to do whatever they could do for you.

      • nutanc 2 hours ago

        Right. But my assistant wouldn't show me a new screen and ask me to do things in it :)

    • zer00eyz 2 hours ago

      Years ago (in the age of flip phones, think pre 2001) I worked at a bank.

      When we launched our mobile banking platform, one of the PM's there swore up and down that we should be piloting banking by text message. He was fabulously wrong at the time and in the end got a lot of things right.

      There are a lot of applications that could fit in a text box provided that your not doing the work rather that your delegating it.

  • stackedinserter 3 hours ago

    The main showstopper here is trust.

    I just can't let anything AI make decisions that have consequences, like spending money, buying anything, planning vacations, flights etc. It's so bad now (I've just tried) that I'm not sure if it will ever gain my trust.

    • sert_121 2 hours ago

      I see your point in user trust, and that's fair, but the same concerns have been prevalent since GPT3 rolled out , that no one would trust these tools to write or edit anything. However, since then users are growing to be more and more attuned to filter and distinguish the quality of the responses (doing invisible A/B testing of these responses), so maybe that's what providers want to capitalize on.

      ChatGPT has become one of the top-most browsed websites, and they want to capitalize on it even if 2% of the people actually trust the new integrations.

    • 3s 3 hours ago

      not to mention the privacy concerns associated with connecting my entire life to OpenAI or Anthropic. If you have the memory feature enabled, it's scary how much ChatGPT knows about you already and can even infer implicit thoughts and patterns about you as a person.

      • sert_121 2 hours ago

        I am sure it already knows a lot regardless of the memory feature, as long you're sharing your chat history/ have your history enabled, but I agree, it'd simply worsen it.

  • jmspring an hour ago

    Searching? Google has become shit and ads.

fidotron 10 hours ago

This conception makes sense iff you believe in ChatGPT as the universal user interface of the future. If anything the agentic wave is showing that the chat interfaces are better off hidden behind stricter user interface paradigms.

  • derekcheng08 9 hours ago

    I suspect there are many, many things for which chat is a great interface. And by positioning ChatGPT as the distributor for all these things, they get to be the new Google. But you're also right that many domains for which a purpose-built interface is the right approach, and if the domain is valuable enough, it'll have someone coming after it to build that.

    • munk-a 9 hours ago

      I have yet to see a chat agent deployed that is more popular than tailored browsing methods. The most charitable way to explain this is that the tailored browsing methods already in place are the results of years of careful design and battle testing and that the chat agent is providing most of the value that a tailored browsing method would but without any of the investment required to bring a traditional UX to fruition - that may be the case and if it is then allowing them the same time to be refined and improved would be fair. I am skeptical of that being the only difference though, I think that chatbots are a way to, essentially, outsource the difficult work of locating data within a corpus onto the user and that users will always have a disadvantage compared to the (hopefully) subject matter experts building the system.

      So perhaps chatbots are an excellent method for building out a prototype in a new field while you collect usage statistics to build a more refined UX - but it is bizarre that so many businesses seem to be discarding battle tested UXes for chatbots.

      • peab 8 hours ago

        agree.

        Thing is, for those who paid attention to the last chatBot hype cycle, we already knew this. Look at how Google Assistant was portrayed back in 2016. People thought you'd be buying starbucks via the chat. Turns out the starbucks app has a better UX

        • ryandrake 8 hours ago

          Yea, I don't want to sit there at my computer, which can handle lots of different input methods, like keyboard, mouse, clicking, dragging, or my phone which can handle gestures, pinching, swiping... and try to articulate what I need it to do in English language conversation. This is actually a step backwards in human-computer interaction. To use an extreme example: imagine instead of a knob on my stereo for volume, I had a chat box where I had to type in "Volume up to 35". Most other "chatbot solved" HCI problems are just like this volume control example, but less extreme.

          • chongli 6 hours ago

            It's funny, because the chat bot designers seem to be continually attempting to recreate the voice computer interface from Star Trek: TNG. Yet if you watch the show carefully, the vast majority of the work done by all the Enterprise crew is done via touchscreens, not voice.

            The only reason for the voice interface is to facilitate the production of a TV show. By having the characters speak their requests aloud to the computer as voice commands, the show bypasses all the issues of building visual effects for computer screens and making those visuals easy to interpret for the audience, regardless of their computing background. However, whenever the show wants to demonstrate a character with a high level of computer mastery, the demonstration is almost always via the touchscreen (this is most often seen with Data), not the voice interface.

            TNG had issues like this figured out years ago, yet people continue to fall into the same trap because they repeatedly fail to learn the lessons the show had to teach.

            • fishpen0 5 hours ago

              It's actually hilarious to think of a scene where all the people on the bridge are shouting over each other trying to get the ship to do anything at all.

              Maybe this is how we all get our own offices again and the open floor plan dies.

              • NBJack 5 hours ago

                Hmm. Maybe something useful will come of this after all!

                "...and that is why we need the resources. Newline, end document. Hey, guys, I just got done with my 60 page report, and need-"

                "SELECT ALL, DELETE, SAVE DOCUMENT, FLUSH UNDO, PURGE VERSION HISTORY, CLOSE WINDOW."

                Here's hoping this at least gets us back to cubes.

              • fragmede 5 hours ago

                Getting our own offices would simply take collective action, and we're far too smart to join a union, err, software developers association to do that.

            • freediver 2 hours ago

              Profound observation, thank you for this.

          • zdragnar 6 hours ago

            Remember Alexa? Amazon kept wanting people to buy things with their voice via assorted echo devices, but it turns out people really want to actually be in charge of what their computers are doing, rather than talking out loud and hoping for the best.

          • BolexNOLA 6 hours ago

            “volume up to 35”

            >changes bass to +4 because the unit doesn't do half increments

            “No volume up to 35, do not touch the EQ”

            >adjusts volume to 4 because the unit doesn’t do half increments

            > I reach over, grab my remote, and do it myself

            We have a grandparent that really depends on their Alexa and let me tell you repeatedly going “hey Alexa, volume down. Hey Alexa, volume down. Hey Alexa, volume down,” gets really old lol we just walk over and start using the touch interface

        • potatolicious 8 hours ago

          It's also a matter of incentives. Starbucks wants you in their app instead of as a widget in somebody else's - it lets them tell you about new products, cross-sell/up-sell, create habits, etc.

          This general concept (embedding third parties as widgets in a larger product) has been tried many times before. Google themselves have done this - by my count - at least three separate times (Search, Maps, and Assistant).

          None have been successful in large part because the third party being integrated benefits only marginally from such an integration. The amount of additional traffic these integrations drive generally isn't seen as being worth the loss of UX control and the intermediation in the customer relationship.

        • jwpapi 2 hours ago

          Omg thank you guys. It felt so obvious to me but nobody talked about it.

          A UX is better and another app or website feels like the exact separation needed.

          Booking flights => browser => skyscanner => destination typing => evaluation options with ai suggestions on top and UX to fine-tune if I have out of the ordinary wishes (don’t want to get up so early)

          I can’t imagine a human or an AI be better than is this specialized UX.

        • JambalayaJimbo 3 hours ago

          Current LLMs are way better at understanding language than the old voice assistants.

      • sanj 3 hours ago

        Hard disagree.

        At least in my domains, the "battle-tested" UX is a direct replication of underlying data structures and database tables.

        What chat gives you access to is a non-structured input that a clever coder can then sufficiently structure to create a vector database query.

        Natural language turns out to be far more flexible and nuanced interface than walls of checkboxes.

      • anal_reactor 8 hours ago

        > the tailored browsing methods already in place are the results of years of careful design and battle testing

        Have you ever worked in a corporation? Do you really think that Windows 8 UI was the fruit of years of careful design? What about Workday?

        > but it is bizarre that so many businesses seem to be discarding battle tested UXes for chatbots

        Not really. If the chatbot is smart enough then chatbot is the more natural interface. I've seen people who prefer to say "hey siri set alarm clock for 10 AM" rather than use the UI. Which makes sense, because language is the way people literally have evolved specialized organs for. If anything, language is the "battle tested UX", and the other stuff is temporary fad.

        Of course the problem is that most chatbots aren't smart. But this is a purely technical problem that can be solved within foreseeable future.

        • robotresearcher 6 hours ago

          > I've seen people who prefer to say "hey siri set alarm clock for 10 AM" rather than use the UI.

          It's quicker that way. Other things, such as zooming in to an image, are quicker with a GUI. Bladerunner makes clear how the voice UI is poor for this compared to a GUI.

        • koreth1 5 hours ago

          > I've seen people who prefer to say "hey siri set alarm clock for 10 AM" rather than use the UI. Which makes sense, because language is the way people literally have evolved specialized organs for.

          I don't think it's necessary to resort to evolutionary-biology explanations for that.

          When I use voice to set my alarm, it's usually because my phone isn't in my hand. Maybe it's across the room from me. And speaking to it is more efficient than walking over to it, picking it up, and navigating to the alarm-setting UI. A voice command is a more streamlined UI for that specific task than a GUI is.

          I don't think that example says much about chatbots, really, because the value is mostly the hands-free aspect, not the speak-it-in-English aspect.

          • hn_throwaway_99 5 hours ago

            Even when my phone is in my hand I'll use voice for a number of commands, because it's faster.

            • NBJack 5 hours ago

              I'd love to know the kind of phone you're using where the voice commands are faster than touchscreen navigation.

              Most of the practical day to day tasks on the Androids I've used are 5-10 taps away from a lock screen, and get far less dirty looks from those around me.

              • majormajor 3 hours ago

                My favorite voice command is to set a timer.

                If I use the touchscreen I have to:

                1 unlock the phone - easy, but takes an active swipe

                2 go to the clock app - i might not have been on the home screen, maybe a swipe or two to get there

                3 set the timer to what I want - and here it COMPLETELY falls down, since it probably is showing how long the last timer I set was, and if that's not what I want, I have to fiddle with it.

                If I do it with my voice I don't even have to look away from what I'm currently doing. AND I can say "90 seconds" or "10 minutes" or "3 hours" or even (at least on an iPhone) "set a timer for 3PM" and it will set it to what I say without me having to select numbers on a touchscreen.

                And 95% of the time there's nobody around who's gonna give me a dirty look for it.

            • fragmede 5 hours ago

              and less mental overhead. Go to the home screen, find the clock app, go to the alarm tab, set the time, set the label, turn it on, get annoyed by the number of alarms that are there that I should delete so there isn't a million of them. Or just ask Siri to do it.

              • jwpapi 2 hours ago

                One thing people forget is that if you do it by hand you can do it even when people are listening, or when it’s loud. Meaning its working more reliable. And in your brain you only have to store one execution instead of two. So I usually prefer the more reliable approach.

                I don’t know any people that do Siri except the people that have really bad eyes

                • fragmede an hour ago

                  God I miss physical buttons and controls. being able to do something without even looking at it.

        • freehorse 7 hours ago

          In an alarm, there is only one parameter to set. In more complex tasks, chat is a bad ui because it does not scale well and it does not offer good ways to arrange information. Eg if I want to buy something and I have a bunch of constraints, I would rather use a search-based UI where i can fast tweak these constraints and decide. Chathpt being smart or not here is irrelevant, it would just be bad ui for the task.

          • anal_reactor 6 hours ago

            You're thinking in wrong categories. Suppose you want to buy a table. You could say "I'm looking for a €400 100x200cm table, black" and these are your search criteria. But that's not what you actually want. What you actually want is a table that fits your use case and looks nice and doesn't cost much, and "€400 100x200cm table, black" is a discrete approximation of your initial fuzzy search. A chatbot could talk to you about what you want, and suggest a relevant product.

            Imagine going to a shop and browsing all the aisles vs talking to the store employee. Chatbot is like the latter, but for a webshop.

            Not to mention that most webshops have their categories completely disorganized, making "search by constraints" impossible.

            • zdragnar 5 hours ago

              Funny, I almost always don't want to talk to store employees about what I want. I want to browse their stock and decide for myself. This is especially true for anything that I have even a bit of knowledge about.

            • array_key_first 5 hours ago

              The thing is that "€400 100x200cm table, black" is just much faster to input and validate versus a salesperson, be it a chatbot or an actual person.

              Also, the chatbot is just not going to have enough context, at least not in it's current state. Why those measurements? Because that's how much room you have, you measured. Why black? Because your couch is black too (bad choice), and you're trying to do a theme.

              That's kind of a lot to explain.

        • WanderPanda 3 hours ago

          Alarm is a good example of an “output only” task. The more inputs that need to be processed the less a pure chatbot interface is good (think lunch bowl menus, shopping in general etc.)

        • majormajor 3 hours ago

          > Not really. If the chatbot is smart enough then chatbot is the more natural interface. I've seen people who prefer to say "hey siri set alarm clock for 10 AM" rather than use the UI. Which makes sense, because language is the way people literally have evolved specialized organs for. If anything, language is the "battle tested UX", and the other stuff is temporary fad.

          I do that all the time with Siri for setting alarms and timers. Certain things have extremely simple speech interfaces. And we've already found a ton of them over the last decade+. If it was useful to use speech for ordering an uber, it would've been worth it for me to learn the specific syntax Alexa wanted.

          Do I want to talk to a chatbot to get a detailed table of potential flight and hotel options? Hell no. It doesn't matter how smart it is, I want to see them on a map and be able to hover, click into them, etc. Speech would be slow and awful for that.

        • alganet 6 hours ago

          > Of course the problem is that most chatbots aren't smart. But this is a purely technical problem that can be solved within foreseeable future.

          Ah yes, it's just a small detail. Don't worry about it.

          • jeremyjh 5 hours ago

            I'm sure some very smart Chatbots are working on it.

          • anal_reactor 5 hours ago

            I don't understand how come that a website for tech people turned into a boomerland of people who pride themselves in not using technology. It's like those people who refuse to use computers because they prefer doing everything the old-fashioned way and they insist on the society following them.

    • foobarian 5 hours ago

      I knew it!

      -diehard CLI user

  • AlphaAndOmega0 9 hours ago

    >If anything the agentic wave is showing that the chat interfaces are better off hidden behind stricter user interface paradigms.

    I'm not sure that claim is justified. The primary agentic use case today is code generation, and the target demographic is used to IDEs/code editors.

    While that's probably a good chunk of total token usage, it's not representative of the average user's needs or desires. I strongly doubt that the chat interface would have become so ubiquitous if it didn't have merit.

    Even for more general agentic use, a chat interface allows the user the convenience of typing or dictating messages. And it's trivially bundled with audio-to-audio or video-to-video, the former already being common.

    I expect that even in the future, if/when richer modalities become standard (and the models can produce video in real-time), most people will be consuming their outputs as text. It's simply more convenient for most use-cases.

    • GoatInGrey 6 hours ago

      Having already seen this explored late '24, what ends up happening is that the end user generates apps that have lots of jank, quirks, and logical errors that they lack the ability to troubleshoot or resolve. Like the fast forward button corrupting their settings config, the cloud sync feature causing 100% CPU load, icons gradually drifting away from their original positions on each window resize event, or the GUI tutorial activating every time they switch views in the app. Even worse, because their app is the only one of its kind, there is no other human to turn to for advice.

      • handfuloflight 5 hours ago

        Hopefully, people, and technology aren't stuck in late '24.

  • notatoad 9 hours ago

    i can't imagine that users will be interested in asking chatGPT to ask zillow things, or ask chatGPT to ask canva to do things. that's a clunky interface. i can see users asking chatGPT to look up house prices, or to generate graphics, but they're not going to ask for zillow or canva specifically.

    and if the apps are trusting ChatGPT to send them users based on those sort of queries, it's only a matter of time before ChatGPT brings the functionality first-party and cuts out the apps - any app who believes chat is the universal interface of the future and exposes their functionality as a ChatGPT app is signing their own death warrant.

    • freakynit an hour ago

      Exactly.

      This is exactly the same playbook as has already been played multiple times in the past(and currently playing) by existing companies.

      These companies initially laid out red carpets for such builders, but once they themselves had enough apps, they started to tighten the rope, and then gradually shifted to complete 100% control and extortion in the name of "security" or other made-up-excuse.

      No-more walled garden. If something like this has to come (which I truly believe is helpful), it should be buiild on open-web and open protocols, not controlled by single for-profit company (ironical since OpenAI is technically non-profit).

    • throwacct 6 hours ago

      This x1000. Are businesses short sided enough to help create and develop another wallet garden just like "Google" and "Amazon" are right now? Time will tell but I think businesses want to own their sales funnel, not just give the user a way to avoid interacting with them.

    • echelon 9 hours ago

      Every company should see OpenAi as a threat. They absolutely will come for you when the time comes.

      It's just like Google and websites, but much more insidious. If they can get your data, they'll subsume your function (and revenue stream).

      • grugagag 3 hours ago

        That and the erosion in privacy make OpenAI somehthing to be very vigilant about.

  • mark_l_watson 5 hours ago

    I agree with you. I think chat interfaces are really good with voice interfaces while walking, asking for a foreign language lesson, effectively doing a web search while walking by speaking and listening to the answer.

    Other app-like interfaces like NotebookLM can be useful, for me one or two real uses a week.

    Then there is engineering small open models into larger systems to do structured data extraction, etc.

    I am skeptical about the current utility of agentic systems, MCP, etc. - even though I like to experiment.

    Someone else said that at least the didn’t go on and on about AGI today - a nice thing. FOMO chasing ASI and AGI will drive us bankrupt, and produce some useful results.

  • asim 10 hours ago

    It's not just as ChatGPT as the interface. It's that Chat with AI will now be the universal interface and every tech company will have their version of it. Everything you want to do will happen in one place. Cards will provide predefined and interactive experience. Over time you'll see entirely dynamic content get generated on the fly. The user experience is going to be one where we've shrunk websites to apps and apps to cards or widgets. Effectively any action you need to take can be done like this and then agents can operate more complex workflow in the background. This is probably the interface for the next 10 years and what replaces the mobile app experience and stronghold that Apple or Google have. This lasts until fully immersive AR/VR become a more mainstream thing. At that point these cards are on a heads up display but we'll be looking at something totally different. Like agents roaming the earth...

    • JumpCrisscross 9 hours ago

      This has been the pitched playbook for decades. (Metamates!) I'm increasingly convinced its driven by a specific generation of tech entrepreneurs who cut their teeth while reading ca. 1980s science fiction.

      I could see chat apps becoming dominant in Slack-oriented workplaces. But, like, chatting with an AI to play a song is objectively worse than using Spotify. Dynamically-created music sounds nice until one considers the social context in which non-filler music is heard.

      • fidotron 9 hours ago

        The thing it reminds me of is those old Silicon Graphics greybeards that were smug about how they were creating tools for people that created wealth when those other system providers "just" created tools for people tracking wealth.

        There's a whole bizarre subculture in computing that fails to recognize what it is about computers that people actually find valuable.

      • neutronicus 9 hours ago

        Chatting with an AI to play a song whose title you know, sure.

        Getting an AI to play "that song that goes hmm hmmm hmmm hmmm ... uh, it was in some commercials when I was a kid" tho

        • JumpCrisscross 9 hours ago

          > Getting an AI to play "that song that goes hmm hmmm hmmm hmmm ... uh, it was in some commercials when I was a kid" tho

          Absolutely. The point is this is a specialised and occasional use case. You don't want to have to go through a chat bot every time you want to play a particular song just because sometimes you might hum at it.

          The closest we've come to a widely-adopted AR interface are AirPods. Critically, however, they work by mimicing how someone would speak to a real human by them.

        • fragmede 9 hours ago

          more abstract than that, "I'm throwing a wedding/funeral/startup IPO/Halloween/birthday party for a whatever year old and need appropriate music". Or, without knowing specific bands, "I want to hear some 80's metal music". "more cowbell!"

          • array_key_first 5 hours ago

            You don't need AI for this, Spotify has, like, infinite playlists.

            Also their playlists are made by real people (mostly...), so they don't completely suck ass.

            • fragmede 5 hours ago

              Playlists aren't interactive though. I can't say "like this but with less guitar".

              Also, following the Beatport top 100 tech house playlist, and hearing how many tracks aren't actually tech house makes me wonder about who makes that particular playlist.

      • echelon 9 hours ago

        It's because Zuck can't own a pane of glass. He's locked out of the smartphone duopoly.

        Everyone wants the next device category. They covet it. Every other company tries to will it into existence.

  • yuriNator 5 hours ago

    The interface of the future is local "AI" in the form of functions embedded in hardware inferred from data sets

    One way to consider it that I like as an EE working in the energy model realm; consider the geometry of an oscilloscope.

    Electromagnetism to be carved up into equations that recreate it.

    Geometric generators that create bulk structure and allow for changing min/max parameters to achieve desired result.

    Consider a hardware system that boots and offers little more than blender and photoshop like parameter UI widgets to manipulate whatever segment of the geometry that isn't quite right.

    Currently we rely on an OS paradigm that is basically a virtual machine to noodle strings. The future will be a vector virtual machine that lets users noodle coordinates.

    Way less resource intensive to think of it all as sync of memory matrix to display matrix and jettison all the syntax sugar developers stuck with string munging OS of history.

  • gapeslape 9 hours ago

    I agree with what you are saying.

    I’m building a tool that helps you solve any type of questionnaire (https://requestf.com) and I just can’t imagine how I could leverage Apps.

    It would be awesome to get the distribution, but it has to also make sense from the UX perspective.

  • JumpCrisscross 9 hours ago

    > conception makes sense iff you believe in ChatGPT as the universal user interface of the future

    Out of curiosity, why iff?

    • ViscountPenguin 5 hours ago

      "iff" means "if and only if". It's common in mathematics.

      • JumpCrisscross 3 hours ago

        Correct. I’m asking why this SDK makes sense <—> ChatGPT becomes a universal interface. Why isn’t it useful for intermediate applications?

  • nextworddev 10 hours ago

    The apps can send any arbitrary HTML / interface back though.

    e.g. Coursera can send back a video player

    • foobarian 5 hours ago

      This will be a bunch of rushed garbage. It will be like Java applets

      • nextworddev 4 hours ago

        Maybe, but don't forget they are godly at iteration.

  • wslh 5 hours ago

    WeChat is the counterexample of your affirmation.

    • esafak 5 hours ago

      Is wechat purely conservational, without visuals? I think not.

  • glenstein 9 hours ago

    There's a lot of appropriate blowback against stupid AI hype and I'm all for it. But I do think in many respects it's a better interface than (1) bad search results, (2) cluttered websites, (3) freemium apps with upgrade nags, as well as the collective search cost of sorting through all those things.

    I remember reading some not-Neuromancer book by William Gibson where one of his near-future predictions was print magazines but with custom printed articles curated to fit your interests. Which is cool! In a world where print magazines were still dominant, you could see it as a forward iteration from the magazine status quo, potentially predictive of a future to come. But what happened in reality was a wholesale leapfrogging of magazines.

    So I think you sometimes get leapfrogging rather than iteration, which I suspect is in play as a possibility with AI driven apps. I don't think apps will ever literally be replaced but I think there's a real chance they get displaced by AI everything-interfaces. I think the mitigating factor is not some foundational limit to AI's usefulness but enshittification, which I don't think used to consume good services so voraciously in the 00s or 2010s as it does today. Something tells me we might look back at the current chat based interfaces as the good old days.

    • jerojero 8 hours ago

      I think you need to be careful here because you shouldn't be comparing chat apps to the current state of search results. Instead you compare it to the ideal or to the state of them before companies decided that instead of providing what people are looking for it was more profitable to provide them with related content that they're paid to show.

      We are at a moment where we're trying to figure out how to design good interfaces, but very soon after that the moment of "okay, now let's start selling with them" will come and that's really what we're going to be left with.

      In that regard, things like adblockers which now a days can be used to mitigate some of these defects you talk about are probably going to be much more difficult to implement in a chat-app interface. What are we going to do when we ask an agent for something and it responds with an ad rather than the relevant information we're seeking? It seems to me like it's going to be even more difficult to be in control for the user.

      • jeremyjh 4 hours ago

        Its fine though, because this technology is a commodity, anyone can run it or resell it. I expect I can continue paying Kagi or someone like them to provide a good experience at a fair price.

      • glenstein 5 hours ago

        I think you're right that it's going to get enshittified (in fact I tried to say a similar thing toward the end of my comment). I'll stand by this though, LLM Chat, as it exists now, is (imo) objectively better than Google Search, as it is now. Google Search at its best (or, say, Kagi), vs LLM Chat at its best, I would say there's an interesting open question, but I can see the case for chat winning.

        But I think it's going to be like Kagi, you'll pay for a subscription to a good-enough one, but the main companies will try to make their proprietary ones too feature rich and too convenient so that you'll have no choice but to use their enshittified version. What we have now might be a golden age that we will miss having.

        But, for better or worse, I do think what's coming may be a paradigm where they are effectively one big omniscient super-app.

        • jeremyjh 4 hours ago

          I'll say it: ChatGPT is better than Kagi, and better than Google Search 1.0 at searching the web and finding relevant sources, even if that is all you use it for is to just find links that you read. Usually its analysis is sound if I don't know anything about the subject matter.

    • dylan604 9 hours ago

      at least with bad search results, you had to look at them to know they were bad or become used to certain domains that you could prejudge the result and move to the next one. LLMs confidently tell you false/made up information as fact. If you fail to follow up with any references and just accept result, you are very susceptible to getting fooled by the machine. Getting outside of the tech bubble echo chamber that is HN, a large number of GPT app users have never heard of hallucinations or any of the issues inherit with LLMs.

  • cube2222 9 hours ago

    Is it? Honestly, most agents and/or ai apps I interact with that are actually useful present some form of chat-like interface.

    I’m not very bullish on people wanting to live in the ChatGPT UI, specifically, but the concept of dynamic apps embedded into a chat-experience I think is a reasonable direction.

    I’m mostly curious about if and when we get an open standard for this, similar to MCP.

    • fidotron 9 hours ago

      The whole value of an actual executive assistant is them solving problems and you not micromanaging them.

      What users want, which various entities religiously avoid providing to us, is a fair price comparison and discovery mechanism for essentially everything. A huge part of the value of LLMs to date is in bypassing much of the obfuscation that exists to perpetuate this, and that's completely counteracted by much of what they're demonstrating here.

    • neutronicus 9 hours ago

      Yes, I certainly prefer "chatting with Claude Code" to "Copilot taking forever to hallucinate all over my IDE, displacing the much-more-useful previous-generation semantic autocomplete."

      The former is like a Waymo, the latter is like my car suddenly and autonomously deciding that now is a good time to turn into a Dollar Tree to get a COVID vaccine when I'm on my way to drop my kid off at a playdate.

  • artursapek 9 hours ago

    Once it's efficient enough, you will be able to just vocally talk to your computer to do all of this. Text chat is just the simplest form of a natural language interface, which is obviously the future of computing.

    • s__s 5 hours ago

      I don’t think natural language is efficient enough. Whether that be text or voice.

      I imagine the Star Trek vision is pretty accurate. You occasionally talk to the computer when it makes sense, but more often than not you’re still interacting with a GUI of some kind.

    • fragmede 6 hours ago

      The ChatGPT phone app has had voice conversation mode for a while now. it's more interactive than a podcast while driving. There are apps (Wispr, non-affiliated) to make talking to your computer easier. The future is definitely a hybrid of them. sometimes I want to talk, other times I want to type.

rushingcreek 11 hours ago

I think this is very interesting, but it is reminiscent of what we built with Phind 2 where the answer could include dynamic, pre-built widgets.

The problem with this approach is precisely that these apps/widgets have hard-coded input and output schema. They can work quite well when the user asks something within the widget's capabilities, but the brittleness of this approach starts showing quickly in real-world use. What if you want to use more advanced filters with Zillow? Or perhaps cross-reference with StreetEasy? If those features aren't supported by the widget's hard-coded schema, you're out of luck as a user.

What I think it much more exciting is the ability to completely create generative UI answers on the fly. We'll have more to say on this soon from Phind (I'm the founder).

  • chatmasta 10 hours ago

    Phind is awesome. I often forget to use it until legacy search engines fail to surface what I’m looking for after a dozen searches. Phind usually finds it.

    That said, I used it a lot more a year ago. Lately I’ve been using regular LLMs since they’ve gotten better at searching.

    • rushingcreek 10 hours ago

      Thanks for the feedback. I think that our main differentiator going forward will be this generative UI on the fly for answering questions as opposed to search alone.

      • dleeftink 10 hours ago

        In a similar boat, but have been increasingly returning to for its quick notebook/charting capabilities. Would be awesome to somehow be able to select between different UI modes offering search, ranking, graphing or else depending on user needs.

  • alvis 10 hours ago

    Given there is already a MCP-UI project, I’m not surprised it can be done. But even that I’m not very convinced that it’s the right approach. After all, it’s still far too slow for real usage…

    • rushingcreek 10 hours ago

      Totally agree that it's too slow with conventional approaches, which is why we're training custom models for this that we can run fast

  • 9dev 9 hours ago

    Ah, that’s interesting. I’m considering building something similar for our product, and my solution to the schema constraints you mentioned thus far is breaking my widgets into blocks as universal as possible, as to still be useful. All of this is just ideas yet mind you, but my thinking was—maybe I can get the model to pick from a range of composable widgets depending on the task that are interoperable?

    For a concrete example, think a search result listing that can be broken down into a single result or a matrix to compare results, as well as a filter section. So you could ask for different facets of your current context, to iterate over a search session and interact with the results. Dunno, I’m still researching.

    Have you written somewhere about your experience with Phind in this area?

    • rushingcreek 6 hours ago

      Yes! We have a blog post here on how we designed these models and widgets: https://www.phind.com/blog/phind-2-model-creation.

      Now that models have gotten much more capable, I'd suggest to give the executing model as much freedom with setting (and even determining) the schema as possible.

  • irrationalfab 9 hours ago

    > If those features aren't supported by the widget's hard-coded schema, you're out of luck as a user.

    Chat paired to the pre-built and on-demand widgets address this limitation.

    For example, in the keynote demo, they showed how the chat interface lets you perform advanced filtering that pulls together information from multiple sources, like filtering only Zillow housers near a dog park.

    • rushingcreek 8 hours ago

      Yes, because it seems that Zillow exposes those specific filters as a part of the input schema. As long as it's a part of the schema, then ChatGPT can generate a useful input to the widget. But my point is that is very brittle.

      • handfuloflight 5 hours ago

        Isn't that as brittle as any system being constrained to providing only some type of outputs? Please elaborate.

        • rushingcreek 4 hours ago

          A fully generative UI with on-the-fly schema would be less brittle because you can guarantee that the schema and the intelligent widget can fully satisfy the user’s request. The bottleneck here is the intelligence of the model computing this, but we are already at the point where this is not much of a problem and it will disappear as the models continue to improve.

          I think most software will follow this trend and become generated on-demand over the next decade.

    • JumpCrisscross 8 hours ago

      > Chat paired to the pre-built and on-demand widgets address this limitation

      The only place I can see this working is if the LLM is generating a rich UI on the fly. Otherwise, you're arguing that a text-based UX is going to beat flashy, colourful things.

  • rco8786 6 hours ago

    That’s solved by MCP though. You can update your MCP’s servers schema dynamically without ever having to touch the app itself but the app will be aware of the new schema.

    • rushingcreek 6 hours ago

      I'm not saying that the schema can't change from time to time, I'm saying that having any fixed schema at request time is not an ideal user experience because it may not be clear what is supported and what is not supported. From first principles, it's much better if the app schema can be created dynamically at request time so we can guarantee that we can fully serve the user's request exactly as they asked it.

  • esafak 10 hours ago

    The problem is not the limitations of the capabilities per se but their discoverability (https://en.wikipedia.org/wiki/Discoverability). The user doesn't know what the capabilities are, as they are added and -- infuriatingly -- removed. Google Assistant is a perfect example of this.

    Conservational user interfaces are opaque; they lack affordances. https://en.wikipedia.org/wiki/Affordance

    • stavros 7 hours ago

      They don't lack affordances, you can do stuff. They lack signifiers, ie it's not easy to discover the stuff you can do.

      • esafak 7 hours ago

        Affordance is not what it can do, it is what it signals that it can do. It needs to be perceptible, by the definition I use (Norman's). I see others go by different definitions that even admit hidden affordances. I do not.

        • stavros 6 hours ago

          From The Design of Everyday Things:

          > Affordances represent the possibilities in the world for how an agent (a person, animal, or machine) can interact with something. Some affordances are perceivable, others are invisible. Signifiers are signals. Some signifiers are signs, labels, and drawings placed in the world, such as the signs labeled “push,” “pull,” or “exit” on doors, or arrows and diagrams indicating what is to be acted upon or in which direction to gesture, or other instructions. Some signifiers are simply the perceived affordances, such as the handle of a door or the physical structure of a switch. Note that some perceived affordances may not be real: they may look like doors or places to push, or an impediment to entry, when in fact they are not.

          With Norman's definition, if a conversational interface can perform an action, it affords that action. The fact that you don't know that it affords that action means there's a lack of a signifier.

          As you say, this is a matter of definition, I'm just commenting on Norman's specific definition from the book.

    • rushingcreek 10 hours ago

      Yep, this is a big problem as well. If the user doesn't know what features will or won't work, they lose confidence overall.

    • beefnugs 9 hours ago

      Thank you for this word. I have felt it my whole life and never learned the exact word.

      I immediately knew the last generation of voice assistants was dead garbage when there was no way to know what it could do, they just expected you to try 100 things, until it worked randomly

  • babyshake 10 hours ago

    I know that AG-UI from copilot kit is in this space. But it hasn't worked well with the MCP model AFAIK

ed 4 hours ago

A bit underwhelming when you see what's actually on offer. "Apps" are really just MCP servers, with an extension to allow returning HTML.

A lot of the fundamental issues with MCP are still present: MCP is pretty single-player, users must "pull" content from the service, and the model of "enabling connections" is fairly unintuitive compared to "opening an app."

Ideally apps would have a dedicated entry point, be able to push content to users, and have some persistence in the UI. And really the primary interface should be HTML, not chat.

As such I think this current iteration will turn out a lot like GPT's.

  • shredprez 4 hours ago

    MCP has this in the spec: it's called "elicitation", and I'm pretty confident this push from OpenAI sets the stage for them to support it.

    Once a service can actively involve you and/or your LLM in ongoing interaction, MCP servers start to get real sticky. We can safely assume the install/auth process will also get much less technical as pressure to deliver services to non-technical users increases.

    • ed 3 hours ago

      > Once a service can actively involve you and/or your LLM in ongoing interaction

      Is there any progress on that front? That would unlock a lot of applications that aren't feasible at the moment.

      Edit: Sampling is a piece of the puzzle https://modelcontextprotocol.io/specification/2025-03-26/cli...

      I also see a lot of discussion on Github around agent to agent (a2a) capabilities. So it's a big use case, and seems obvious to the people involved with MCP.

mhl47 11 hours ago

There was a recent post here about how deeply ingrained the chat interface is in OpenAIs organization. This really doubles down on that, but does anyone really like to interact with so much language instead of visual elements? Also feels horrible that you are supposed to remember a bunch of app names like "zillow" and punch them in the chat. And like an opportunity for them to slowly introduce ads for this apps or "preferential discovery", if you will, as monetization strategy.

Personally I don't hope thats the future.

  • baby_souffle 6 hours ago

    I feel like we're rehashing the debate around whether or not a GUI or terminal is more powerful.

    For a large number of tasks that cleanly generalize into a stream of tokens, command line or chat is probably superior. We'll get some affordances like tab auto completion to help remember the name of certain bots or mCP endpoints that can be brought in as needed...

    But for anything that involves discovery, graphical interaction feels more intuitive and we'll probably get bespoke interfaces relevant to that particular task at hand with some sort of partially hidden layers to abstract away the token stream?

  • agentcoops 10 hours ago

    Very much agreed. I think the dominance of the chat interface to LLMs has materially impaired the general usefulness of these tools — the sooner it goes away the better. It’s almost impossible to explain to a non-engineer how the illusion of a continuous conversation is crafted through context management and why past moments in a conversation might fall out of memory. My general advice to non-technical friends is to create a new conversation for each prompt so that they can get a more deterministic sense of how to formulate instructions and which are successful.

    I was really hoping Apple would make some innovations on the UX side, but they certainly haven’t yet.

  • drdrey 6 hours ago

    counterpoint: a lot of people around me just type "zillow" in google to access it, so maybe it's not absurd to refer to it by name in a chat interface

    • fishpen0 5 hours ago

      Right, but if you just search for "house listings" you find zillow and redfin and other stuff. Becoming the new word for "listings" will tie specific brands to our use of language in very interesting ways. What happens if I register my app to a common word. In this example, can I take "listings" and astroturf my app to the top? Is this a new DNS "buying all the domains" race?

      • aabhay an hour ago

        I mean ultimately you’re in OpenAI’s world, they have even more innate control of language, meaning, and truth

emilsedgh 10 hours ago

I see a lot of negative comments here but to me, it was obvious this is where OAI should land.

They want to be the platform in which you tell what you want, and OAI does it for you. It's gonna connect to your inbox, calendar, payment methods, and you'll just ask it to do something and it will, using those apps.

This means OAI won't need ads. Just rev share.

  • dewitt 10 hours ago

    > This means OAI won't need ads. Just rev share

    If OpenAI thinks there’s sweet, sweet revenue in email and calendar apps, just waiting to be shared, their investors are in for a big surprise.

    • dawnerd 2 hours ago

      Zapier has been doing this for how long and no one talks about them like some hot new startup.

  • nicce 10 hours ago

    > This means OAI won't need ads.

    Ads are defenitely there. Just hidden so deeply in the black box which is generating the useful tips :)

    • aniviacat 3 hours ago

      I wonder what the ad labeling (according to EU law) would look like in that case.

      In my (non-lawyer) understanding, each message potentially containing sponsored content (which would be every message, if the bias is encoded in the LLM itself,) would need to be marked as an ad individually.

      That would make for an odd user interface.

    • thebigkick 10 hours ago

      If you ask it to build a headless frontend web app, it immediately starts generating code with Next.js. I’ve always wondered how it was trained to default to that choice, given the smorgasbord of web frameworks out there. Next.js is solid, but it’s also platform-ware, tightly coupled to commercial interests. I wish there were more bias toward genuinely open-source technologies.

      • jerojero 8 hours ago

        There's probably different ways the LLM converged to it.

        One could be for example: from people asking online which tools they should use to build something and being constantly recommended to do it with Next.js

        Another could be: how many of the code that was used to train the LLM is done in Next.js

        Generally, the answer is probably something along the lines of "next.js is kind of the most popular choice at the time of training".

      • b_e_n_t_o_n 8 hours ago

        To me it feels like the default choice in the industry, perhaps it's not and I'm wrong but if I could have that feeling I can see how the AI can as well.

        • array_key_first 5 hours ago

          I've never seen next.js in the wild. I have seen plain React plus dotnet, though, a million times.

        • nicce 8 hours ago

          It is a trap. But once you realise that you are already too deeply invested.

      • intrasight 4 hours ago

        Just append to your prompt "not using a framework developed by a company that supports a genocidal fascist regime"

    • GoatInGrey 7 hours ago

      Because the AI labs are just hovering up all internet text that they can, I've been seeing more and more marketing pilots that deliberately seed marketing material in thousands of fake, AI-generated blogs and tutorials. The intention here is to get new LLMs to train on these huge numbers of associations between specific use cases and the company's product. All in a way that gets their marketing information into the final weights.

      You may have started seeing this when LLMs seem to promote things based entirely on marketing claims and not on real-world functionality.

      More or less, SEO spam V2.

  • jimmydoe 9 hours ago

    > This means OAI won't need ads. Just rev share.

    They obviously want both. In fact they are already building an ad team.

    They have money they have to burn, so it makes sense to throw all the scalable business models in the history, eg app store, algo feed, etc, to the wall and see what stick.

  • seydor 10 hours ago

    A platform requires a user moat or unfair advantage. Having a better quality model is neither

    • og_kalu 8 hours ago

      Consumer LLM apps have moat. As it is, ChatGPT (the app) spends most of its compute on Personal Non work messages (approx 1.9B per day vs 716 for Work)[0]. First, from ongoing conversations that users would return to, then to the pushing of specific and past chat memories, these conversations have become increasingly personalized. Suddenly, there is a lot of personal data that you rely on it having, that make the product better. You cannot just plop over to Gemini and replicate this.

      [0] https://www.nber.org/system/files/working_papers/w34255/w342...

    • typpilol 10 hours ago

      How's having the best model not a most?

      • maleldil 7 hours ago

        Because it changes all the time. A few weeks ago, it was Gemini 2.5 Pro, then Claude Opus 4.1, GPT-5 Thinking, now maybe Claude Sonnet 4.5, etc[1]. Having a good model isn't enough when they're basically interchangeable now. You need something else.

        [1] This is an example. Which model was the best when is not important.

      • zackangelo 9 hours ago

        Because it depends on how much better “best” is. If it’s only incrementally better than open source models that have other advantages, why would you bother?

        OpenAI’s moat will only come from the products they built on top. Theoretically their products will be better because they’ll be more vertically integrated with the underlying models. It’s not unlike Apple’s playbook with regard to hardwares and software integration.

fny 10 hours ago

It’s remarkable that will inevitably rush to build free apps that only reinforce OpenAI’s moat while cannibilizing their own opportunities.

  • tantalor 10 hours ago

    When the iPhone came out, there were like 6 apps, and no app store.

    In 2024, iOS App Store generated $1.3T in revenue, 85% of which went to developers.

    • codybontecou 10 hours ago

      Will this have a revenue share / marketplace built into it?

      • JumpCrisscross 9 hours ago

        > Will this have a revenue share / marketplace built into it?

        I'm genuinely surprised these companies went with usage-based versus royalty pricing.

      • rco8786 4 hours ago

        Altman mentioned an App Store is coming

    • hmate9 10 hours ago

      That figure sounds way too high

      Edit: yes I understand it is correct, but still it sounds like an insane amount

      • IncreasePosts 10 hours ago

        They're confusing "sales facilitates by the app store" with sales from the app store itself.

        That 1T figure is real, but it includes things like if you buy a refrigerator using the Amazon iOS app.

        • bangaladore 10 hours ago

          Yeah, the article itself even lists the reality at about 20% of the 1.3T.

      • moralestapia 10 hours ago

        It's true, though.

        It is now evident why Flash was murdered.

        • tracker1 10 hours ago

          Because it was buggy, known for security holes and the single biggest source of application crashes in all software in the late 90's through early 00's.

          • jjtheblunt 10 hours ago

            you missed the "it drained battery like there was no tomorrow" argument.

            • tracker1 9 hours ago

              I never really used it detached from a wall... mostly from work projects.

          • moralestapia 10 hours ago

            We get it, you drank the kool-aid.

            • tracker1 9 hours ago

              Drank the kool-aid?!? I worked in the eLearning space, I was a prominent user and developer for Flash/Flex content... there was some interesting tooling for sure, I also completely disabled it on my home computers as a result of working with it.

              I had a lot of hopes after the Adobe buyout that Flash would morph into something based around ActionScript (ES4) and SVG. That didn't happen. MS's Silverlight/XAML was close, but I wasn't going to even consider it without several cross-platform version releases.

              • moralestapia 8 hours ago

                >I was a prominent user and developer for Flash/Flex content

                I was as well. It wasn't as bad as people describe it. It was an amazing platform, HTML5 just recently caught up.

                In retrospective, Adobe should have open sourced it.

                >MS's Silverlight/XAML was close

                Hahahahahha, yeah sure! That tells me everything I need to know.

                • tracker1 8 hours ago

                  I agree it should have been open-sourced (at least the player portion)...

                  As for Silverlight, I mean the technology itself was closer to where I wanted to see Flash go. I'm not sure why you're laughing at that.

                  edit: as for not being as bad as people describe it... you could literally read any file on the filesystem... that's a pretty bad "sandbox" ... It was fixed later, but there were different holes along the way, multiple times.

        • JumpCrisscross 9 hours ago

          > We now know why Flash was murdered

          This is a stupid conspiracy given Apple decided not to support Flash on iPhone since before Jobs came around on third-party apps. (The iPhone was launched with a vision of Apple-only native apps and HTML5 web apps. The latter's performance forced Cupertino's hand into launching the App Store. Then they saw the golden goose.)

          • moralestapia 8 hours ago

            You ignore the state of things back then.

            HTML5 was new and not widely supported, the web was WAY more fragmented back then, to put things in perspective, Internet Explorer still had the largest market share, by far. The only thing that could provide the user with a rich interactive experience was Flash, it was also ubiquitous.

            Flash was the biggest threat to Apple's App Store; this wasn't a conspiracy, it was evident back then but I can see why it is not evident to you in 2025. Jobs open letter was just a formal declaration of war.

            • JumpCrisscross 8 hours ago

              > HTML5 was new and not widely supported

              Yes. It was a bad bet on the open web by Apple. But it was the one they took when they decided not to support Flash with the original iPhone's launch.

              > Flash was the biggest threat to Apple's App Store

              Flash was not supported since before there was an App Store. Since before Apple deigned to tolerate third-party native apps.

              You can argue that following the App Store's launch, Apple's choice to not start supporting Flash was influenced by pecuinary interests. But it's ahistoric to suggest the reason for the original decision was based on interests Cupertino had ruled out at the time.

  • jjtheblunt 10 hours ago

    what's their moat that you refer to?

  • mrcwinn 9 hours ago

    This is nonsense. Why would they destroy the incentive to get real-time, live data and MCP actions that help their users?

    Connecting these apps will, at times, require authentication. Where it does not require payment, it's a fantastic distribution channel.

cefboud 10 hours ago

This is an interesting branding exercise. Presenting MCP as 'Apps' makes it sound more accessible, while tools and MCP server sound very technical. Add a demo with Expedia and Spotify and you have an MCP that's end-user ready.

  • lossolo 9 hours ago

    Ye, that's basically an MCP server, that can be used by ChatGPT.

pu_pu 2 hours ago

This really feels like a missed opportunity to build something genuinely new, something that actually plays to the strengths of LLMs, instead of just embedding a fixed set of app screens inside chat.

Ideally, users should be able to describe a task, and the AI would figure out which tools to use, wire them together, and show the result as an editable workflow or inline canvas the user can tweak. Frameworks like LlamaIndex’s Workflow or LangGraph already let you define these directed graphs manually in Python where each node can do something specific, branch, or loop. But the AI should be able to generate those DAGs on the fly, since it’s just code underneath.

And given that LLMs are already quite good at generating UI code and following a design system (see v0.app), there’s not much reason to hardcode screens at all. The model can just create and adapt them as needed.

Really hope Google doesn’t follow OpenAI down this path.

darajava 8 hours ago

I don't understand, what could be built with this platform that wouldn't be made obsolete by conceivable updates to ChatGPT?

Another commenter suggested a hotel search function:

> Find me hotels in Capetown that have a pool by the beach .Should cost between 200 dollars to 800 dollars a night

ChatGPT can already do this. Similarly, their own pizza lookup example seems like it would exist or nearly exist with current functionality. I can't think of a single non-trivial app that could be built on this platform - and if there are any, I can't think of any that would be useful or not in immediate danger of being swallowed by advances to ChatGPT.

  • mindwok 4 hours ago

    ChatGPT can only do this now because the information is essentially freely available. Booking.com etc post their pages on the web to get traffic. In the world OpenAI is imagining, people will rarely if ever interact with the internet directly, it’ll instead all be through intermediary LLMs. In that world, the organisations that own authoritative information about hotel prices and locations will not make that freely available to LLMs, they will sell it. ChatGPT is trying to get ahead by encouraging them to embed themselves directly into their platform so they get first dibs on this kinda stuff before they put up the walls.

  • dworks 3 hours ago

    > Find me hotels in Capetown that have a pool by the beach .Should cost between 200 dollars to 800 dollars a night

    I built this 18 months ago at an OTA platform. We parse the query and identify which terms are locations, which are hotel features, which are room amenities etc. Then we apply those filters (we have thousands of attributes that can be filtered on, but cannot display all of them in the UI) and display the hotel search results in the regular UI. The input query is also through the normal search box.

    This does not need and should not be done in a chatbot UX. All the implementation is on the backend and the right display is the already existing UI. This is semantic search and it comes as a standard capability in ElasticSearch, Supabase etc. Though we built our own version.

    • spullara 3 hours ago

      You want it in a chat with other tools and intelligence so that you can give softer preferences and for it to judge reviews and the like. Perhaps even look at the room layout and photos to see if it is something you would like. There are good reasons to surround the tool you describe with AI.

  • bonoboTP 7 hours ago

    There are multiple branches they are exploring. This is a more structured one. But they also work on Agents that load the website and produce clicks to do the task. Also, this requires hand design, but they also work on generating the gui just-in-time, based on context.

    They also have this new design gui for visual programming of agents, with boxes and arrows.

    It's going to be a hybrid of all these. Obviously the more explicit work done for interoperability, the easier it is, but the gaps can be bridged with the common sense of the AI at the expense of more time and compute. It's like, a self driving car can detect red lights and speed limit signs via cameras but if there are structured signals in smart infrastructure, then it's simpler and better.

    But it's always interesting to see this dance between unstructured and structured. Apparently any time one gets big, the other is needed. When theres tons of structured code, we want AI common sense to cut through it because even if it's structured, it's messy and too complicated. So we generate the code. Now if we have natural language code generators we want to impose structure onto how they work, which we express in markup languages, then small scripts, then large scripts that are too complex and have too much boilerplate so we need AI to generate it from natural language etc etc

  • rco8786 4 hours ago

    There’s an incredibly long tail of profitable software business that would like to have a dynamic presence on ChatGPT that OpenAI would never have any interest in stealing. OpenAI wants to be the entry point to the internet, much like Google has been for the last couple decades.

Illniyar 7 hours ago

I can't understand the documentation. How are the interactive elements embedded in the chat? Are they just iFrames?

The docs mention returning resources, and the example is returning a rust file as a resource, which is nonsensical.

This seems similar to MCP UI in result but it's not clear how it works internally.

  • willtheperson 6 hours ago

    If the connector is enabled by the prompt or via a UI interaction, it calls your MCP server. They have created some meta fields your tool can respond with, one of which is something about producing a widget along with a field for html.

    In the current implementation, it makes an iframe (or webview on native) that loads a sandboxed environment which then gets another iframe with your html injected. Your html can include meta field whitelisted remote resources.

LudwigNagasena 5 hours ago

Glad to see no AGI hubris in this presentation, but we also haven’t see anything groundbreaking: their own version of GUI plugins, their own version of a workflow builder, and an aspiration to take cut of every transaction on the web.

I hope their GUI integration will be eventually superseded by native UI integration. I remember such well thought out concepts dating back to 2018 (https://uxdesign.cc/redesigning-siri-and-adding-multitasking...).

  • zmmmmm 5 hours ago

    AGI is so last year. Now it's all ASI which is great because it was achieved in like 1968 or something so nobody trying to achieve it can possibly fail

WillieCubed 10 hours ago

It's poetic that Google attempted to pursue apps within Google Assistant years ago, but the vision of apps within an AI assistant is more feasible now with LLMs that (whether actually or not) understand arbitrary user intents and more flexible connectors to third party apps via MCP (and a viral platform with 700+ million weekly active users).

Custom GPTs (and Gemini gems) didn't really work because they didn't have any utility outside the chat window. They were really just bundled prompt workflows that relied on the inherent abilities of the model. But now with MCP, agent-based apps are way more useful.

I believe there's a fundamentally different shift going on here: in the endgame that OpenAI, Anthropic et al. are racing toward, there will be little need for developers for the kinds of consumer-facing apps that OpenAI appears to be targeting.

OpenAI hinted at this idea at the end of their Codex demo: the future will be built from software built on demand, tailored to each user's specific needs.

Even if one doesn't believe that AI will completely automate software development, it's not unreasonable to think that we can build deterministic tooling to wrap LLMs and provide functionality that's good enough for a wide range of consumer experiences. And when pumping out code and architecting software becomes easy to automate with little additional marginal cost, some of the only moats other companies have are user trust (e.g. knowing that Coursera's content is at least made by real humans grounded in reality), the ability to coordinate markets and transform capital (e.g. dealing with three-sided marketplaces on DoorDash), switching costs, or ability to handle regulatory burdens.

The cynic in me says that today's announcements are really just a stopgap measure to: - Further increase the utility of ChatGPT for users, turning it into the de facto way of accessing the internet for younger users à la how Facebook was (is?) in developing countries - Pave the way for by commoditizing OpenAI's complements (traditional SaaS apps) as ChatGPT becomes more capable as a platform with first-party experiences - Increase the value of the company to acquire more clout with enterprises and other business deals

But cynicism aside, this is pretty cool. I think there's a solid foundation here for the kind of intent-based, action-oriented computing that I think will benefit non-technical people immensely.

bonoboTP 9 hours ago

This is part of the fight regarding whether we will have utility apps inside the chat app or chatboxes inside the utility apps. Obviously OpenAI would prefer that they are in the driver seat and delegate to passive apps, while regular apps like Booking would prefer to be the app the user uses and to run an AI chatbox nested inside their own app UI, so they can swap it out etc.

Convenience-wise probably this model is more viable, and things will get centralized to the AI apps. And the nested utilities will be walled gardens on steroids. Using custom software and general computing (in the manner of the now discontinued sideloading on Android) will get even further away for the average person.

  • somuchdata 8 hours ago

    They also released ChatKit today for building in-app chat UI experiences, so it seems like OpenAI is trying to make sure they get a larger slice of the pie no matter which interaction model wins.

wiradikusuma 10 hours ago

In 2018, I founded a startup specializing in chatbot for events. At the time the platforms were Alexa Skills, Actions on Google, and Messenger Platform (and LINE Bot, for people in Asia). I guess what's old is new again, but with fancier tech.

This time will be different?

  • jerf 10 hours ago

    We've actually got systems that can understand English now. Chatbots don't have to be glorified regular expression matches or based on inferior NLP. I've thought more than once that the true value of LLMs could well be that they essentially solve the language comprehension problem and that their ability to consume language is relatively underutilized compared to our attempts to get them to produce language. Under all the generative bling their language comprehension and ability to package that into something that conventional computing can understand is pretty impressive. They've even got a certain amount of common sense built in.

    • b_e_n_t_o_n 8 hours ago

      Yeah this seems accurate to me. All the talk of a bubble etc, but LLMs see genuinely useful at tasks like this and I'm sure we'll find more uses as time goes on.

  • apt-apt-apt-apt 4 hours ago

    Chatbots with and without GPT is like comparing a car with round vs triangular wheels

  • rco8786 4 hours ago

    You can’t think of anything that’s changed in the Chatbot space since 2018?

  • Traubenfuchs 9 hours ago

    Do people even want chatbots for events?

    I personally prefer well curated information.

    • cruffle_duffle 2 hours ago

      "I personally prefer well curated information."

      The LLM will do the curation.

dawnerd 2 hours ago

I’ve still yet to see how this improves anything? I saw someone mentioning it can use Spotify. Okay but like so can older gen assistants. Seems like they’re just trying to sell a much more expensive way of doing something that already exists.

sailfast 5 hours ago

Why would I want to enable OpenAI to collect an Apple Tax from me down the road?

Sure, this helps app partners access their large user base and grows their functionality too - but the end game has to be lock-in with a 30% tax right?

  • mrcwinn 4 hours ago

    For the same reason everyone's fine with an Epic tax down the road. It costs you nothing today.

spullara 10 hours ago

We have been building MCP servers and this looks very good directionally. Fills a bunch of holes in the protocol and gives meaning to something that were kind of like placeholders. Being able to return UI to the client is fantastic and will make lots of things possible. We have been working on these kinds of things assuming that the clients would improve to meet us.

https://lukew.com/ff/entry.asp?2122

skeeter2020 8 hours ago

Seems wild to have an App SDK for a technology that's 1. supposed to free us from purpose-built APIs and interfaces, and 2. comprised entirely of a single textbox. Feels perhaps more like a MS-type strategy of standards and formal rules intended to lock down the extended ecosystem?

  • reed1234 7 hours ago

    I think they want businesses to be more tightly integrated with ChatGPT to open up future opportunities for monetization.

whinvik 8 hours ago

Ads. They created ads. Now (or eventually) they can charge app developers to be featured first for a specific use case.

  • sumedh 7 hours ago

    Ads was always the end goal, they have an opportunity to become a bigger player than Google in the ad space.

    Instead of the user wasting time, ChatGpt can come up with the recommendations.

  • risyachka 8 hours ago

    How else would the company sell their product? and keep people employed.

    Of course ads will be there and this is good. A bad thing would be if they took a bunch of traffic from google and then gave no way to promote your products.

    That would lead to companies closing and layoffs and economy decline.

MaxPock 10 hours ago

This is honestly useful.

"Find me hotels in Capetown that have a pool by the beach .Should cost between 200 dollars to 800 dollars a night "

  • zzo38computer 3 hours ago

    I would not want to use LLMs for such a thing like that. Something like SQL queries or other kind of computer codes would be better. You would have to read the documentation, but it can be specified more precisely and more accurately. If you have a local program that can manage these queries (and then convert them to the remote service's format; a service could provide a file to specify the schema and the estimated cost of different fields) and interact with multiple services (including local files), then that will be better, without having to worry about problems with OpenAI, require as much power that OpenAI uses, more privacy violations than is necessary, etc.

    However, it might be useful for people who do want to use that instead.

  • pphysch 10 hours ago

    [injected with guerilla ads]

    I don't see how this is a significant upgrade over the many existing hotel-finder tools. At best it slightly augments them as a first pass, but I would still rather look at an actual map of options than trust a stream of generated, ad-augmented text.

    • elpakal 9 hours ago

      The benefit I see is that it meets users where they presumable already are (GPT). As other comments allude to here, it's clear they see themselves as a staple of the user's online experience.

    • AlBentley 5 hours ago

      exactly. Booking.com etc can just use OpenAI APIs to enable a similar voice/ chat interface on top of their search, and then the UX is not limited to 'cards'.

      The UI 'cards' will naturally becoming ever increasing, and soon you end up back with a full app within ChatGPT or ChatGPT just becomes an app launcher.

      The only advantage I can see is if ChatGPT can use data from other apps/ chats in your searches e.g. find me hotels in NYC for my upcoming trip (and it already knows the types of hotels you like, your budget and your dates)

    • b_e_n_t_o_n 8 hours ago

      I think the end game is that rather than spitting out text back, the LLM transforms your plaintext request to something processable, and then chooses some relevant widgets to display the results.

outlore 8 hours ago

remember when custom GPTs would just need an OpenAPI spec to be compatible with any existing API out there? we've been through this app store journey once before, maybe it's different this time since we now have agents and MCP

naiv 11 hours ago

Remember "GPTs" and the thing before it which I don't even remember now. I think this will go the same route .. to nowhere

  • minimaxir 10 hours ago

    The GPT App Store (which is technically now obsolete with this SDK) was funny.

  • elpakal 10 hours ago

    Are they still expecting us to get paid based on “revenue sharing”?

irrationalfab 10 hours ago

This feels like the death of the app, and the rise of the micro-app.

itsnowandnever 10 hours ago

this seems kinda silly, especially given their previous app store flop. but I'm just happy there's some spark and competition in tech again. it's felt like the industry has been pretty stagnant since web 2.0 (more stagnant than any other time in the last 40-50 years, anyway). but this AI stuff feels like another "1977 Trinity" moment

so, best of luck to OAI. we'll see how this plays out

helloguillecl 9 hours ago

Chat offers a far better experience than using Google—no more searching through spam-filled results, clicking between sponsored links, accepting endless cookie banners, and trying to read a tiny bit of useful content buried among ads and clutter.

It has the potential to bridge the gap between pure conversation and the functionality of a full website.

  • d4mi3n 9 hours ago

    I’m just worried they we’ll go from very obvious advertising to advertising that’s a lot harder to spot.

    I can block adds on a search engine. I cannot prevent an LMM from having hidden biases about what the best brand of vodka or car is.

    • helloguillecl 9 hours ago

      I agree. But Google has gone in that direction long ago: ads are now harder to distinguish from genuine search results. In many cases, the organic results are buried so deep that they don’t even appear in the first visible section of the page anymore.

    • somuchdata 8 hours ago

      Google could also have allowed invisible pay-for-placement without marking it as an ad. Presumably they didn't do that because undermining the perceived trustworthiness of their search results would have been a net loss. I wonder if chat will go in that same direction or not.

      • jerojero 8 hours ago

        Pretty sure it's illegal to present advertisement and not label it as such in some form.

        But as with everything, as new technologies emerge, you can devise legal loopholes that don't totally apply to you and probably need regulation before it's decided that "yeah, actually, that does apply to me".

disiplus 10 hours ago

Honestly I see how somebody like kayak.com would build a "app" they work through commission, they don't care from where is the booking coming from. But they will sort the flight tickets based where do they earn the best commission. What's in there for me as a user ?. Also will openai let different providers pay for the top placement when somebody tries to buy ticket on chatgpt ?

chvid 11 hours ago

Discovery, monetization. What is in it for developers?

  • spongebobstoes 10 hours ago

    deploying an app to 700M people?

    • artisin 9 hours ago

      Not only do you get to deploy your app to 700M users; you also get to provide responsive support for every single one of them!

      Per the docs: 'Every app comes from a verified developer who stands behind their work and provides responsive support'

      That's thinly veiled corporate speak for, Fortune 500 or GTFO

    • saberience 9 hours ago

      That's like saying making a website is like deploying an app for 7B people.

      Sure, but deploying a website or app doesn't mean anyone's going to use it, does it?

      I could make an iOS app, I could make a website, I could make a ChatGPT app... if no one uses it, it doesn't matter how big the userbase of iOS, the internet, or ChatGPT is...

      • handfuloflight 3 hours ago

        Right this same sleight of hand is encoded in the language used in the announcement to make building on this platform to be attractive seeming.

      • jryle70 4 hours ago

        Well, if you don't make it nobody would use it for sure.

ttoinou 11 hours ago

Does anyone think small players (like an independent developer) will be accepted ? Sounds like it will only for the big whales

hubraumhugo 10 hours ago

Why does everyone think chat is better UX than traditional interfaces? I get the AI hype, but so many products are not a fit for chat interfaces.

Why would I use a chat to do what could be done quicker with a simple and intuitive button/input UX (e.g. Booking or Zillow search/filter)? Chat also has really poor discoverability of what I can actually do with it.

  • throwacct 9 hours ago

    This x100. This is HCI 101. I'm glad I took that class during my master's program. It opened my eyes to a new world.

ttoinou 10 hours ago

That’s a great idea and Im wondering if Telegram can follow this path too, since they’re so advanced in mobile UX / UI, constantly updating their app and have some kind of crypto payments support.

hamonrye 3 hours ago

1GK AMD chips will accelerate

saberience 9 hours ago

What is the incentive for developers to build apps for this platform? I don't see any way of monetizing them at all.

  • jimmydoe 9 hours ago

    fear of missing out, as always, be the first flappy bird in the store.

nextworddev 10 hours ago

Your SaaS / Business is my Tool

MaxPock 10 hours ago

Tencent already has this with WeChat.Good to see it on chatgpt finally

benatkin 11 hours ago

They're looking like Facebook did with their phone project and later the metaverse - too big for their britches.

  • MaxPock 10 hours ago

    Lmfao..you've reminded me of the phone they made with HTC that had a Facebook button .

    • sieep 9 hours ago

      We've already sorta come full circle with the Meta glasses having a physical button to interact with the Facebook AI

danjl 10 hours ago

If only this somehow resulted in fewer, better apps. <sigh>

nthypes 8 hours ago

chat is the best interface for information retrieval and REPL-like experiences. for all the rest, chat is horrible.

mirzap 9 hours ago

Is it just me, or does it seem odd that if you truly believed AGI would be achieved within a few years, you wouldn’t launch an app store for AI apps? I don’t think an app store makes any sense in a post-AGI world.

klysm 4 hours ago

I guess openai is trying to execute the google playbook?

  • saxelsen 33 minutes ago

    I'll bet $100 they're seeing an opportunity to dethrone Google as the entrance point to the web and this is a big part of it.

    It feels like OpenAI's mission has changed from "We want to do do AGI" to

    "it'll be easier to do AGI with a lot of money, so let's make a lot of money first" to

    "we have a shot at becoming bigger than Google and stealing their revenue. Let's do that and maybe do AGI if that ever works out"

compacct27 11 hours ago

“Build our platform for us!”

Dig1t 6 hours ago

Just let the AI control my mouse and keyboard, let it use my device like a human. There's a huge swath of software already designed to be used by humans and anyone who uses ChatGPT knows that it's already been trained on every scrap of knowledge on how to use any existing complex software.

siva7 10 hours ago

This feels like a fever dream. As a developer everything changes every week. A new model, a new tool, a new sdk, paradigm we have to learn. I'm getting tired of all that shit.

  • jampa 9 hours ago

    As a JS developer for over 10 years who has seen multiple hype waves, here is my advice: You don't need to ride the first wave. You can wait until technology matures and see if it has staying power.

    For example, React and TypeScript were hard to set up initially. I deferred learning them for years until the tooling improved and they were clearly here to stay. Likewise, I'm glad I didn't dive into tech like LangChain and CoffeeScript, which came and went.

    • pyuser583 9 hours ago

      LangChain has gone? I thought it was still around.

      • jampa 9 hours ago

        It's still around, but the hype has faded. Users discovered numerous issues with the project and began abandoning it. I remember one month when everyone was all, "LangChain is the future," and another month when the sentiment became: "LangChain is terrible."

        You can see the hype cycle's timeline in HN's Algolia search: https://hn.algolia.com/?dateRange=all&page=0&prefix=true&que...

      • ajcp 9 hours ago

        It is but I feel it's main value prop as a developer friendly abstraction layer has been very well solved for by the actual model providers themselves, while LangChain itself have become more bloated, clunky, and to under-opinionated.

    • awesome_dude 9 hours ago

      This is how I feel about Rust.

      The big hype wave has finished now (we still have the "how dare you criticise our technology bros" roaming around though), the tooling is maturing now. It's almost time for me to actually get my feet wet with it :)

  • nlarew 10 hours ago

    Who says you have to learn this? You are free to ignore it if it's overwhelming.

    I'd much rather see a thriving ecosystem full of competition and innovation than a more stagnant alternative.

    • throwacct 9 hours ago

      With what exactly? They are desperately trying to create a "marketplace" and become gatekeepers on the backs of developers and businesses alike. There's no innovation here.

      • serial_dev 9 hours ago

        I guess what’s implied is that developers and businesses would innovate, not OpenAI directly.

        • throwacct 9 hours ago

          Knowing OAI's history, only big whales could survive being copied by the platform's owner—case in point: Amazon Basics. They're so big that most of the time, SMBs can't escape them and don't have a choice but to cave to Amazon's demands. Is your product successful? Great, I'll copy you, add the "Amazon basics" label, and start bombarding users with my "product".

          • pkaye 9 hours ago

            Amazon basics is a private label just like Costco and the Kirkland Brand. Same thing with Walmart, Target, Trader Joes, etc. And if these SMBs don't have to deal with Amazon, they will have to deal with a dozen copycats from China for anything that becomes a hit.

            • throwacct 9 hours ago

              Please check how Amazon Basics works and what SMBs are saying.

    • 65 9 hours ago

      For me the most annoying thing is APIs arbitrarily changing all the time. Completely change the entire Tailwind, ESLint, AWS SDK, etc APIs every 6 months? Why not! Heaven forbid you don't touch a project for a few months, blink and all your code is outdated.

  • cube2222 9 hours ago

    You just point your AI agent at the docs and have it build the integration with your app for you :)

    On a more serious note, it remains to be seen if this even sticks / is widely embraced.

  • garbawarb 9 hours ago

    Just get an LLM to do it for you.

  • alvis 9 hours ago

    The question is, whether having UI in chatgpt a game changer, fundamentally?

  • esafak 10 hours ago

    Specialize, escape, or accept.

    • falcor84 9 hours ago

      Like "Abort, Retry, Fail"? And same as there, what's the difference between the first and the third? Is there a way of accepting a new sdk every week without specializing?

  • apwell23 9 hours ago

    nothing really changed much here though. re llms nothing really has changed either, its mostly just scaling. there is really not much to learn as a consumer and app builder.

OtherShrezzing 9 hours ago

OpenAI launched an App Store in Nov 2023. A 23 month turnaround from major feature launch, to deprecation, to relaunch is a commitment to product longevity that’d put Google to shame.

  • AlphaAndOmega0 9 hours ago

    I found it genuinely impressive how useless their "GPTs" were.

    Of course, part of it was due to the fact that the out-of-the-box models became so competent that there was no need for a customized model, especially when customization boiled down to barely more than some kind of custom system prompt and hidden instructions. I get the impression that's the same reason their fine-tuning services never took off either, since it was easier to just load necessary information into the context window of a standard instance.

    Edit: In all fairness, this was before most tool use, connectors or MCP. I am at least open to the idea that these might allow for a reasonable value add, but I'm still skeptical.

    • CharlieDigital 9 hours ago

          > I get the impression that's the same reason their fine-tuning services never took off either
      
      Also, very few workloads that you'd want to use AI for are prime cases for fine-tuning. We had some cases where we used fine tuning because the work was repetitive enough that FT provided benefits in terms of speed and accuracy, but it was a very limited set of workloads.
      • apwell23 9 hours ago

        > fine tuning because the work was repetitive enough that FT provided benefits in terms of speed and accuracy,

        can you share anymore info on this. i am curious about what the usecase was and how it improved speed (of inference?) and accuracy.

        • CharlieDigital 9 hours ago

          Very typical e-commerce use cases processing scraped content: product categorization, review sentiment, etc. where the scope is very limited. We would process tens of thousands of these so faster inference with a cheaper model with FT was advantageous.

          Disclaimer: this was in the 3.5 Turbo "era" so models like `nano` now might be cheap enough, good enough, fast enough to do this even without FT.

  • kbar13 9 hours ago

    product roadmap was also ai generated

alvis 11 hours ago

So it’s take 2 for Open AI’s App Store moment. But this time surfing Anthropic’s MCP wave. Smart interop.. or just chasing the cool kids?

  • apwell23 10 hours ago

    mcp was a dud

    • consumer451 10 hours ago

      What is the superior way for an LLM to interact with your product?

      • apwell23 10 hours ago

        llm can call my existing apis fine. curious what kind of problems you are running to with your existing apis?

        • consumer451 an hour ago

          I want LLM chat integration with my product.

          So far, it seems that if you give an LLM a few tools to create projects and other entities, they seem to be very good at using them. The user gets the option of chat driven ui for our app, with not that much work for limited features.

          Currently building internal MCP servers to make that easy. But I can imagine having a public one in the future.

Handy-Man 10 hours ago

This is them trying to build ChatGPT into platform, from which they will take some portion of revenue generated by these apps...hmm where have I seen this before.

jasonsb 11 hours ago

They promised AGI and delivered SDKs. I think I'm gonna skip this one..

  • jsheard 11 hours ago

    Hey don't sell them short, they also delivered a TikTok clone with vertically integrated slop generation. It's the 5D Chess path to AGI, they just need to rot the average human brain until the bar for super-human intelligence is reduced to an attainable level.

    • Narciss 10 hours ago

      This was funny

testfrequency 10 hours ago

Wow.

“CEO” Fidji Simo must really need something to do.

Maybe I’m cynical about all of this, but it feels like a whole lot of marketing spin for an MCP standard.

throwacct 10 hours ago

Yeah... no. I'm going to pass. The premise is bad from any angle. In the case of businesses, why "create" another "Amazon" and compete with other brands when the focus should be on getting customers through my sales funnel? For developers is much worse since they are going to copy Amazon's model with brands that found a niche: Amazon Basics. In this case, it'll be OpenAI "core" (or something like that), where you do all the work, and when your "app" is somewhat famous enough or getting traction, they'll copy it, rebrand it, and bombard all old and new customers to use it instead of yours.

I'mma call it now just for the fun of it: This will go the way of their "GPT" store.

darkwater 10 hours ago

Oh, I guess tomorrow when American HQs come online we will get some new shiny thing barely tested that needs to be deployed in production ASAP. Or maybe there is already something waiting for me in Slack...

markab21 11 hours ago

The skepticism is understandable given the trajectory of GPTs and custom instructions, but there's a meaningful technical difference here: the Apps SDK is built on the Model Context Protocol (MCP), which is an open specification rather than a proprietary format.

MCP standardizes how LLM clients connect to external tools—defining wire formats, authentication flows, and metadata schemas. This means apps you build aren't inherently ChatGPT-specific; they're MCP servers that could work with any MCP-compatible client. The protocol is transport-agnostic and self-describing, with official Python and TypeScript SDKs already available.

That said, the "build our platform" criticism isn't entirely off base. While the protocol is open, practical adoption still depends heavily on ChatGPT's distribution and whether other LLM providers actually implement MCP clients. The real test will be whether this becomes a genuine cross-platform standard or just another way to contribute to OpenAI's ecosystem.

The technical primitives (tool discovery, structured content return, embedded UI resources) are solid and address real integration problems. Whether it succeeds likely depends more on ecosystem dynamics than technical merit.