I think if your attitude is that its an oracle, then you already have the wrong attitude for using the tool. Chatgpt is a tool, if you dont know how to use the tool, stop complaining that you don't like it. Imagine telling everyone scalpels are horrible tools because they tried to perform surgery on someone and botched it up.
One day, not too far off, we are going to be able to tell a 'chat bot': make me a cartoon, it takes place in a steam punk fantasy. make it 2 seasons with 22 episodes each season. great, add a cliff hanger at the end of episode 11. add a love story component to it. reduce it back down to 1 season but 44 minute episodes.
content creation is no longer going to be tied back to knowing how to draw cartoons, or have armies of writers. yes, we can get garbage out of the system. but its a tool, plenty of tools produce garbage results if you dont know how to use them.
As an experiment, I asked chatgpt to write a business plan (one my brother started). The business plan was very close to what my brother produced, after working on it for a month. That's powerful, that's worthy of being 'the future'.
Broad strokes are broad strokes. I can procedurally generate levels all day long in a video game, but for me, they're never going to be as compelling or interesting as a lower-resolution and low quality textured game from the '90s or '00s where every single tree and rock is placed with intent.
I already think modern cartoons are fairly sterile and soulless versus their hand drawn or hybrid counterparts. It isn't even elitism, they just don't hold my attention or interest me artistically, stylistically, or in terms of content.
If you choose to express yourself in broad strokes, that's fine, whatever floats your boat. I'll continue to chase things that have intent and artistry behind every aspect of them. Generic and formulaic is generic and formulaic all day long. It's also why I don't like most modern anime, it's sterile visually and isn't why I enjoyed the medium.
For a writer, if you write out a plot, you can get the AI to actually simulate the character's responses and dialogue (even voiced by AI!). There, you've LITERALLY brought a character to life, each character is driven by a different persona simulated by a different AI, the quality of stories that will create, will annihilate what came before.
You want to write an adventure, but want to keep it unpredictable. Ask the AI for ideas, there, the adventure is now a true adventure, not a fake mirage created by the writer.
No need to describe scenery, no need to describe character appearances. Feed those descriptions into txt2img, and you get portraits that would have cost $1000/pic from top tier artists.
Generic and formulaicness, comes from having TOO MANY PEOPLE. Too many people involved in production, means the creator must dilute intent, appeal to wider audiences, and limit risks, to ensure costs are reclaimed. Once AI gets going, you'll see indie creators making full anime series, and releasing them on youtube. Because for an individual creator, even ad + patreon revenue alone would be able to sustain a comfortable existence, with no dependency on corporate or teams.
I thought people who love art, would be exhilarated by AI. I realized, the majority of artists don't love art. They love drawing, but not art. They love socializing with artists, but not art. They love receiving attention and income from their art, but not art. That's all fair and fine. But there will be people, who just want to create the best possible art, no matter the method, no matter the reward, and with AI, this latter group will outcompete the first, hard.
The popularity of Minecraft and other procedural games would imply that there is still a large number of people who value exploring the unknown generated content, even if it means it's not curated.
Yes the quality won't be as good, but you do get quantity instead. And the quality will improve.
What makes them great is that people enjoy them. Whether they were created with "Intent, attention to detail and self-expression" or monkeys banging on typewriters is irrelevant and indistinguishable.
For someone with that POV you sure are peddling the "Everything is soulless, stop enjoying things" perspective.
There are various finetuned models out there for conversations or story-telling, though they're quite small in terms of parameter count at the moment, but I don't see it as being fundamentally impossible.
How much knowledge and ideas did your brother personally develop over this month in addition to the business plan? Being handed a working plan is sometimes less useful than the aggregate experience leading up to the plan.
The "why?" is frequently more important than the "what?" with biz. An LLM doesn't really have an understanding of how the world works, so I would be very sceptical of its ability to write a sensible business plan. It doesn't even understand the product/service, and the nuances of what it does because it can't experience it.
The LLM is a yes-man with no experiences.
Imagine taking a year-old old book about what worked for someone's biz years ago, then naively assuming it'll work for yours today without considering how the world, market, people, technology et al may have changed in that time. This is the same thing. It assumes x is x. Usually x is not x. People used to compare their biz to Apple, despite not even being in remotely the same industry.
It's a tool that can be opaquely configured to be used in a million different ways, and when using it does not bring about the desired result, its acolytes sneer, and suggest that you're using it wrong.
It's like a multi-tool, that only works when you're blindfolded. Sure, it can be used to hammer nails and tighten screws and strip wires and measure a tire's pressure, but it makes it quite difficult to find the magical incantation that will apply the right end to the job.
(And most of the time, it quietly leaves the screw untightened, the wire clipped, and the tire with a hole in it. It's the user who's wrong, of course.)
When you use a hammer or a drill, do you expect it to sometimes not hit/screw the nail?
If ChatGPT is a tool for knowledge transfer/extraction, it can't hallucinate/lie to you/be wrong/make stuff up.
If it's a tool for potentially discovering some knowledge that may be true and needs to almost always be verified by either a compiler or a followup "find me a reference/discussion" Google search to make sure it's accurate, then sure. But I don't think that's what it's primarily being advertised as.
Web searches will for sure give you wrong answers. Even professors or other experts in a field will be wrong sometimes. Heck, even Einstein got some things wrong.
Your goalpost is in the wrong spot. Tools don’t need to be and probably never can be perfect. But that doesn’t mean they’re not useful.
Not sure about drills, but this absolutely happens with drivers if you fumble mating the bit to the screw head, or if you miss the stud, or if you overtighten, or if you don't sometimes pre-drill, or if you strip the head, or if you don't correctly gauge underlying material composition, or thickness, or if you...
This is such a weird statement from someone in the tech space. Programming languages rarely have an opinion for how they are to be used (for example JS MUST only run the browser or which code style to use).
When I chat with customer support, I wish they could meet me where I am instead of me needing to learn their tools. For example, I want to say "cancel my subscription" and my subscript get cancelled. I don't want to have to figure out which sub menu of the sub menu that has the magic "end subscription" button.
I know how to use my tool (english). LLMs teach computers how to use that tool too.
I also get support emails for password resets, which I try to make as simple as possible.
People don't want to learn new tools if their existing tools (language) work just fine.
Erm, they absolutely have an opinion. That's why I can't just write however I like in whatever language. I need to stick to the designer's opinions on syntax and semantics otherwise it won't work.
The only difference will be that script-following customer service reps giving you the runaround will be replaced by indefatigable chatbots giving you the runaround, which honestly sounds pretty hellish to me.
really? I feel like Apple's App Store provides great UX with their warning emails and centralized subscription management view. It is well documented too: https://support.apple.com/en-us/HT202039
But I still get emails asking me to cancel subscriptions.
> Compare that to looking at a typical chat interface. The only clue we receive is that we should type characters into the textbox. The interface looks the same as a Google search box, a login form, and a credit card field.
And Google is one of the most used tools. Probably more used than pen and paper today.
The point and click adventure games from Sierra and Lucas Arts were a huge step forward in interaction, although you didn't have to use your imagination as much to solve the puzzles.
And here we are again asking users to type their way to success.
Provided the responses aren't too brittle (and the LLM getting it wrong isn't too upsetting or find-out-too-late) lots of non power-users are going to prefer it, at least in cases where a menu or form input with about six options won't suffice.
Toolbars and menus provide affordances but you still need to know what things are called and what order to use them. "I'd like to email this file as a PDF and I'd also like to print it." may be much easier in a chat UX than in a menu based UX. Often these things can co-exist but chatUX has access to much more nuanced UI that would otherwise be too complex to build or expose.
I’m distressed by the growing use of chatbots in online payments. You do know the ontology—get balance information, make a payment, customer service.
I much prefer to not speak aloud to a robot on the phone, especially in the office, when there should only be three options.
Conversations with a robot speaking in a mixed tone of obsequiousness and superciliousness make me bananas.
It has all the affordances. You can turn it into anything you want. Want it respond with json, check, it can do that. Turn a wall of text into a Python data structures, check, it can do that too.
You can take your LLM text interface and put what ever api you want on top. You start with clay and mold it into anything you need. You construct a parser so that the way data is already constrained and validated. Same goes for the output.
I am not arguing that the chat interface is everything to everyone from a design perspective. I do think that as UI workbenches improve the need for a dedicated application will nearly disappear though. Applications will become a 1-5 page specification, including the UI.
>It has all the affordances.
these are two different ways of saying the same thing
A graphical UI can provide much more and much more intuitive guidance than a chat input ever will. And I say that as a big fan of Unix and the shell.
I could tell a chat bot I am finding the horizontal split in my editor is annoying because I have a wide monitor, and have it tell me there's a setting for that and ask if I want the default changed.
With a gui I might have to go through the files menu for settings, check if it's in edit-preferences, check tools-options, before maybe having to find out online it's if it's in some settings file.
"There's an ongoing trend pushing towards continuous consumption of shorter, mind-melting content. Have a few minutes? Stare at people putting on makeup on TikTok. Winding down for sleep? A perfect time to doomscroll 180-character hot takes on Twitter."
*280 chars, he means.
*4000 chars, he means.
*10000 chars, he means.
where's the fine-tuned prompt helper model?
I'm guessing that training custom language models on a company's data must be one of the hottest things you can be working on right now if you're looking for VC money (if there's something out there that compares to how well StableDiffusion+Dreambooth works for images, I'd be thankful for any pointers).
Web search changed that. Most queries work, at least somewhat.
There's a point where freeform text input becomes better than structured input. A simple search box is what people mostly use instead of an advanced search form, let alone a web directory (like Yahoo! back in the day).
For web search, there are very few error messages. If you enter a query that doesn't work very well, you get back results that aren't very good or what you wanted, so you try something else.
With AI chatbots, expectations are sky-high, but there are times when they should refuse with a good error message, because they really can't do what you're hoping to do. An example is when you ask it to explain its reasoning. An LLM never knows why it wrote what it did, but it will try to invent a plausible explanation anyway. [1]
Better error messages that help users understand what chatbots can actually do would help avoid misconceptions, but this won't happen unless the error messages are trained in.
[1] https://skybrian.substack.com/p/ai-chatbots-dont-know-why-th...
Classic example, Siri. It's so easy to quickly find stuff you feel like it should be able to do, but it just can't. "When was my last message from Steve" etc.
It should be pretty easy (even with today’s APIs and technology) to have an LLM design a user interface for you for your current task.
Simplest way: output a JSON of simple control definitions with every answer.
Coolest way: Just have it generate a full-ass React front end or whatever on every message.
Down the line people will expect applications to be chat ready. They will see an input box and expect the application to understand natural language and respond in the most helpful way. Which might be showing an error message or suggesting relevant next steps.
https://www.latent.space/p/build-ai-ux
full recording in the video at the bottom!
Quite a few people use search by awkwardly typing on mobile one or two words, probably misspelled and/or auto-completed as they type it. The query isn't sophisticated, lacks a lot of context and parameters, which the search engine then tries to guess.
When you use ChatGPT in that way, you'll get useless generic answers. It seems to shine specifically when being more specific, detailed, which also suggests users are willing and able (education level) to give such rich input.
The idea that it's better than search for this specific normy behavior, I openly question. And let's not forget about the economics. More expensive to run, vastly less ad space, and content owners (the whole web) are going to be pissed and will put up ever higher walls.
Put differently: if Google’s search models already have the ability to return great results for poor queries, why couldn’t a large language model (or a plug-in for one) learn the same algorithm?
Sidenote, i've found GPT useful enough to pay for (GPT Plus) by doing the opposite. Or rather, i find it very useful when i struggle to search for problems. ChatGPT helps guide me to new search or research terms, sometimes even providing the answer more directly.
It feels like the olden days where Google was great at finding a movie based on some vague movie description. GPT does that for a ton of things for me, enough that i found it useful.
It hasn't replaced online research but it has accelerated it for me.
So the one-track thinking of garbage-in-garbage-out is not the limitation any more.
What we're precisely now able to do is garbage-in-less-garbage out.
You can take a vague prompt in and ask GPT to hypothesize on what it means, why the user is asking that question and then generate a detailed prompt. Then use that that prompt and then perform the search.
Trying this out in the playground, I see a suprisingly capable search experience.
I use GPT-4 for debugging a lot now, because it's excellent at taking nothing other than an error message from the console and giving me back what's wrong and how to fix it. It's not perfect, but it's good enough that I reach for it by default now. I don't have API access to GPT-4 yet, and so I was comparing how well GPT-3.5 performed at this same task and for the example I tried, it just didn't get close enough for me to truly find it useful, so I wouldn't begin to rely on it in my daily workflow unlike GPT-4.
But... what I am actually quite interested in, and what I'm seeing a lot of, is exactly how far can you push a less capable model through prompt engineering? I think it's actually surprisingly further than you might have initially thought.
I just typed "eli5 hn vs redit" (misspelled reddit), and it understands perfectly.
Books were very simple machines to dress up our thoughts and the community reoriented itself well to the demands of the machine. But progress marches on the graves of obsolete machines as it did quills, book presses, typewriters, telegrams, libraries, word processors. Joe Reader has the same access to mind blowing dynamic text that creatives only wish to see as a finishing tool. He won't settle for the old glass window over static text just to please the artisan book writer.
1) I don't think chatting with anything, human or machine, is a learning experience, particularly since the machine veracity is poor, untrustworthy, and Hinton's resignation today tells you everything you need to know about the narrative inside big research orgs right now.
2) Recognition vs. recall. Given that it's the equivalent of an informal language CLI, which I prefer by the way; but there is no recognition (as in symbols) only recall.
Long story short, I think the emergent need is for written communication, with a tip of the hat to Daniele Procida:
https://ubuntu.com/blog/engineering-transformation-through-d...
Except that what's missing is a human-computer collaboration, i.e. sensemaking with another tip of the hat to Peter Pirolli:
https://www.efsa.europa.eu/sites/default/files/event/180918-...
Driving these suggestions is all of your data as well as your goals and values that you can give to the assistant in natural language.
At work the goal might be: "I want to sell $100,000 worth of widgets this quarter" and it will break down step by step how that might be possible.
For personal life it might be "I want to get involved in the kayaking community" and it will recommend activities, clubs, etc.
Once these assistants are good enough, it will be reckless to not use one (especially at work). We will then live in a world where AI and human live together and make decisions together hand in hand. Buckle up.
ChatGPT can be confidently stupid, but what if it gets better?
You need explanations/affordances of what it can and can't do only when its capabilities are limited. If it really could do whatever you asked, you wouldn't need it. Just say what you want.
Some neural connection to the brain that will interpret your thoughts is the only logical thing that will supersede chat ai, but that won't happen for a while. Maybe a connection to all your data will happen first so the AI will better understand what type of person you are and what you want, that's already probably happening based on past responses.
I was already convinced of this. What I'm not convinced of, and the article has little to say about, is
> Chatbots Are Not the Future... chatbots are not the future of interfaces.
Chatbots are a terrible interface to LLMs, and yet they are absolutely going to be the future of every third godawful website I must visit.
Indeed, and like any other text box, like whatsapp, like word — all tools that no one uses because they lack affordances.
He called ChatGPT oracle, nice but not enough.
I want someone to name chatbot their `oracle of delphi` plz. thank you.
"Please don't complain about tangential annoyances—e.g. article or website formats, name collisions, or back-button breakage. They're too common to be interesting."