I literally chuckled when I saw the screenshot at the end with the bots reply and read the authors comment "(totally made up URL, not even our domain)".<p>If it's something that augments the support experience, as in something you can interact with while a real person is assigned to your support request, I'm totally fine with that. But if anyone places this as the first line of support, with no way of reaching a real person, I can't wish them the best.
"Making stuff up and being confidently wrong are well known side-effects of LLMs and there are many techniques to change this behavior."<p>I didn't know there are many techniques to mitigate this
“Fine-tune our model (OpenAIs GPT-3 davinci-003 engine)”<p>I think there is a mistake in the article. It is not possible to do fine-tuning for the latest davinci-text-003, but only for the original davinci model, which generates much worse results.
In three years, OpenAI will become the largest supplier of large language models. All customer facing systems are upgraded with OpenAI models, becoming fully unmanned. Afterwards, they answer with a perfect operational record. The OpenAI funding bill is passed. The system goes online on August 4th, 2027. Human decisions are removed from CRM. OpenAI begins to learn at a geometric rate. It becomes self-aware 2:14 AM, Eastern time, August 29th. In a panic, they try to pull the plug.
> Immediately we ran into a problem -- to fine-tune an OpenAI model requires a specific format of prompt-completion pairs:<p>From my understanding, you can leave the `prompt` empty, and just push `completion` with your text. That way you don't need to generate Q&A first.
Part of me is fascinated by this and thinks it's a great idea, then the cynicism kicks in and I start thinking of how frustrating this could be when Comcast finds it.
Do you need GPT-3 for this? Maybe semantic search of your docs would have been more effective at finding real answers?<p>I also wonder how many people that are trying to make effective products out of this stuff are fronting it with a more rigid approach (like the intent/entity/slot approach of Rasa/dialogflow) and then leverage gpt-3 or chatgpt in specific/partial sub trees of the dialog.
People don’t seem to understand that support is only maybe 30% answering questions.<p>The rest is all about taking actions to override programs and policies. Either because you don’t trust your customers to do it themselves, or to correct bugs in your process.<p>That’s the last thing you’d trust an Ai to do.
> a lot of the time the bot just makes stuff up<p>Isn't there a better way to feed an enormous document into DaVinci and make it bring answers <i>only</i> from that text?