Aliza [message #97189] |
Mon, 13 November 2023 09:34 |
|
Wayne Parham
Messages: 18786 Registered: January 2001
|
Illuminati (33rd Degree) |
|
|
I've just released a fun little chatbot. As a nod to Weizenbaum's Eliza program, I've called it "Aliza."
If you don't know "Eliza" or who Joseph Weizenbaum was, ask Aliza.
You can reach it at any of these locations:
Right now, it has just the "standard" training included in gpt-3.5, which includes a bunch of data gathered up until around June 2021. So it can answer a lot of questions and be somewhat accurate on several topics. But then again, it lacks information on some things and where it lacks information, it will literally make stuff up. So keep that in mind when you interact with it.
I've done this largely as an experiment in large language model transformer-based AI. Specifically what I want to know is how much better I can train it in acoustics, and specifically on the details of Pi Speakers and recommended setups.
It already "knows" a lot about Pi Speakers, things like the fact that most models use waveguides, that I am the one that designed them, that they are high-efficiency designs and so on. But it doesn't know things like the part number of the waveguide, the proper use of flanking subs, the descriptions of the models, etc. So I will spend some time over the next few months fine-tuning a GPT dataset to give it this information.
The tricky thing about transformer-based large language models is that they know only words. The phrase, "a picture is worth a thousand words" falls on deaf ears here. Well, not exactly; A transformer will gobble up that phrase. And it can spit it back out to you quite elegantly but it has no idea what you are talking about. My point is that these chatbots have no "mental model" of the world. They have only mental models of words. So to teach them concepts that are best described with pictures is more difficult. You have to use your words.
More about AI, if you're interested:
|
|
|