You are viewing a single comment's thread from:

RE: AI is stupid: Humans Judgment Matters

in #steemit7 years ago

I toying with making something like this part of my research program in the next year or two. Though the secret sauce is how we help the AIs understand when they are having an aesthetic experience. Otherwise, we're basically just putting randomness into feedback loops to see what sticks. Whether or not you think that randomness gets you anywhere interesting I suppose depends on how much of a platonist you are.
I did write an interactive fiction where some AIs decided it was their religious duty to search for true randomness to see whether or not it existed.

Sort:  

That sounds cool! Is it up on the web somewhere?

No. It's an embryonic thought experiment for now. I'm going to run it past some AI researchers before I try to implement it.

No, I meant the interactive fiction.
The AI is cool as well though. I've been mentally wrestling with how one might simulate emotional responses in the last few days myself. Not for the usual fictional reason, but to help learning and creation. In science fiction they often have some android struggling with emotion, but I wonder if it might be far more important than we realize, and may need to be solved sometime soon. Not so we can have a computer cry because puppies, but so it can subjectively decide.

If you're interested, the interactive fiction The Entropy Cage is available online and for Android.
There's some theory that emotion is quick thinking and conscious thought is slow thinking. I'm not sure how I feel about that. The trick for building such AIs is coming up with a model that can be converted into code.

Perhaps it's simpler than I'm thinking. Perhaps it's a side effect of the complex reward system of the human brain, and something similar will come about in all significantly advanced general AI.

What are you ideas for emulating that in code? Our brains have complex reward systems as you say and also a few major parts. I thought the reward system was mostly tied to the simplest parts of our brain (the reptilian brain) with some ability to override that in the higher brain. We also have super-neurons to help with the think-fast. At least starting with bio-mimicry might lead somewhere.
(Please don't take this as me knowing anything particularly useful: I'm assembling words on the small odds that the conversation might spark off some useful inspiration)

I honestly have no idea. That's likely why my brain has been focused on trying to figure it out for days.
The "reward" system for a standard neural network is simple, from what I understand of it. It's just adjusting certain numbers in the neural net to make certain outcomes more likely. Our brain's reward system is far more complex. Although it does use various means to reinforce certain patterns, just like in an artificial neural net, there's also pleasure involved. I have absolutely no idea how we could possibly program pleasure into an AI or bot, yet I have a feeling it might actually be a necessity for certain AI tasks. Like when you make an AI that creates art. You could train it to create art based on famous artists. You could even use data on what people like, or how much they pay for different pieces of art, to make the AI more likely to create beautiful works of art that people will pay heaps for. You could even feed in data on the influences of different artists, so it could likewise create art that's based on certain famous paintings, while also being altogether different. But how do you teach it to "like" certain things? How do you teach it to enjoy certain artwork, and use that to create it's own?