Socially Intelligent Robots12
Nov 12, 2021Would you like a robot to assist you with tasks around the home? What kinds of jobs would you be comfortable leaving a robot to do? Would you trust one to take care of your child or an elderly parent? This week's show is "The Social Lives of Robots."
Comments (11)
Harold G. Neuman
Saturday, October 2, 2021 -- 12:19 PM
They need not have socialThey need not have social lives. Because they do not think...no matter what anyone tries to tell you. AI is contrivance. Turing would have told you that. Actually, he did. But no one was paying attention.
Tim Smith
Thursday, October 7, 2021 -- 7:11 PM
Turing said and did manyTuring said and did many things and was ignored for various reasons, but to characterize his view of AI as a contrivance and that no one noticed is like saying Barack Obama was a racist and Donald Trump loved his country and we all missed it. All these statements are valid, but all are equally untrue.
This is a cool little sub-thread, though. Alan Turing won WWII, was one of the greatest philosophers, mathematicians, and human beings of the last century… PT should do a show on him.
Harold G. Neuman
Sunday, October 3, 2021 -- 7:34 AM
I think professionals get soI think professionals get so wrapped in their disciplines, they are not able (willing?) To separate reality: 'how things probably are' from fantasy: ' how they might possibly be'. Sure, I tend to think outside the box and/or jump out of the system on a few things myself. It is good mental gymnastics;helps keep me sharper in advanced years. But, when we talk of notions such as computer social skills---or the lack thereof, we are in the realm of What Does it Matter?, not how things probably are. This is more Asimov than is either practical or pragmatic. Is it useful to speculate on the improbable? I don't think so. Artificial Intelligence is making differences-this much is inarguable. Mental gymnastics are good therapy. And, idle chat IS a social skill, banal as it can often be. Computers are marvelous tools, SKA, algorithims. I would, however, rather play chess with another human being.
Tim Smith
Thursday, October 7, 2021 -- 9:13 PM
Exoskeleton tech can getExoskeleton tech can get people with paraplegia mobile. Assistive robots have infinite patience for autistic kids to learn from and play with even. Eldercare robots give failing parents peace of mind. Medical rescue robots provide people with epilepsy dignity and timely care. There are way more applications than this for robots. These roles call for social skills and sensitivity to perform well. Philosophers must consider the limits and possibilities.
Advanced neuromorphic chips are in the pipeline to add neural networking units to process the cues that could detect a seizure, log depression, and find avenues to engage a young autistic mind. These and many other tasks can be done and done right. It only takes a few fails for people to lose their trust in tech to solve and address these high-need use cases.
I agree Robots don’t need to be social like humans, but they need to respond to human emotion and traits (some of which humans might not even perceive.) The most dangerous thing a robot could do would be to imitate humans, perhaps. Sometimes you want a robot to act and feel like a human; in general, robots should communicate socially with humans that they are not themselves humans. If that seems odd, it won’t very soon. Robots are becoming more and more human. That, I agree, is a problem.
Harold G. Neuman
Wednesday, October 20, 2021 -- 5:02 PM
Tim:Tim:
I do not often look to film or popular literature for philosophy.. But, I was moved by I Robot. The hi-tech Audi was neat, too.
Tim Smith
Wednesday, October 20, 2021 -- 5:55 PM
Harold:Harold:
"The Imitation Game" is worthy wrt Turing.
tinkwelborn@mac.com
Tuesday, November 2, 2021 -- 11:06 AM
Nice banter, here.Nice banter, here.
Daniel
Saturday, November 6, 2021 -- 6:14 PM
And informative! Not clearAnd informative! Not clear to me why more scholars are not making use of this excellent resource, which I describe as a public utility of intellectual exchange. At any rate, the notion of socialized robotics seems closely related to that of artificial intelligence, a term conspicuous for its ambiguity. If intelligence consists in the capacity for discrimination among relevant alternatives, e.g. in the case of a motion sensor signaling a person's body entering a structure rather than a flurry of leaves blown in by the wind, there's no detectable difference between artificial and natural intelligence. If, however, as forum participant Smith suggests in the first paragraph of his October 7 post above, artificial intelligence picks up where that of normal intelligence leaves off, then it's a continuation of the latter, and therefore distinguished only by degree. But the concern many have over how such technology is used in our society I think is very important. That is that artificial intelligence constitutes a consumer product of natural stupidity. Take self-driving cars. A product of natural intelligence in relation to existing conditions in the U.S. would be to reduce the economic externality of traffic congestion and damage to the atmosphere by hydrocarbon combustion by means of a rapid transition at the national level from private to public transportation. To represent artificially operated private transportation as a kind of intelligence would only then be plausible insofar as natural intelligence plays no role in its voluntary deployment.
Devon
Wednesday, November 17, 2021 -- 4:35 PM
Fanya in Aptos, CA sent usFanya in Aptos, CA sent us this question:
"Would you please ask the guest whether she's thought about using what we know about mirror neurons in her research to get a robot to model human beings and respond to humans?"
Unfortunately we received it just a bit too late to include in the program, but our guest Elaine Short sent this response:
"HRI researchers are often inspired by human cognition, including mirror neurons! In our work, that might look something like modeling how the robot would do a task while watching a person do that task. So a robot trying to figure out how to help someone put groceries away might make a plan for how to put away the oranges, which the robot is holding, AND a plan for how to put away the pasta, which the person is holding. It can then use the plan it made "for the person" to either detect when something surprising has happened ("The person put the box in the freezer! Maybe it was a box of mochi, not a box of pasta") or to make better decisions about its next actions ("The person is going to put the pasta in the cabinet, which means opening the cabinet by the fridge, so I should wait before I try to open the fridge")."
Tim Smith
Thursday, November 18, 2021 -- 9:50 PM
Mirror neurons wereMirror neurons were discovered in macaques. The only problem is … macaques don’t imitate. The fundamental papers on this were straightforward and unassuming. Later, mirror neurons were proposed to explain language, theory of mind, and imitation based loosely on a debunked theory of motor learning in speech.
An excellent book that explains this is ‘The Myth of Mirror Neurons' by Gregory Hickok. I read it a few years ago and can’t remember all the points. Still, he pretty much squelches any hope of using the mirror network to do anything other than help explain rudimentary motor/action function, which may be the use case here.
There are lessons to learn from human consciousness. These are not lessons for robots or artificial intelligence, for the most part.
As brought out in the show, the more common issue is that humans assume a robot is social, that it can learn and approach human expectations. As Elaine said, that would require good old-fashioned AI to accomplish at this point, and maybe all points. There’s too much goodness in robots to waste time living up to false and unrealistic expectations. Currently, we should focus on safety, ease of use, and easily automated assistive algorithms for high-need use cases.
Harold G. Neuman
Sunday, January 16, 2022 -- 1:53 PM
Whoa, now. This choice ofWhoa, now. This choice of subject matter implies that not only could robots be people, they are. A bit premature, don't you think? I am not there yet. I might countenance a notion of life extension through improved circuitry if only I could see some percentage in it. But I am not so self-indulgent (or self-assured) to believe extension of my life would make a difference that made a difference. That would amount to loathsome egomania---which we have seen more than enough of. I like to speculate about an array of topics. I even like dancing robots. Humans are social animals, when they choose. So are cats. Society; social; & socialization are words and concepts that fit certain life forms more precisely than others. Unless the definitions of the words and concepts have changed, they do not fit machines---even sophisticated ones. (Please. Make no mistake. AI and robots are machinery.) Those who elect to use such words and concepts in discussions around AI, robots and so on, need both a faster horse and better cart to put before it.
If this is absurdity, I'll take it over contextual reality. The next big thing is notoriously illusive.