Tufts University Develops Robots That Can Learn How To Trust
Researchers at Tufts University's Human-Robot Interaction Lab are programming robots that would give sci-fi author Dr. Isaac Asimov a bit of concern, according to his fictive Three Laws of Robotics: the Tufts engineers are "teaching" their robots the ability to say "no," or at least how to question what they are ordered to do, based on a certain list of criteria that they have been wired to consider before following a command.
The bipedal bots, which move with the eerily-precise mannerisms of a full-grown human being, use felicity conditions (i.e., the conditions that make a directee follow an order) to determine which commands are optimal for the robot to follow, as reported by the IEEE Spectrum:
- Knowledge: Do I know how to do X?
- Capacity: Am I physically able to do X now? Am I normally physically able to do X?
- Goal priority and timing: Am I able to do X right now?
- Social role and obligation: Am I obligated based on my social role to do X?
- Normative permissibility: Does it violate any normative principle to do X?
While it wouldn't be completely accurate to say that these robots have anything close to free will — at least how humans define it — they can supposedly use the latter felicity conditions to assess whether they will follow a given order, namely if the commander has the authority to have a directive carried out, and whether it will put the robot in any danger. So, while the bot might not have any sentience in a way that is entirely understandable to the average layman, it can develop, in a way, a level of trust — a very human characteristic, indeed.
Check out the rabble-rousing robots in the videos below.
Via: The Next Web
From Our Sponsor
Under The Tree: Smart Christmas Packaging Tips From Packsize, The Pros In On-Demand Custom PackagingTips on how to celebrate a merrier Christmas this year.