PRE2018 3 Group6: Difference between revisions
Line 115: | Line 115: | ||
Margot sketched the current situation within human robot interaction mainly focused on the industrial sector. According to her there is a lot of improvement possible within the way robots and humans interact on the work floor. At this point both robots and humans do their own work independent of each other. Where often the robots do not signal what they are currently working on. This increases the uncertainty for human workers. And when humans become uncertain their anxiety increases and their productivity drops. This is because human workers who do not know what their robotic colleague is working on at this moment have only one way to reduce this uncertainty. And that is waiting to see what the robot is going to do next. | Margot sketched the current situation within human robot interaction mainly focused on the industrial sector. According to her there is a lot of improvement possible within the way robots and humans interact on the work floor. At this point both robots and humans do their own work independent of each other. Where often the robots do not signal what they are currently working on. This increases the uncertainty for human workers. And when humans become uncertain their anxiety increases and their productivity drops. This is because human workers who do not know what their robotic colleague is working on at this moment have only one way to reduce this uncertainty. And that is waiting to see what the robot is going to do next. | ||
“That is what people do when they do not understand the robot: they wait. Margot Neggers” | “That is what people do when they do not understand the robot: they wait. -Margot Neggers” | ||
Margot also explained previous research she has done for human-robot interaction. This was an experiment where a subject and a robot both would navigate concurrently through the same room. For this experiment they tried three cues: blinking leds, speech and gestures. From this experiment it became clear that subjects prefer speech as a communication tool over gestures and light signals. They found this the most pleasant way for the robot to communicate to them. But if performance was also taken into account it was found that speech was actually not the fastest gestures where. These findings were extremely interesting to us and formed the basis of this research project. | Margot also explained previous research she has done for human-robot interaction. This was an experiment where a subject and a robot both would navigate concurrently through the same room. For this experiment they tried three cues: blinking leds, speech and gestures. From this experiment it became clear that subjects prefer speech as a communication tool over gestures and light signals. They found this the most pleasant way for the robot to communicate to them. But if performance was also taken into account it was found that speech was actually not the fastest gestures where. These findings were extremely interesting to us and formed the basis of this research project. | ||
Line 123: | Line 123: | ||
Besides this Margot also pointed us towards the antromorpistic side of human-robot interaction. As she stated that humans working with robots often assign personalities to that robot. Which have a very large impact on how humans feel about the robot in future interactions. This could both be in a positive or in a negative sense. Which is of course strange when considering what a robot is. A simple combination of code and hardware with no personality attached. We thought this would be interesting and therefore decided to try to gain some information on this during the rest of the experiment. | Besides this Margot also pointed us towards the antromorpistic side of human-robot interaction. As she stated that humans working with robots often assign personalities to that robot. Which have a very large impact on how humans feel about the robot in future interactions. This could both be in a positive or in a negative sense. Which is of course strange when considering what a robot is. A simple combination of code and hardware with no personality attached. We thought this would be interesting and therefore decided to try to gain some information on this during the rest of the experiment. | ||
“The effect now is that the image of robots is defined by what people see. They might be quick to label him a dumb robot when he does something that ,in their eyes, is unlogical. Increasing their feeling of discomfort. Partly because they can not predict what the robot is going to do. Which means that they can not do their own tasks as they have to be conscious about the location of the robot. But if they do know what the robot is going to do then you can move comfortably in the same space. -Margot Nrggers” | |||
Revision as of 19:00, 6 April 2019
Group members
Name | ID |
---|---|
Pim van Berlo | 0957823 |
Timo Boer | 0965729 |
Charlotte Bording | 1246089 |
Luuk Roozen | 0948743 |
Panagiotis Kyriakou | 1256416 |
Problem statement
How can the uncertainty humans have about robots actions be reduced, using vocal queues.
- Can a vocal queue help reduce the anxiety people feel, because of uncertainty, towards these robots?
- Does the type of voice have influence? That is a more human like voice or a full on (Steven Hawking) robot voice.
Define Uncertainty:
employees working with robots have trouble understanding the robot.
- They do not know how dumb they are at the moment and give a robot a made-up character and intelligents based on what they know.
- Robot is not competing, trying to be better than human(humans can feel like they have to)(Margot)
- humans don’t understand the robot → feel uncomfortable → can't predict its actions → human stops doing his own task (Margot)
- “The ability to estimate engagement and regulate social signals is particularly important when the robot interacts with people that have not been exposed to robotics, or do not have experience in using/operating them: a negative attitude towards robots, a difficulty in communicating or establishing mutual understanding may cause unease, disengagement and eventually hinder the interaction.” https://link.springer.com/article/10.1007/s12369-016-0357-8
(So in short, People don’t understand what the robot is thinking/doing therefore the human is going to put false assumptions on the robot and will act on those assumptions. Or the human will just stop doing his/her tasks.)
Users
Full rundown of stakeholders involved with introduction of robots in warehouse environments. Since there are many different titles for warehouse workers I’m mentioning the ones that I think that need to be distinguished because of their relationship with the robot.
General Warehouse Laborer
- Description: Responsible for general duties involving physical handling of product, materials, supplies and equipment.
- Education: nothing more than a high school diploma needed.
- Much of their work is moving things around in the warehouse but in smaller scales like forklift drivers.
- Relatively safe (employment wise) since they are the jack of all trades in a warehouse
- Logically their interest is against the introduction of robots and if someone was causing inefficiencies because of the robot introduction it would be them. But since they will be the ones working the most with the robots, it makes sense to believe that they will have a positive stance against the robots having higher speaking skills since their job will be easier and they’ll socialize more.
Forklift driver, material handler
- Description: Operate forklifts to move pallets of products, materials, and supplies between production areas, shipping areas, and storage areas, load and unload trailers
- Education: again nothing more than a high school diploma but this time a forklift driver’s license is required and usually experience is required.
- The main group that will be getting replaced for their transporting role but could still remain because of loading and unloading
- Logically their interest is against the introduction of robots since this is the group that is mostly getting replaced, if one stays, their interest are the same as above
Shipping and Receiving Associate, Warehouse Clerk
- Jobs that mostly have to do with overlooking, organizing and preparing the shipments and with speaking with clients to make new deals respectively.
- Education: again, nothing more than a high school diploma but computer handling skills are required
- This group of employees (not employers) will probably be the one that is going to get affected (replaced or change of responsibilities) the least since they do mostly work with human clients or the make higher level decisions
- I believe their interest again should be somewhere in the middle, since their colleagues are mostly getting replaced but they can do their jobs better since the introduction of robots will bring efficiency. Now for them, that their jobs are fairly stable, it makes sense to believe that they will have a positive stance against the robots having higher speaking skills because they will bring better collaboration and thus their organizing tasks will be easier.
Sources (Links for now) https://www.horizonstaffingsolutions.com/different-types-warehouse-jobs https://www.friday-staffing.com/blog/warehouse-job-titles-and-descriptions/ https://www.indeed.com/
What do the users require?
- users require
- efficiency
- worth of their work
- safety
- robots doing tedious tasks
- time to adapt
- absence of stress
- understandable robots
- Do not require:
- losing their job
- make working with a robot less efficient
- annoying robot
Approach(last edit 2/17)
A qualitative design research.
Trough literature we will broaden our scope of topic. We will look into related work and based on that create a prototype that can give us more knowledge of the issue we are looking into. We will test the prototype on an small group of people(6 to 8) and will evaluate it with them very thoroughly. The results will be analyzed and based on the results together with the literature we will create a discussion and conclusion.
Milestones
- writing introduction,
- prototype finished
- user-test and/or user interview
- using and analyzing results of user-test
- finalizing report - conclusion
- creating presentation
Interviews
A large of the orientation and validation process has been done using interviews with subject experts. This ranged from interviewing philosophy professors to Robotics R&D engineers. Combining these interviews with the knowledge gained from our literary research served to create a broad base of knowledge concerning robotics and human-robot interactions. Here you can find short the mayor findings and synopses from these interviews.
Margot Neggers
Margot Neggers is a PhD Candidate at the Technical University of Eindhoven. Where she does research for the Human Technology Interaction groep. Mainly concerning robots and their interaction and communication with humans.
Margot sketched the current situation within human robot interaction mainly focused on the industrial sector. According to her there is a lot of improvement possible within the way robots and humans interact on the work floor. At this point both robots and humans do their own work independent of each other. Where often the robots do not signal what they are currently working on. This increases the uncertainty for human workers. And when humans become uncertain their anxiety increases and their productivity drops. This is because human workers who do not know what their robotic colleague is working on at this moment have only one way to reduce this uncertainty. And that is waiting to see what the robot is going to do next.
“That is what people do when they do not understand the robot: they wait. -Margot Neggers”
Margot also explained previous research she has done for human-robot interaction. This was an experiment where a subject and a robot both would navigate concurrently through the same room. For this experiment they tried three cues: blinking leds, speech and gestures. From this experiment it became clear that subjects prefer speech as a communication tool over gestures and light signals. They found this the most pleasant way for the robot to communicate to them. But if performance was also taken into account it was found that speech was actually not the fastest gestures where. These findings were extremely interesting to us and formed the basis of this research project.
Another important insight gained from Margot was that ,especially in this discipline, it is crucial for your results that subject do not have to imagine what something would be like. Because when subjects think deeply about what how they would experience something the results would differ from when they actually experience it. And for human-robot interaction this could have a very large impact on your final conclusion. Therefore we started to work towards doing an actual experiment to give more validity to our conclusion.
Besides this Margot also pointed us towards the antromorpistic side of human-robot interaction. As she stated that humans working with robots often assign personalities to that robot. Which have a very large impact on how humans feel about the robot in future interactions. This could both be in a positive or in a negative sense. Which is of course strange when considering what a robot is. A simple combination of code and hardware with no personality attached. We thought this would be interesting and therefore decided to try to gain some information on this during the rest of the experiment.
“The effect now is that the image of robots is defined by what people see. They might be quick to label him a dumb robot when he does something that ,in their eyes, is unlogical. Increasing their feeling of discomfort. Partly because they can not predict what the robot is going to do. Which means that they can not do their own tasks as they have to be conscious about the location of the robot. But if they do know what the robot is going to do then you can move comfortably in the same space. -Margot Nrggers”
Jilles Smids
Jilles Smids is a PhD candidate in the ethics of persuasive technology at the section of Philosophy and Ethics at the Technical University of Eindhoven. He has been working on a paper named: Robots in the warehouse: working with or against the machine. This got our interest and kick-started our project. Jilles gave us confirmation in the issues that are at hand with humans and robotics in warehouses. He mentioned that the fear of human workers being taken over by robots is unjustified. Warehouses are actually expanding and cannot find the the excessive workforce needed to fill the warehouse so use robots to co-exist with the current workers. According to Jilles, people tend to fear robots because they are afraid of losing their jobs, whilst robots can actually take over the tasks that are too heavy or uninteresting for humans to do.
Jilles helped us realize that a big problem in automation in the warehouse is not informing workers of what exactly is going on. This leads to cases where robots are being vandalized by human co-workers. It helped give us the idea that humans should be more aware of the actions of robots and gave us a clear ground to start our project from.
Jilles also gave us certain examples of where humans and robots did not get along; such as the Waymo’s self driving car incidents in Arizona. This made it clear that humans and robots are still in early phases of understanding and working together.
Bas Coenen
Bas Coenen is a graduate from the Technical University of Eindhoven, who is now currently working at Vanderlande as Team Leader Robotics, Senior R&D integration Engineer.
Bas helped us get a more clear understanding of what was actually being done with warehouse-type robotics. He gave insight on current issues with warehouses and robotic involvement around the world. He mentioned that multinational companies that have warehouses/factories all over the world have to deal with different cultures. Something as simple as a positive gesture here could mean something negative somewhere else. So robots that are created have to be able to fit-in in any culture over the world. Otherwise, there will always have to be changes in every warehouse over the world and as a company this can be inconvenient.
He also took us into a demonstration area where different setups for different supply lines were displayed. This gave us a better sense of how professionals have made solutions to warehouse robotics. However, most lines still gave the sense that robots and humans are separate from each other and not really co-existing.
His colleague explained about her work on adding eyes to robots to give a certain sense of being. This way people would notice when they are being noticed. For example, if you are trying to cross a pedestrian crossing and a car is coming, many people try to make eye contact with the driver to see if they are being noticed. This way you know you are seen and won’t be run over. Robots tend to just carry out their task and not give a sense of knowing someone is there. People tend to be afraid or uncertain of robots actions so they stop working to make sure they are safe. If a person notices it is noticed it could save valuable time.
Speaking to people actually working on robotics on a large scale definitely gave us some ideas on how to carry out our experiment. And allowing us to use some of their time, increased the interest we have in robotics.
Who does what
- everybody works on writing report
- Charlotte(Industrial design) --> Can do user tests, can help create prototypes.
- Luuk(Web science) --> Can analyze test data, if it is available. Can code for the prototype.
- Timo(Software Science) --> Can also analyze test data, if it is available. Can code for the prototype.
- Pim (Electrical Engineering) -> Contacts in the field of Robotics
Report
Abstract
Introduction(draft)(last edit 2/17)
Industrial robots are taking over more and more tasks in the workforce. Where there were 1.2 million Industrial Robots around the world, there were 1.9 million in 2017 (West, D. M. (2015). The kind of tasks are usually very repetitive (Tamburrini, G. ,2019), jobs that can be seen as very soul-sucking. The robots are making the process faster and cheaper. For example, robots can work 24 hours a day without getting tired. Robots will take in more space in the workforce, robots are getting smarter, cheaper and are starting to take over more complex tasks (West, D. M. (2015). The transition to robot co-workers can often bring troubles due to the human resistance of robots like humans having trouble understanding the robot, the robot working in a different pace (Weiss, A., et al. 2016) or the fear of being replaced by the robot.(Salvini, P. et al. 2010).
In this research we are looking at creating a more efficient and pleasant transition for the human co-workers by giving the robot a human aspect. It will be tested if a voice, and certain usages of the voice can create more trust and understanding of the robot and therefore less resistance. In a study done by Sauppé, A., et al It is already shown how giving the robot eyes can make its functions clearer for humans (2015). Peoples jobs often change due to the introduction of a robot in their work (Salvini, P. et al. 2010). An application of the voice could be explaining the new tasks in a suitable way. This research will mainly focus on industrial robots in warehouses as this is a very common place where robots are taking over tasks(..., ….)
Citation
Tamburrini, Guglielmo. (2019). Robot ethics: a view from the philosophy of science.
Weiss, A., & Huber, A. (2016). User Experience of a Smart Factory Robot: Assembly Line Workers Demand Adaptive Robots. In AISB2016: Proceedings of the 5th International Symposium on New Frontiers in Human-Robot Interaction.
Sauppé, A., & Mutlu, B. (2015, April). The social impact of a robot co-worker in industrial settings. In Proceedings of the 33rd annual ACM conference on human factors in computing systems (pp. 3613-3622)
West, D. M. (2015). What happens if robots take the jobs? The impact of emerging technologies on employment and public policy. Centre for Technology Innovation at Brookings, Washington DC.
Salvini, P., Laschi, C., & Dario, P. (2010). Design for acceptability: improving robots’ coexistence in human society. International journal of social robotics, 2(4), 451-460.
Related work
In the introduction saw that there is not just an increase in the amount of robots, but also in the difficulty of the tasks they perform. In this section we will cover some of the many works that relate to the problems emerging from these increases. We will also cover research that has been done to specific aspects of our problem.
One of these papers focusses specifically on how helpful vocal interaction can be in learning. In the paper “Effects of voice-based synthetic assistant on performance of emergency care provider in training”[1] the researchers measure whether a person can learn how to provide emergency care more quickly if it receives assistance from a voice-based synthetic assistant. This research paper is related to our work in that it shows that a vocal stimulant can help people in situations that are normally difficult for them to deal with. Giving us an indication that our research might result in a positive outcome.
Another paper that is related to ours is the paper called “Intelligent agent supporting human-multi-robot team collaboration”[2]. This paper gives an in-depth explanation of how an intelligent agent can help humans in their collaboration with robots. This article is especially interesting since one of the scenarios they focus on is warehouses. In this scenario the intelligent agent suggest possibly difficult situations for robots, such as a box that has fallen. The human can then judge whether this hinders the robots or not, and possibly change the robots behavior based on this. At some point the paper discusses whether the human found these suggestions helpful or annoying. Another paper that writes about a robot giving feedback to humans is the paper “Inspector Baxter: The Social Aspects of Integrating a Robot as a Quality Inspector in an Assembly Line”[6]. However this paper focusses on the influence movement and facial expressions can have on the social interaction.
The paper “The Social Impact of a Robot Co-Worker in Industrial Settings”[3], gives a discussion of the social phenomena that emerge when a robot co-worker is introduced in an industrial environment. It does this by discussing the relationship that each employee developed with the robot. In the end it suggests some improvements for future robot implementations that would improve further social interactions. This article relates to our problem in that it discusses the problem but in a broader sense. I.e. not focusing on a specific solution. An paper called “Ripple effects of an embedded social agent: A field study of a social robot in the workplace”[4] also discusses these social phenomenon that emerge in a human-robot cooperative industrial environment. However the robots in this paper perform the same basic tasks as the robots in a warehouse. I.e. picking up goods and delivering them to people.
One paper that we found tries to show the importance of movement in the social interaction between human-robot interaction. As cited “they [robots] also need the ability to model and reason about human activities, preferences and conventions. This knowledge is fundamental for robots to smoothly blend their motions, tasks and schedules into the workflows and daily routines of people. We believe that this ability is key in the attempt to build socially acceptable robots for many domestic and service applications”[5] This relates to our work, because it indicates a variable we need to keep track off.
[1] Damacharla, P., Dhakal, P., Stumbo, S., Javaid, A., Ganapathy, S., Malek, D., . . . Devabhaktuni, V. (2019). Effects of voice-based synthetic assistant on performance of emergency care provider in training. International Journal of Artificial Intelligence in Education : Official Journal of the International Aied Society,29(1), 122-143. doi:10.1007/s40593-018-0166-3 [2] Rosenfeld, A., Agmon, N., Maksimov, O., & Kraus, S. (2017). Intelligent agent supporting human-multi-robot team collaboration. Artificial Intelligence, 252, 211-231. doi:10.1016/j.artint.2017.08.005. [3] Allison Saupp´e, Bilge Mutlu, The Social Impact of a Robot Co-Worker in Industrial Settings, Department of Computer Sciences, University of Wisconsin
[4] Min Kyung Lee, Sara Kiesler, Jodi Forlizzi, Paul Rybski, Ripple Effects of an Embedded Social Agent: A Field Study of a Social Robot in the Workplace, Human-Computer Interaction Institute , Robotics Institute, Carnegie Mellon University, Pittsburgh USA
[5] Gian Diego Tipaldi, Kai O. Arras, Planning Problems for Social Robots, Social Robotics Lab Albert-Ludwigs-University of Freiburg [6] Amy Banh , Daniel J. Rea , James E. Young , Ehud Sharlin, Inspector Baxter: The Social Aspects of Integrating a Robot as a Quality Inspector in an Assembly Line, Proceedings of the 3rd International Conference on Human-Agent Interaction, October 21-24, 2015, Daegu, Kyungpook, Republic of Korea
Research questions and hypothesis
Based on the work already done it is possible to formulate some research questions. The main question will be: “Do vocal queues increase the likeability of a warehouse robot, compared to the use of a so called pick list.” The reason that this is the main research question, is that it covers every problem discussed in the introduction. If people like the robot more, then they are less likely to feel anxious about it.
Since we are able to borrow the robot, and doing the same experiment with visual queues instead of vocal queues. We decided that we would do both. For this reason the second research question is defined as follows: “Do visual queues increase the likeability of a warehouse robot compared to the use of a so called pick list”.
If the answer to either of these two questions is true, then we can look at which method is better. Which is why the third research question is the following: “Do vocal queues increase the likeability of a warehouse robot more than visual queues?”
Now that we have defined some research questions, we can define a hypothesis for the experiment that will be explained a little bit further on. Because of the paper “Effects of voice-based synthetic assistant on performance of emergency care provider in training”[1]. We predict that “Vocal queues increase the likeability of a warehouse robot.” Because of the research Margot Neggers has done, we also predict that “Visual queues increase the likeability of a warehouse.” With this, we can formulate the following hypothesis for the experiment.
Hypothesis: Adding a vocal queue or a visual queue to a robot in a warehouse setting will increase the likeability people feel towards it; compared to the likeability they have towards the same robot without these features.
Health & Safety Plan
During the experiment, it is essential to keep the participants safe. Since we are working with a robot which is around 50kg, there could be a chance that heavy bruising could take place. To avoid this certain safety measures have to be taken:
• There is a clear line on the floor that the test subject may not pass once the robot is moving. Only until the robot has a complete stand still the test subject may cross the line. The line is not directly in the robot’s path so the chances of crossing the line are diminished.
• There is always one person responsible for the safety measures taken. This person task is to have access to an external “Kill-switch”. This person solely focusses on the experiment and pressing the button. Once the button is pressed it will immediately turn off the robot.
Questionnaires
During the experiment, the test subjects were given a couple of questionnaires. This was done as follows:
1. Fill in intake Questionnaire
2. Scenario A
3. Fill in Trial Questionnaire
4. Scenario B
5. Fill in Trial Questionnaire
6. Scenario C
7. Fill in Trial Questionnaire
8. Ending Short Interview
The test subjects would fill in a general intake questionnaire to ask general questions about themselves; such as age, education, etc. Then they would start with one of the scenarios, this could be the list, voice, or LEDs. Then they would fill in the trial questionnaire. The trial questionnaire consisted of a combination of two validated questionnaires: the Co-existence questionnaire similar to that used in the PEPPER GAME (needs to be sourced) and the Godspeed questionnaire. The Godspeed questionnaire gives a better insight on what people think of the robot. After the test subjects repeated each scenario they got another trial questionnaire. After finishing all three questionnaires, a short interview was conducted. This interview was done to see if additional interesting observations could be made.
A link to the Questionnaires and Interview can be found at Questionnaires Link
Procedure Experiment
Context
In order to test our hypothesis, an industrial-like robot (FAST platform) was used to see how people would react. The robot drove packages to a person who would place the packages in the correct box. The idea of the experiment is modeled after the warehouse robots used at Amazon.
Procedure
This experiment is conducted in different parts. First, the setup of the experiment is described. Secondly, the steps taken in the experiment shall be explained. Thirdly, the separate scenarios (picklist, voice, LED) that have been tested shall be described. And finally, how the theoretical experiment happened in practice.
Setup
Firstly, the experiment was setup as can be seen in Figure. Test subject X is placing numbered packages in the corresponding boxes (A, B, C, D). The robot Y hands over these packages. Out of sight, a researcher Z places packages on the robot. This is done out of sight to simulate an automated warehouse.
Figure 1 Schematic drawing of the experiment
Steps
There were three different scenarios that we tested: picklist, voice, LEDs. These scenarios were tested with each test subject in three trails. In order to keep it fair the experiment was counterbalanced for the different test subjects. For example, test subject 1 would first get a pick list, then experience a robot using voice, and finally a robot with LEDs. Test subject 2 would first get the voice, then the LEDS, and finally the Pick list. This can be done in six different combinations. Since, twelve test subjects were used, each possible order combination was done twice.
A test subject would enter the room and would be given the intake questionnaire as described in the subsection Questionnaire. After they had filled this out they were given half a page of information on what was going to happen; it can be found as information sheet. Once they thoroughly read the information sheet the experiment would start. The person would not be told which scenario would take place. The test subject would wait behind the safety line and the robot would drive to the table. Once the robot stood still the scenario would happen. This would be repeated three times. Then, the test subject would fill in a questionnaire also described in the subsection Questionnaire. This is done by again by the two other scenarios.
Once all three scenarios and all trails within the scenarios have been completed, a final interview is conducted.
Scenario’s
In one scenario the test subject (x) has only the pick-list it received at the start to help him/her with finding the right packets for the right box. An example of the pick list can be found here list. In an other scenario the robot (y) will use led lights at the side to indicate (using color) in which box each packet goes. It does this by lighting up the corners at which the packet is located in the colors corresponding to the correct boxes. The order of the colored boxes can be seen here: LED script In the third scenario the robot (y) will use audio communication to indicate which packet goes into which box. It could say, using the speakers, “Packet 138 goes into box blue”. The human (x) can then place this packet in the corresponding box. The script of the voice can be found here Voice Script. After the human (x) feels it has finished its job, it will say “Done” and the robot will drive away.
Notes: The person doing the experiment should not be (color) blind, or deaf. Since they would not be able to do all scenarios. The test will be done in "wizard of oz"style. The robot will be remote controlled by us. Since the experiment is about which scenario the subjects find most useful/comforting, and not about testing the capabilities of the robot. Each subject only receives one scenario. Each subject does the experiment three times, each time the number of packets meant for that person will differ.
Comparison to real Experiment
The experiment was adjusted slightly in practice, as can be seen in Figure2 compared to Figure1. The table is rotated 90 degrees. But remains unseen from researcher Z on the right in Figure 2. Test Subject X is on the left waiting behind the line until the robot Y stops. The colored boxes can be clearly seen on the left and the packages are on the robot.
Figure 2 Actual situation during the experiment
Method
Figure 2: A drawing of the setup to clarify the explanation given below.
With this method we are trying to see if there is a difference in the interaction with a
warehouse robot when the robot has;
•no communication,
•communicates with lights
•or communicates with voice.
The goal is to get a deeper understanding of human robot interaction focused on voice cues. To do so we will create a scenario were testers have interaction that represents the interaction between man and robot in warehouses.
Definitions
Packets: each packet has a unique number to identify it with; just like in a warehouse. Boxes: each box has a unique color and needs to be filled with certain packets. Subject: The person taking part in the experiment. Denoted by x in figure 2.
Any references to (A), (B), (C), (D), (x), (y), (z) are a reference to their corresponding drawing in figure 2. Thus (y) is a reference to the robot. Experimental setup The robot (y) will start at (z), there researcher (z) will place a packet on each corner of the robot, thus in total 4. An x number of these 4 packets need to be placed by human test subject (x) in the boxes (A) through (D). Test subject (x) has received a list with packet numbers, and a corresponding box color. For example In the list below, the packet with ID 138 needs to go into the blue Box.
Experiment
The robot will drive to the the table containing the boxes (A), (B), (C) and (D). There are three different scenarios. In each scenario the robot (y) uses a different form of communication to aid the test subject (x) in placing the correct packets into the correct box.
Procedure Plan for Team(draft)
Test subjectX is located in front of a table with 4 boxes colored pink, yellow, blue and white. The robot will start every trial at location z. location z is hidden from the view of subject. In location x the team will have the packages they will put on the robot.
Figure(X): Schematic robot
A list of possible combinations will be created, per trial different combinations will be used.
Example: E 38, F 39, G 40, H 41
Results
Observations From the Experiment
During the experiment, most of the team members were present. This allowed for different observations that did not appear in the questionnaires or the interviews to be made that could not only be interesting for our discussion but could also influence the final answer to our research question. Further investigation of some of these observations could also lead to some interesting new findings about robot-human interaction.
Firstly, the subjects saw a clear difference in the perceived fluidness of motion by the robot if the robot used voice or light signals to communicate.(2.6 to 3.5) Where in all cases the robot was controlled in the same way.
Also, if the lights where the first of the scenarios there was notable confusion for some of the subjects. And in two cases they even waited on a second "go ahead" signal from the observers (team members) before starting the trial.
Sometimes with light signals the subjects handle the boxes with significantly less care then when compared to the list and the voice signals or some subjects took two boxes at the time resulting in confusion as to which package should go into which box.
Most subjects did not notice that the robot was remotely controlled but instead assumed that it was completely autonomous.
Finally, subjects that said they had experience with robotics assigned higher anthropomorphic values to the robot then those that said they had no experience with robotics.
Discussion of Results
Because of the use of the questionnaires after every set of trials in the experiment and the fact that the team was close to the participants during the trials there was a large number of data collected. From analyzing this data, that was seen in previous sections, we can reach some results and analyzing them along with our observations from the experiment we can reach a conclusion to answer our research question. Not all results are directly relevant in answering the research question but help to understand the thought process of someone that would be in a scenario that we are examining.
One of the most important results which was also one of the easiest to collect and process, as it was quantitative, was the amount of time it took for the participants to place the packages in the correct boxes. The speed at which someone completes task not only reflects how quick each robot-human interaction method is to communicate the placement information to the participant but also is a good indication of how understandable and clear the method itself is. The results showed that using vocal queues is clearly the slowest, using the list is slightly faster and finally using the LEDs is the fastest of all three. In comparison with the voice, LEDs need 54% of the time the voice needs to convey the same amount of information. The explanation for this is fairly straightforward. The LEDs not only were conveying the correct information instantly without needing to create a whole sentence but they also could inform the participant for the correct package placement for all four boxes simultaneously. This level of speed, was not possible with the voice as the robot had to make individual sentences for every single box. That doesn’t necessarily make the LEDs the best choice though. As was seen in the observation section, some participants, that had the LEDs in their first set of trials, were notably confused as to what they should do and needed extra assistance from the team members even though the procedure was explained to them before the experiment. This was not observed when the list or the voice were used for interaction. Also, another observation regarding the LEDS in the experiments was that people were more aggressive with their package placement as participants tried to be as quick as possible in their task. That behavior sometimes led to participants grabbing two packages simultaneously and then being confused as to where to place each. Of course, there is trade off of accuracy for speed which is not always favorable and in a warehouse scenario could come down to the priorities of the warehouse owners.
The results from the personal evaluation section of the questionnaires are really not what was expected. The participants found the LEDs more clear, pleasant and calmer. This could be for a variety of reasons. In terms of clarity, the voice could have been hard to hear or understand for the participants. . This of course could be the case because a fairly robotic voice (as it can be seen from the videos) was chosen for the robot and because the volume of the speakers was not maximized. That would mean that in a warehouse scenario the voice of the robot would have to be both loud but also clear enough to communicate with the employees without needing repetitions since that would seriously hinder efficiency. It is important to note though that no participant ask for the robot to repeat itself or to speak louder. Pleasantness and calmness on the other hand could be explained with the way that people viewed the robot itself. It was shown by the “Perceived Intelligence” questions that some participants viewed the robot as an authority (responsible) that was commanding them to do something and not as a co-worker that was just telling them what to do because it couldn’t do it. That may have made participants feel stressed out and generally have negative emotions for the robot which was in a sense “looking over their shoulder” to see their work output thus resulting in them thinking it was not calm and unpleasant. In general, other parts of the questionnaires reflected that people did not have the positive reaction to the voice that was expected.
As it can be seen from the questionnaires, there were questions on how safe the participants felt during and after the experiment. From these questions we can see how comfortable and certain participants around the robot. Results show that participants seemed to feel safe and relaxed in all scenarios. However LEDs seem to stand out as more calmer, relaxed and safer. This difference between the LEDs and the voice could be because of the feeling of authority that some participants felt while working the robot but also because, as we can see from the results of our “Animacy” questions, people thought of the voice as the more alive, thus resulting in them anthropomorphizing the robot and feeling uneasy around it.
The fact the participants anthropomorphized the robot more when voice was used can also be seen from the results of the “Anthropomorphism” questions. There, it can be seen that even though LEDs were considered more natural by participants, the voice was considered more conscious and in general the voice was rated as more anthropomorphic. This difference in results could be caused by the robotic nature of the voice chosen. Even though the voice was not like lifelike and natural, people still attributed signs of conscious thought to it. This could mean that people don’t need signs of nature to think of something as conscious but that is something that should be furthered investigated.
In general, participants ranked the voice as the most alive, lifelike and lively as it can be seen from the results of the “Amimacy” questions. This was on par with what was expected as the voice is closer to the forms of communication we have as humans. That result though, did not result in the voice getting ranked as the most liked robot-human interaction technique, getting passed by the LEDs, which was going against our hypothesis. This is mostly because of reasons that have already been mentioned. Namely, the slow speed at which the voice was talking frustrated the participants who wanted to put the packages in the boxes, the fact that the robot using the voice seemed like an authority and also the fact that people felt uneasy and unsafe while working with a speaking robot.
Finally, as indicated by the results in the “Perceived Intelligence” questions, participants ranked the voice and the LEDs equally. Looking into the subcategories specifically we can see that as mentioned before, people found the voice more responsible but also more intelligent and competent. Especially the later, thinking that the robot is competent, could result in the same problems this research is trying to solve, namely, people overestimating the capabilities of the robot.
Conclusion
SotA
Pim
The Human Side of Robotics: How Worker's React to a Robot http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.462.3947&rep=rep1&type=pdf
Short Summary: An research from 30 years ago that should give a good insight on how expectations match what actually happened. Focusing on the working with robots and the dangers that emerge from it being unsafe.
Toward Safe Close-Proximity Human-Robot Interaction with Standard Industrial Robots
https://dspace.mit.edu/handle/1721.1/106035#files-area
Short Summary: Increasing HRI in the industry will make people work in close-proximity to industrial robots. This article continues on how to make an Industrial robot better for close-proximity interaction.
Robot ethics: Mapping the issues for a mechanized world
https://ac.els-cdn.com/S0004370211000178/1-s2.0-S0004370211000178-main.pdf?_tid=4b18b191-4efd-4bf0-a9e3-526f16e94091&acdnat=1549713537_b2857f75deb01b6954139886023ad36cShort
Summary: This paper focusses on the ethics of robotics in mainly computers and industrialization and how it intervenes with human jobs
Robot Ethics: A View from the Philosophy of Science
short Summary: This paper mentions the factory workers working together with humans and collisions that arise (including robots hurting factory workers and the other way around)
Ethics by Design: A Conceptual Approach to Personal and Service Robot Systems
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.105.3665&rep=rep1&type=pdf
Short Summary: Robots in factories are different from what we should have on a day-to-day basis but working together on the same work floor states safety can become an issue. It explores factories a several other work-related robots.
ROBO-PARTNER: Seamless Human-Robot Cooperation for Intelligent, Flexible and Safe Operations in the Assembly Factories of the Future
Short Summary: This article takes an Industry related issue and tries to solve it. Explaining how effective robot-human interaction takes place.
Pepper Game https://pure.tue.nl/ws/portalfiles/portal/109006948/Scheffer_Tim_Report_Pepper_Game_Project_V1.2.pdf
Luuk
Human-robot interaction in rescue robotics https://ieeexplore-ieee-org.dianus.libr.tue.nl/document/1291662
Short Summary: This paper gives an overview about how robots are present at multiple levels of a rescue process. They can help during the search, but also by giving information about an environment that is hard to access by normal people.
Human-Robot Interaction: Issues in the Design of Interfaces for Work in Distant Environments https://journals-sagepub-com.dianus.libr.tue.nl/doi/pdf/10.1177/154193120905300401
Short Summary: This paper talks about the importance of interfaces. They should be simple enough such that they are self-explanatory. However they should also cover all functionality.
Robots in the workplace: Threat or opportunity? https://search-proquest-com.dianus.libr.tue.nl/docview/1807429396
Short Summary: This paper talks about how robots are not as bad as you might think. They do not just cost jobs, but they also create them. And most jobs that are covered by robots, are simple but tedious jobs.
Robots and Machine Vision in Packaging https://doi.org/10.1108/eb004402
Short Summary: This paper talks about how robots help people in the workplace by doing tedious tasks. It describes the robots as “The new robot is a smart, fast, pick‐and‐place device which frees workers to do other tasks.”
People Meeting Robots in the Workplace https://ieeexplore-ieee-org.dianus.libr.tue.nl/stamp/stamp.jsp?tp=&arnumber=5481097
Short Summary: This paper mentions how robots can be used to interact with people. It mentions how it is important for robots to be seen as social. If it is seen as annoying people will not use it no matter how handy it can be. However it is seen as a social creature, people might look past its flaws and actually care for it.
HUMAN-ROBOT FACTORS: ROBOTS IN THE WORKPLACE https://journals-sagepub-com.dianus.libr.tue.nl/doi/pdf/10.1177/154193120605000902
Short Summary: This article describes how robots can help make the workplace better, by doing tedious tasks, it also describes what are important factors in robot acceptance. Such as their ability to do task that are to dangerous for humans, and about how important expectations about this robot are in acceptance.
Charlotte
Weiss, A., & Huber, A. (2016). User Experience of a Smart Factory Robot: Assembly Line Workers Demand Adaptive Robots. In AISB2016: Proceedings of the 5th International Symposium on New Frontiers in Human-Robot Interaction. https://arxiv.org/ftp/arxiv/papers/1606/1606.03846.pdf
Short Summary:
The goal was to find out what kinds of suggestions the assembly line workers – who actually use the new robotic system – propose in order to improve the human-robot interaction the cooperation with a robot that executes predefined working steps actually impedes the user in terms of flexibility and individual speed. that cooperative robots in a dynamic factory context have to adapt to their human co-workers by taking their individual working steps and speed into account.
Buchner, R., Wurhofer, D., Weiss, A., & Tscheligi, M. (2013). Robots in time: How user experience in human-robot interaction changes over time. In Proceedings of ICSR2013, pp. 138-147. https://link.springer.com/chapter/10.1007/978-3-319-02675-6_14
Short Summary:
Our results show an increasing positive UX towards the newly deployed robots with progressing time
Obrist, M., Reitberger, W., Wurhofer, D., Förster, F., Tscheligi, M.: User experience research in the semiconductor factory: A contradiction? In: Campos, P., Graham, N., Jorge, J., Nunes, N., Palanque, P., Winckler, M. (eds.) INTERACT 2011, Part IV. LNCS, vol. 6949, pp. 144–151. Springer, Heidelberg (2011) https://link.springer.com/content/pdf/10.1007%2F978-3-642-23768-3_12.pdf
Short Summary:
(1) Investigate user experience of workers within the factory context. (2) Apply a creative approach, inspired by probing, which is applicable for this context in order to investigate workers’ experiences. on the interaction and user experience (UX) in factories, The results showed that the absence of stress significantly contributes to a perfect working day and that the end of a shift is experienced as the most stressful part of the daily working routine, as this is the most critical point of time for a fluent working process.
Brogårdh, T. (2007). Present and future robot control development—An industrial perspective. Annual Reviews in Control, 31(1), 69-79. https://www.sciencedirect.com/science/article/pii/S1367578807000077
Short Summary:
One scenario is that light-weight robot concepts could have an impact on future car manufacturing and on future automation of small and medium size enterprises (SMEs). Such a development could result in modular robots and in control schemes using sensors in the robot arm structure, sensors that could also be used for the implementation of redundant safe control. Introducing highly modular robots will increase the need of robot installation support,
Sauppé, A., & Mutlu, B. (2015, April). The social impact of a robot co-worker in industrial settings. In Proceedings of the 33rd annual ACM conference on human factors in computing systems (pp. 3613-3622). ACM. https://dl.acm.org/citation.cfm?id=2702181
Short Summary:
workers relate to the robot as a social entity and rely on cues to understand the robot's actions, which we observed to be critical for workers to feel safe when near the robot.
Orendt E.M., Henrich D. (2018) An Architecture for Intuitive Programming and Robust Execution of Industrial Robot Programs. In: Schüppstuhl T., Tracht K., Franke J. (eds) Tagungsband des 3. Kongresses Montage Handhabung Industrieroboter. Springer Vieweg, Berlin, Heidelberg
Short Summary:
Intuitive robot programming and robust task execution. Our architecture enables users to create robot programs by guiding a robot kinesthetically through tasks.
(Dauth, W., Findeisen, S., Südekum, J., & Woessner, N. (2017). German robots-the impact of industrial robots on workers.)
Timo
Robots in Society, Society in Robots
https://link.springer.com/article/10.1007/s12369-010-0066-7
An analysis of human robot interaction and how they are influenced by social and cultural factors. And describes a range of methodologies and design that support a socially robust understanding of technological development/robots
Robot ethics, The ethical and Social Implications of Robotics
An extensive ethical analysis of most branches of robotics. States that human-robot interaction would benefit greatly if robots could express emotions to humans and read emotions shown by humans.
Artificial intelligence and robotics and their impact on the workplace
100+ pages of the effects of robots in the private sector containing: new forms of employment: an analysis of new employment possibilities created by automatization health and safety issues: The interaction between employees and robots and how this will need to change The Impact of New Technology on the Labour Market: A analysis of the jobs that are endangered by automatization and jobs that become more important because of automatization.
Robot ethics: A view from the philosophy of science
This paper analyses the effect of the need for a reduced work force (by automatization). And in particular how people outside the workforce for a lengthy period of time are going to get social benefits as these are usually obtained though employment.
Experimental investigation into influence of negative attitudes toward robots on human–robot interaction
http://rins.st.ryukoku.ac.jp/~nomura/docs/FormalPaper.pdf
A paper exploring the correlation between negative attitudes towards robots and their behavior toward robots. Where they found a gender difference within both negative attitudes towards robots and in the relation between negative attitudes towards robots and their behavior toward robots.
Robots in Time: How User Experience in Human-Robot Interaction Changes over Time
https://link.springer.com/chapter/10.1007/978-3-319-02675-6_14
An analysis of how employees experience their robotic colleagues in a study over time. (week, six months and 1.5 years after implementation of the robot)