PRE2018 3 Group11: Difference between revisions

From Control Systems Technology Group
Jump to navigation Jump to search
Line 305: Line 305:
# A Motion User Interface. Can interpret motion gestures.
# A Motion User Interface. Can interpret motion gestures.


== Distance Estimation==
Since a camera can be considered to be a part of a drone, we present a formula for distance computation, given some parameters. The only assumption made is that the type of lens is '''rectilinear'''. The explained formula does not hold for other types of lenses.


An integral part of the operation of a drone is distance computation and estimation.  
Since <math> D \cdot \phi = H_{obj} </math> and <math> f \cdot \phi = h_{obj} </math> hold, we can factor <math>\phi</math> out and get <math> D = \frac{ f \cdot H_{obj} }{ h_{obj} }</math>. After rewriting <math>h_{obj}</math> to be in terms of mm, we finally end up with:


[[File:distance_estimation_diagram.png| center]]
<math>
D(mm) = \frac { f(mm) \cdot H_{obj}(mm)} { \frac{ h_{obj}(px) } { h_{sensor}(px) } \cdot h_{sensor}(mm) }
</math>


Since a camera can be considered to be a part of a drone, we present a formula for distance computation, given some parameters. The only assumption made is that the type of lens is '''rectilinear'''. The explained formula does not hold for other types of lenses.
The parameters are as follows:
# <math> D(mm) </math>: The real-world distance between the drone and the user in millimeters. This is what is computed.
# <math> f(mm) </math>: The actual focal length of the camera in millimeters. If you see a ''35mm equivalent'' focal length, then refrain from using it. If you simply see two values, a quick test will show you which one should be used.
# <math> H_{obj}(mm) </math>: The real-world height of the user in millimeters.
# <math> h_{obj}(px) </math>: The height of the user in pixels.  
# <math> h_{sensor}(px) </math>: The height of the sensor in pixels, in other words, the vertical resolution of the video file (e.g. for 1080p videos this would be 1080, for 480p this would be 480 and so on).
# <math> h_{sensor}(mm) </math>: The sensor size in millimeters, which is the pixel size in millimeters multiplied by <math> h_{obj}(px) </math>.


[[File:d_equation_0.png| 600px| center]]
All these parameters should be present on the spec-sheet of a well documented camera, such as the OnePlus 6 (which is a phone that was used for shooting our own footage).  


<!-- <math> H_{obj} = D \cdot \frac{ h_{obj}(px)}{h_{sensor}(px)} \cdot V_{FOV}(radians) </math> -->
When looking at the '''amount of error''' this method gives when implementing, we find that on scales less than 500mm, it is unfeasible. For a scale from 1000mm and up, it can be used. The results of our tests can be found in following table.


The parameters
{| class="wikitable"
# ''H<sub>obj</sub> : The (real-world) height of the object or user in a metric system.
|+ Distance Computation Evaluation
# ''D'' : The distance between the lens on the drone and an object or user in a metric system.
! Scale (mm)
# ''h<sub>obj</sub>(px)'' : The height of the object or user in pixels.
! Computed Distance (mm)
# ''h<sub>sensor</sub>(px)'' : The pixel-height of the sensor of the camera. This can be computed in terms of the sensor’s pixel-size (which should be present in the specification of the used lens) and resolution of the digital image (which should easily be retrievable) [http://www.ni.com/product-documentation/54616/en/].  
! Absolute Error (m, rounded)
# ''V<sub>FOV</sub>(radians)'' : The vertical ''field of view'' of the lens. This is denoted by ''&alpha;'' in the image above, and is either in the specification of the used lens, or can be computed using ''f'', the focal length and sensor size which both must be in said specification. It possibly needs a conversion to radians afterwards.
|-
| 100  || 7989.680302875845  || 7.89
|-
| 200  || 2955.9793551540088 || 2.76
|-
| 500  || 936.64000567      || 0.44
|-
| 1000 || 687.851437808      || 0.31
|-
| 2000 || 1692.22216373      || 0.30
|-
| 3000 || 3183.46590645      || 0.18
|-
| 4000 || 4143.14879882      || 0.14
|}


Rewriting the equation for ''D'', the distance gives the following.
If the drone is to be equipped with 2 cameras, computing the distance towards an object is possible using stereo vision as proposed by Manaf A. Mahammed, Amera I. Melhum and Faris A. Kochery, with an error rate of around 2% [http://warse.org/pdfs/2013/icctesp02.pdf]. This does, however, increase the number of cameras that need to be used, which is a further constraint on power.
<!-- <math> D = \frac{ H_{obj} }{ \frac{ h_{obj}(px)}{h_{sensor}(px)} \cdot V_{FOV}(radians) } </math> -->


[[File:d_equation_1.png| 600px| center]]
As the distance estimation described is to be used solely for user tracking purposes, and not for detecting and avoiding obstacles, we believe that an error rate of 3.5% ((0.14/4)*100) on a scale of 4m is acceptable.


== Obstacle detection ==
== Obstacle detection ==

Revision as of 15:38, 27 March 2019

<link rel=http://cstwiki.wtb.tue.nl/index.php?title=PRE2018_3_Group11&action=edit"stylesheet" type="text/css" href="theme.css"> <link href='https://fonts.googleapis.com/css?family=Roboto' rel='stylesheet'>

Organization

The group composition, deliverables, milestones, planning and task division can be found on the organization page.

Brainstorm

To explore possible subjects for this project, a brainstorm session was held. Out of the various ideas, the Follow-Me Drone was chosen to be our subject of focus.

Background

The ability to navigate around one’s environment is a difficult mental task whose complexity is often left unacknowledged by most of us. By difficult, we are not referring to tasks such as pointing north while blindfolded, but rather much simpler ones such as: following a route drawn on a map, recalling the path to one’s favourite cafe. Such tasks require complex brain activity and fall under the broad skill of ‘navigation’, which is present in humans in diverse skill levels. People whose navigation skills fall on the lower end of the figurative scale are slightly troubled in their everyday life; it might take them substantially longer than others to be able to find their way to work through memory alone; or they might get lost in new environments every now and then.

Navigation skills can be traced back to certain distinct cognitive processes (to a limited extent, nevertheless, as developments in neurosciences are still progressing [1]), which means that a lesion in specific brain regions can cause a human’s navigational abilities to decrease dramatically. This was found in certain case studies with patients that were subject to brain damage, and labeled “Topographical Disorientation” (TD), which can affect a broad range of navigational tasks [2]. Within the last decade, cases started to show up in which people showed similar symptoms to patients with TD, without having suffered any brain damage. This disorder in normally developed humans was termed “Developmental Topographical Disorientation” (DTD) [3], which, as was recently discovered, might affect 1-2% of the world's population (this is an estimate based on informal studies and surveys) [4] [5].

Navigational Map Reading

Navigational Map Reading is the term used by the geographer Amy Lobben of the University of Oregon in her 2007 article [6] to describe the set of spatial abilities humans use to navigate using a map. She identified three skills: 1) Self-location: the ability to determine the spot on a map where one's location in the real world is. We do this by observing our surroundings, e.g., roads, buildings, intersections and so on. 2) Route memory: the ability to remember the way. 3) Map rotation: Map rotation is the mental process of re-orienting or rotating a map as one changes orientation in the real world, without physically rotating the map.

We can identify the deficiency of DTD patients with the three aforementioned map-reading skills. For example, the deficiency of the two skills ‘self-location’ and ‘route memory’ is an evident symptom in DTD patients; the third identified skill ‘map rotation’ is an even further developed skill that depend on the first two.

Problem Statement

Many of the affected people cannot rely on map-based navigation and sometimes even get lost on everyday routes. To make their everyday life easier, one solution is to use a drone. The idea of the playfully-named “follow-me drone” is simple: By flying in front of its user, accurately guiding them to a desired location, the drone takes over the majority of a human’s cognitive wayfinding tasks. This makes it easy and safe for people to find their destination with minimized distraction and cognitive workload of navigational tasks.

The first design of the “follow-me drone” will be mainly aimed at severe cases of (D)TD, who are in need of this drone regularly, with the goal of giving those people some of their independence back. This means that the design incorporates features that make the drone more valuable to its user with repeated use and is not only thought-of as a product for rare or even one-time use as part of a service.

Examination of our user base

While we are well aware that the target user base is often not congruent with the real user base, we would still like to deliver a good understanding of the use cases our drone is designed for. Delivering this understanding is additionally important for future reference and the general iterative process of user centered design.

As clarified by our problem statement, we specifically designed this drone for people that for various reasons are not able to navigate through their daily environment without external help. This concerns mostly people that suffer a form of topographical disorientation. Since this form of mental disorder is still relatively new and does not have an easily identifiable cluster of symptoms, Aguirre & D’Esposito (1999) made an effort to clarify and classify cases of TD and DTD (whereas TD is caused by brain damage and DTD is the developmental form). To give a better understanding of the problems that affected people are suffering from, we listed the significant categories of topographical disorientation and a short description of their known symptoms and syndromes below:

  • Egocentric disorientation
Patients of this category are not able to perform a wide range of mental tasks connected to navigation. They might be able to point at an object they can see, however, they lose all sense of direction as soon as their eyes are closed. Their route-descriptions are inaccurate and they are not able to draw maps from their mind. Most patients are not able to point out which of two objects is closer to them. This results in often getting lost and generally poor navigation skills, which contributes to many of the affected people staying home and not going out without a companion. (Aguirre & D’Esposito, 1999)
  • Heading disorientation
Patients with heading disorientation are well able to recognize landmarks but can not derive any directional information of them. This might also just show up occasionally, as one patient describes: “I was returning home when suddenly, even though I could recognize the places in which I was walking and the various shops and cafés on the street, I could not continue on my way because I no longer knew which way to go, whether to go forward or backward, or to turn to the right or to the left...”. (Aguirre & D’Esposito, 1999)
  • Landmark agnosia
People affected by landmark agnosia are not able to use environmental features such as landmarks for their way-finding or orientation. As described by someone landmark agnostic: “I can draw you a plan of the roads from Cardiff to the Rhondda Valley... It's when I'm out that the trouble starts. My reason tells me I must be in a certain place and yet I don't recognize it.” (Aguirre & D’Esposito, 1999)

In all three cases above, we can identify the need for additional guidance in navigating through their environment in everyday life. Also different solution besides a drone may fulfill that need, more details about the viability of the drone and alternative solutions can be found in the section Alternative Solutions. However, our drone can act as a mentally failsafe help for people affected by topographical disorientation, since all of the mentioned cases showed an ability to recognize objects and point at them. This will give people who lost faith in their ability to freely and independently walk to a desired nearby location hope and security to do so again.

Approach

In order to get to a feasible design solution, we will do research and work out the following topics:

  • User requirements
    • Which daily-life tasks are affected by Topographical Disorientation and should be addressed by our design?
    • What limitations does the design have to take into account to meet the requirements of the specified user base?
  • Viability
    • What alternative solutions exist currently?
    • What are the pros and cons of the drone solution?
    • In which situations can a drone provide additional value?
  • Human technology interaction
    • What design factors influence users’ comfort and safety with these drones?
    • How can an intuitive and natural guiding experience be facilitated?
      • How can the salience of the drone be maximized in traffic?
      • What distances and trajectories of the drone will enhance the safety and satisfaction of users while being guided?
    • What kind of interface is available in which situations for the user to interact with the drone?
  • User tracking
    • How exact should the location of a user be tracked to be able to fulfill all other requirements of the design?
    • How will the tracking be implemented?
  • Positioning
    • How to practically implement findings about optimal positions and trajectories?
  • Obstacle avoidance
    • What kind of obstacle avoidance approaches for the drone seem feasible, given the limited time and resources of the project?
    • How to implement a solution for obstacle avoidance?
  • Special circumstances
    • What limitations does the drone have in regards to weather (rain, wind)?
    • How well can the drone perform at night (limited visibility)?
  • Physical design
    • What size and weight limitations does the design have to adhere to?
    • What sensors are needed?
    • Which actuators are needed?
  • Simulation
    • How could the design be implemented on a real drone?
    • How well does the implemented design perform?
    • How can the internal workings of the drone be visualized?

Limitations

There are certainly some limitations in the manufacturing and usage of the follow-me drone. These limitations are the following:

  • Hazard of malfunction. If the drone malfunctions, by making a wrong move, making a miscalculation, failing to avoid an obstacle, or makes a wrong maneuver based on a false positive obstacle, then it could potentially harm the user or the traffic. This poses more significance on the safety-critical aspect of its design and implementation, which adds more cost thereon.
  • Distracting the user or the traffic increases the risk of accidents.
  • Negative impact on the environment. A flying drone poses unnecessary usage of energy that could be avoided through more energy-efficient solutions. Moreover, it uses materials in its body, batteries and solar cells, if any, that are hard to recycle such as Nickel and Cadmium and Polypropylene. This merely poses extra cost on the recycling process of such drone that has to be taken into account with respect to the benefit harnessed from it which is limited due to the small targeted user group.
  • Weather conditions

Alternative Solutions

  • Google Maps

Google Maps has announced and demonstrated a new feature that works in tandem with Street View [7][8]. Now the user will be able to navigate streets with a live view on his phone’s screen overlaid with navigational instructions. The user can walk around streets pointing their camera towards buildings and receiving information about those buildings. This is perfect for DTD patients because they have trouble recognizing landmarks. On top of that, it provides personalized suggestions, so the user can essentially get a notification once they point their phone at the direction of a building what this building is and why it is suggested. The overlay becomes more effective when combined with an earlier project of Google discussed below, Google Glass, especially for DTD patients.

  • Smart Glasses

Multiple projects aim to overlay a display on glasses and most of them focus particularly on feeding the user with navigational information. These projects are Google Glass [9], a Japanese GPS glass [10][11], and a research prototype glass for the blind [12]. This last project is particularly helpful for the visually-impaired, but its technology is just as much useful for DTD patients. That is because the glasses have a pair of cameras that can see the world and recognize the surroundings. This helps anyone who gets lost due to not recognizing surroundings. More advanced glasses are virtual reality ones such as Microsoft HoloLens [13]. This not only lay a 2D picture on your glasses, but they use advanced technology that tracks the movement of pupils to adjust the 2D overlay such that it appears 3-Dimensional. These technologies can be combined to deliver a virtual reality navigation solution for everyone including the visually impaired and the DTD.

In the Maps app on iPhone, the user can fly over many of the world's major cities and landmarks in full-fledged 3D world.

3D Navigation is an alternative technology that can be used instead of the Follow-Me Drone. TomTom 3D Navigation is already used in new generation Peugeot cars. Apple has a Flyover feature that gives you 3D views of a certain area. While it does not act as a navigator, it lays a foundation for the development of 3D mobile navigation.

“WRLD 3D maps provide a simple path to create virtual worlds based on real-world environments”.

Tomtom and WRLD 3D both provide platforms for further app development. That means that even if Google and Apple do not provide APIs to the backends of their respective technologies, then Tomtom and WRLD 3D platforms would facilitate that functionality.

Why these alternatives work

The reason why these technologies are believed to solve DTD problems is clear when dissecting the three navigational map-reading skills which the 2007 Amy Lobben article identified. The three skills are simple: 1) Self-location: This is a central feature of the mentioned technologies that uses GPS geolocation. 2) Route memory: With navigation apps, the user does not need to remember or deduce routes. Furthermore, some of these technologies can identify buildings and landmarks using image recognition combined with GPS. 3) Map rotation: These apps automatically detect the user’s orientation and they rotate the map accordingly.

When is a drone better or worse than a phone app

Since the most central feature of a phone app is geolocation using GPS signals, a drone would work better in environments where GPS signals cannot reach, e.g. underwater or underground. In 2017, two divers lost their way inside an underwater cave system [14]. A drone able to measure its location displacement in a 3D space without relying on GPS [15] would have been able to guide those divers out the same way the came in, or it can even be used to conduct a search and rescue operation. The same applies to underground caves, catacombs, or mines. A similar 2017 incident occurred in the Paris catacombs. Two teenage boys lost their way for three days in the catacombs with no phone signal or GPS signal [16]. Since the Paris catacombs are a tourist attraction visited by all types of people, such a hazard is not unlikely. In such a situation, a drone accompanying the boys would have guided them out. Technologies such as Doppler Velocity Loggers, Inertial Navigation Systems and Acoustic Sensing [17] are recent developments meant for autonomous vehicles to navigate when GPS signals are absent, and in some cases radio signals and even visuals.

On the other hand, a drone serves worse than a phone app in the following situations: in crowded areas, in traffic, in buildings with doors and other impediments, and in trips that are longer than the drone's battery life which is typically 30 minutes for a flying drone.

A case against hand-held devices

Holding a phone in hand while navigating is not an ideal solution, for as simple a reason as not having a free hand, or needing one's attention on the road or surroundings to avoid possible accidents. The solution to this can indeed be a flying drone guiding the user. Yet an even better solution is the glass overlay mentioned in Alternative Solutions, or any other hands-free display. The follow-me drone itself can act as a hands-free device with its own display or even holding the user's phone and utilizing it as a display.

Divide and Conquer

In fact, the hands-free display is a separate problem from the navigation app. once we make a hands-free display, then the navigation app can be displayed on it, thereby splitting the work into two independently developed modules. Needless to say, this more modular approach is more efficient from a developmental perspective and even from a commercial perspective.

Solution

Requirements

Aside from getting lost in extremely familiar surroundings, it appears that individuals affected by DTD do not consistently differ from the healthy population in their general cognitive capacity. Since we can't reach the targeted group, it is difficult to define a set of special requirements.

  1. The Follow-Me Drone has a navigation system that computes the optimal route from any starting point to the destination.
  2. The drone must guide the user to his destination by taking the decided route hovering in direct line of sight of the user.
  3. The drone must guide the user without the need to display a map.
  4. The drone must provide an interface to the user that allows the user to specify their destination without the need to interact with a map.
  5. The drone announces when the destination is reached.
  6. The drone must fly in an urban environment.
  7. The drone must keep a minimum distance of two? meters to the user.
  8. The drone must not go out of line of sight of the user.
  9. The drone must know how to take turns and maneuver around obstacles and avoid hitting them.
  10. The drone should be operable for at least an hour between recharging periods back at home.
  11. The drone should be able to fly.
  12. (Optional, if the drone should guide a biker) The drone should be able to maintain speeds of 10m/s.
  13. The drone should not weigh more than 4 kilograms.

Human factors

Drones are finding more and more applications around our globe and are becoming increasingly available and used on the mass market. In various applications of drones, an interaction with humans who might or might not be involved in the task that a drone carries out, is inevitable and sometimes even the core of the application. Even though drones are still widely perceived as dangerous or annoying, there is a common belief that they will get more socially accepted over time [18]. However, since the technology and therefore the research on human-drone interaction is still relatively new, our design should incorporate human factors without assuming general social acceptance of drones.

The next sections will focus on different human factors that will influence our drone’s design with different goals. This will include the drone’s physical appearance and its users’ comfort with it, safety with respect to its physical design, as well as functionality-affecting factors which might contribute to a human's ability to follow the drone.

Physical appearance

The physical design of the drone depends on different factors. It is firstly defined and limited by technical requirements, as for example the battery size and weight, the size and number of propellers and motors and the system’s capacity for carrying additional weight (all discussed in Technical considerations). As soon as the basic functionality is given, the safety of the drone’s users has to be considered in the design. This is done in the section Safety. Taking all these requirements into account, further design-influencing human factors can now be discussed. However, the design process is not hierarchical, but rather iterative, which means that all mentioned factors can influence each other.

  • Drone size
For a human to feel comfortable using a drone as a guiding assistant, the size will play an important role. Especially in landing or storing processes a drone of a big size will come in unwieldy, whereas a small drone might be more difficult to detect during the process of following it. In the study conducted in 2017 by Romell and Karjalainen [19], it was investigated what size of a drone people prefer for having a drone companion. The definition of a drone companion in this study lies close enough to our follow-me drone to consider the results of this study for our case. Given the choice between a small, medium or large drone, the majority of participants opted for the medium size drone, which was approximated at a diameter of 40cm. Giving priority to the technical limitations and the drone’s optimal functionality, we took this result into account.
  • Drone shape
In the mentioned 2017 study by Romell and Karjalainen [20], also the optimal shape for such a drone companion was evaluated. It came to the conclusion that particularly round design features are preferred, which we considered as well in our design process.
Additionally, while the drone is not flying, a user should intuitively be able to pick up and hold the drone, without being afraid to hurt him/herself or damage the drone. Therefore a sturdy frame or a visible handle should be included.
  • Drone color
Since the color of the drone greatly contributes to the visibility of it, and is therefore crucial for the correct functioning of the system, the choice of color is discussed in the section Salience in traffic.

Safety

The safety section considers all design choices that have to be made to ensure that no humans will be endangered by the drone. These choices are divided into factors that influence the physical aspects of the design, and factors which concern dynamic features of the functionality.

  • Safety of the physical design
The biggest hazard of most drones are accidents caused by objects getting into a running propeller, which can damage the object (or injure a human), damage the drone, or both. Propellers need as much free air flow as possible directly above and below to function properly, which introduces some restrictions in creating a mechanical design that can hinder objects from entering the circle of the moving propeller. But most common drone designs involve an enclosure around the propeller (the turning axis), which is a design step that we also follow for safety reasons.
Other than the propellers, we are not aware of another evident safety hazard that can be solved by changes in the physical design of the drone.
  • Safety aspects during operation
Besides hazards that the drone is directly responsible for, like injuring a human with its propellers, there are also indirect hazards to be considered that can be avoided or minimized by a good design.
The first hazard of this kind that we identified, is a scenario where the drone leads its user into a dangerous or fatal situation in traffic. As an example, if the drone is unaware of a red traffic light and keeps on leading the way crossing the trafficked road, and the user shows an overtrust into the drone and follows it without obeying traffic rules, this could result in an accident.
Since this design will not incorporate real-time traffic monitoring, we have to assure that our users are aware of their unchanged responsibility as traffic participants when using our drone. However, for future designs, data from online map services could be used to identify when crossing a road is necessary and subsequently create a scheme that the drone follows before leading a user onto a trafficked road.
Another hazard we identified concerns all traffic participants besides the user. There is a danger in the drone distracting other humans on the road and thereby causing an accident. This assumption is mainly due to drones not being widely applied in traffic situations (yet), which makes a flying object such as our drone, a noticable and uncommon sight. However, the responsibility lies with every conscious participant in traffic, which is why we will not let our design be influenced by this hazard greatly. The drone will be designed to be as easily visible and conspicuous for its user, while keeping the attracted attention from other road users to a minimum.

Salience in traffic

As our standard use case described, the drone is aimed mostly at application in traffic. It should be able to safely navigate its user to her destination, no matter how many other vehicles, pedestrians or further distractions are around. A user should never get the feeling to be lost while being guided by our drone.

To reach this requirement, the most evident solution is constant visibility of the drone. Since human attention is easily distracted, also conspicuity, which describes the property of getting detected or noticed, is an important factor for users to find the drone quickly and conveniently when lost out of sight for a brief moment. However, the conspicuity of objects is perceived similarly by (almost) every human, which introduces the problem of unwillingly distracting other traffic participants with an overly conspicuous drone. We will focus on two factors of salience and conspicuity:

  • Color
The color characteristics of the drone can be a leading factor in increasing the overall visibility and conspicuity of the drone. The more salient the color scheme of the drone proves to be, the easier it will be for its user to detect it. Since the coloring of the drone alone would not emit any light itself, we assume that this design property does not highly influence the extent of distraction for other traffic participants.
Choosing a color does seem like a difficult step in a traffic environment, where certain colors have implicit, but clear meanings that many humans got accustomed to. We would not want the drone to be confused with any traffic lights or street signs, since that would lead to serious hazards. But we would still need to choose a color that is perceived as salient as possible. Stephen S. Solomon conducted a study about the colors of emergency vehicles [21]. The results of the study were based upon a finding about the human visual system: It is most sensitive to a specific band of colors, which involves ‘lime-yellow’. Therefore, as the study showed, lime-yellow emergency vehicles were involved in less traffic accidents, which does indeed let us draw conclusions about the conspicuity of the color.
This leads us to our consideration for lime-yellow as base color for the drone.
To increase contrast, similar to emergency vehicles, red is chosen as a contrast color. This is implemented by placing stripes or similar patterns on some well visible parts of the drone.
  • Luminosity
To further increase the visibility of the drone, light should be reflected and emitted from it. Our design of the drone will only guarantee functionality by daylight, since a lack of light drastically limits the object and user detection possibilities. But an emitting light source on the drone itself is still valuable in daylight situations.
We assume that a light installation which is visible from every possible angle of the drone should be comparable in units of emitted light to a standard LED bicycle lamp, which has about 100 lumen. As for the light color, we choose a color that is not preoccupied by any strict meaning in traffic and can be seen normally by color blind people. This makes the color blue the most rational choice.

Positioning in the visual field

A correct positioning of the drone will be extremely valuable for a user but due to individual preferences, there will be difficulties in deriving a default positioning. The drone should not deviate too much from its usual height and position within its users attentional field, however, it might have to avoid obstacles, wait, and make turns without confusing the user. According to a study of Wittmann et. al. (2006) [22], it is easier for the human visual-attentional system to track an arbitrarily moving object horizontally than it is vertically. For that reason, our drone should try to avoid objects with only slight movements in vertical direction, but is more free to change its position horizontally on similar grounds. Since the drone will therefore mostly stay at the same height while guiding a user, it makes it easier for the user to detect the drone, especially with repeated use. In a study that investigated drone motion to guide pedestrians [23], a small sample was stating their preference on the distance and height of a drone guiding them. The results showed a mean horizontal distance of 4 meters between the user and the drone as well as a preferred mean height of 2.6 meters of the drone. As these values can be individually configured for our drone, we see them merely as default values. We derived an approximate preferred viewing angle of 15 degrees, so we are able to adjust the default position of the drone to each users body height, which is entered upon first configuration of the drone.


Legal Issues

In this project, a flying drone is designed to move autonomously through urban environments to guide a person to their destination. It is important however to first look at relevant legal issues. In the Netherlands, laws regarding drones are not formulated in a clear way. There are two sets of laws: recreational-use laws and business-use laws. It is not explicitly clear which set of laws applies to this project if any at all. The drone is not entirely intended for recreational use because of its clear necessity to the targeted user group. On the other hand, it does not fall under business use either it is not “used by a company to make money”---this does not include selling it to users. In that case, it is in the users’ ownership and is no longer considered used by the company. If the drone is rented, however, it still might fall under business use. In conclusion, neither set of laws applies to our project. Since drones are not typically employed in society, this is not unexpected and we might expect new laws to adapt to cases such as ours.

According to law, drone users are not allowed to pilot drones above residential areas regardless of their use. Our drone, however, flies through residential areas for its intended purpose, not above them.

Legislation is subject to change that can be in accordance with our project as well as against it. That being said, we can safely design our drone as intended without worrying about legislation. Our design does not explicitly violate any laws.

Technical considerations

Battery life

One thing to take into consideration while developing a drone is the operating duration. Most commercially available drones can not fly longer than 30 minutes [24],[25]. This is not a problem for a walk to the supermarket around the block, but to commute for example to work longer battery life would be preferred. Furthermore wind and rain have a negative influence on the flight time of the drone but the drone should still be usable during these situations so it is important to improve battery life.

In order to keep the battery life as long as possible, a few measures could be taken. First of all a larger battery could be used. This however makes the drone heavier and thus also would require more power. Practical energy density of a lithium-polymer battery is 150Wh/kg [26]. However if the battery is discharged completely it will degrade, decreasing its lifespan. Therefore the battery should not be drained for more than 80% meaning that the usable energy density is about 120Wh/kg.

Another way to improve flight time is to keep spare battery packs. It could be an option in this project to have the user bring a second battery pack with them for the trip back. For this purpose the drone should go back to the user when the battery is nearly low and land. The user should then pick the drone up and swap the batteries. This would require the batteries to be easily (dis)connectable from the drone. A possible downside to this option is that the drone would temporarily shut down and might need some time to start back up after the new battery is inserted. This could be circumvented by using a small extra battery in the drone that is only used for the controller hardware.

For this project first it is researched whether a continuous flight of around 60 minutes is feasible. If this is not possible it is chosen to use the battery swapping alternative. Unfortunately no conclusive scientific research is found about the energy consumption of drones so instead existing commercially available drones are investigated. In particular the Blade Chroma Quadcopter Drone [27] is researched. It weighs 1.3kg and has a flight time of 30 minutes. It uses 4 brushless motors and has a 11.1V 6300mAh LiPo battery. The amount of energy can be calculated with Battery equation.png which gives about 70Wh. The battery weighs about 480g [28] so this seems about right with previously found energy density. The drone has a GPS system, a 4K 60FPS camera with gimball and a video link system. The high-end camera will be swapped for a lighter simpler camera that uses less energy since the user tracking and obstacle avoidance work fine with that. The video link system will also not be used. The drone will only have to communicate simple commands to the user interface. This does spare some battery usage but on the other hand a computer board will be needed that runs the user tracking, obstacle avoidance and path finding. For now it is assumed that the energy usage remains the same. Unfortunately the flight time of 30 minutes is calculated for ideal circumstances and without camera. The camera would decrease the flight time by about 7 minutes . For the influence of weather more research is needed but for now it is assumed that the practical flight time is 15-20 minutes.

If a flight time of 60 minutes is required that would mean that a battery at least 3 times as strong is needed. However since these battery packs weigh 480g, that would increase the drone weight almost by a factor of 2. This would in itself require so much more power that even more battery and motor power is required. Therefore it can be reasonably assumed that this is not a desirable option. Instead the user should bring 2 or 3 extra battery packs to swap these during longer trips.

It is however still possible to design a drone that has sufficient flight time on a single battery. Impossible aerospace claims to have designed a drone that has a flight time of 78 minutes with a payload [29]. To make something similar a large battery pack and stronger motors would be needed.

Hardware

The drone is controlled by an Raspberry Pi (Model 3??) which weighs 22g and requires 12.5W of energy.

For the drone to effectively use the octomap framework for the obstacle detection one needs at the very least a rotational lidar (which is a rangefinder), preferably one that can scan in 360 degrees. These lidars range in price from 100 to a 7000 euros. The Sweep V1 360° Laser Scanner [30] is a good example of a lidar that is not too expensive, and should be good enough for our purposes. It can scan up to 40 meters and does not require a lot of power. A simply 5V supply is needed and it draws a current of about 0.5A which means it requires 2.5W of energy. It weighs 120g.

The drone needs a camera to track the user. This camera does not have many requirements so it can be as simple and light as possible. That is why a camera module designed for use with Raspberry Pies is used. The Raspberry Pi Camera Module F [31] can take videos of resolutions up to 1080p which is more than needed for our project. Weight is not specified but other modules weigh about 3g so this module won't weigh more than 10g. It uses power from the arduino directly. 00g.

Motors and propellers, batterypack

To have a starting point which propellers and motors to use, the weight of the drone is assumed. Two cases are investigated. One is where the drone has flight time of 15-20 minutes and its battery needs to be swapped during the flight. In this case the drone is assumed to have a weight of approximately 1.5kg which is slightly heavier than standard for commercially available long flight time drones. The other case is a drone that is designed to fly uninterrupted for one hour. The weight for this drone is assumed to be 4kg, the maximum for recreational drones in the Netherlands.

With a weight of 1.5kg the thrust of the drone would at least have to be 15N. However, with this thrust the drone would only be able to hover and not actually fly. Several fora and articles on the internet state that for regular flying (no racing and acrobatics) at least a thrust twice as much as the weight is required (e.g. [32] and [33]) so the thrust of the drone needs to be 30N or higher. If four motors and propellers are used, this comes down to 7.5N per motor-propeller pair.

APC propellers has a database with the performance of all of their propellers [34] which will be used to get an idea of what kind propeller will be used. There are way too many propellers to compare individually so as starting point a propeller diameter of around the 10" is used (the propeller diameter for the previously discussed Blade Chroma Quadcopter) and the propeller with the lowest required power at a static thrust (thrust when the drone does not move) of 7.5N or 1.68lbf is searched. This way the 5.5x2 propeller is found. It has the required thrust at around 24400RPM and requires a torque of approximately 0.511in-lbf or 60mNm. This requires about 150W of mechanical power. The propeller weighs 7g.

With the required torque and rotational velocity known a suiting motor can be chosen. For this the torque constant in mNm/A and the speed constant in RPM/V are required along with the operation range (like in the specifications for Maxon motors [35]). Unfortunately drone motor suppliers do not give the torque constant making it very difficult to determine the amount of thrust a motor-propeller pair could deliver. Therefore it is for now assumed that size (and thus the weight) of the motor is linearly related to the amount of thrust generated. The motor for the Blade Chroma Quadcopter weighs 68g and supports a quarter of the 1.3kg weighing drone. If the linear relation between weight and thrust were true our drone would require motors of 78g. Efficiency specifications are also hard to find so it is unknown how much power the motors actually require. For now an efficiency of 80% is assumed which would mean that every motor-propeller combination would require 190W.

In total the drone motors require 760W and the hardware requires an additional 20W. The drone without the battery pack weighs 800g meaning that 700g is left over for the battery pack. With an energy density of 120Wh/kg approximately 84Wh of energy is available. With that a flight time of about 6 minutes could be achieved. This is highly insufficient. There seems to be something wrong with the calculations and assumptions used here so the design will be done in a different way. It will be based on an existing design that works sufficiently and only where necessary will be deviated from that design.

The existing drone the design will be based upon the Blade. Of the drones with a flight time of 30 minutes it has the second highest weight of 1.3kg and it's parts are findable on the internet. Because the drone weighs only 1.3kg instead of the desired 1.5kg our design is then also not allowed to weigh more than 1.3kg. Otherwise the flight duration will be shorter. The motors used are 4 Chroma, 1.2M Opterra brushless motors and 10 inch propellers are used. These will thus also be used for our design. The motors weigh 68g per piece and the propellers 14g per piece. The battery pack used is a 11.1V 6300mAh LiPo battery which weighs 480g.

Since motor specifications are hard to find it is difficult to design a drone that could fly uninterrupted for one hour. Therefore this design will no longer be pursued. Our drone will have to land when it is empty and its battery needs to be swapped.

Overview with weight and size estimates

The total weight of the drone will be 1.3 kg. The drone uses 10" (25.4cm) propellers. Since the propellers should not overlap the frame too much the total width of the drone will be 60cm.

The weight will consist of the following parts:

  • 4 motors + propellers: 328g
  • Raspberry Pi controller: 22g
  • Lidar: 120g
  • Camera: 10g
  • Frame: 150g?
  • Miscellaneous: 100g
  • Over for battery pack: 570g

The required power for the components is:

  • 4 motors: 140W
  • Raspberry Pi: 12.5W
  • Lidar: 2.5W
  • Miscellaneous: 5W
  • Total: 160W

User Interface Design

The user interacts with the drone via multiple interfaces. Here we list the specific design elements and dynamics of each of these interfaces.

  1. A Natural Language User Interface (NLUI) to answer questions, make recommendations on desired destinations, take permission to change route mid flight, and similar speech-based interactions.
  2. A Graphical User Interface displayed through a web page that provides enter or change the destination and current location, add favorite locations for quick access. Allows user to change the altitude of the drone, stop and start the drone. Search bar: search for place or address, icons buttons that say food and drink, shopping, fun, and travel. Pressing food and drink shows a list of restaurants and bars and cafes with yelp reviews. Pressing travel shows gas stations, landmarks, airports, bus stations, train stations, hotels. Pressing fun shows cinemas and other entertaining places. Favorites show frequently used addresses.
  3. A Motion User Interface. Can interpret motion gestures.

Since a camera can be considered to be a part of a drone, we present a formula for distance computation, given some parameters. The only assumption made is that the type of lens is rectilinear. The explained formula does not hold for other types of lenses.

Since [math]\displaystyle{ D \cdot \phi = H_{obj} }[/math] and [math]\displaystyle{ f \cdot \phi = h_{obj} }[/math] hold, we can factor [math]\displaystyle{ \phi }[/math] out and get [math]\displaystyle{ D = \frac{ f \cdot H_{obj} }{ h_{obj} } }[/math]. After rewriting [math]\displaystyle{ h_{obj} }[/math] to be in terms of mm, we finally end up with:

[math]\displaystyle{ D(mm) = \frac { f(mm) \cdot H_{obj}(mm)} { \frac{ h_{obj}(px) } { h_{sensor}(px) } \cdot h_{sensor}(mm) } }[/math]

The parameters are as follows:

  1. [math]\displaystyle{ D(mm) }[/math]: The real-world distance between the drone and the user in millimeters. This is what is computed.
  2. [math]\displaystyle{ f(mm) }[/math]: The actual focal length of the camera in millimeters. If you see a 35mm equivalent focal length, then refrain from using it. If you simply see two values, a quick test will show you which one should be used.
  3. [math]\displaystyle{ H_{obj}(mm) }[/math]: The real-world height of the user in millimeters.
  4. [math]\displaystyle{ h_{obj}(px) }[/math]: The height of the user in pixels.
  5. [math]\displaystyle{ h_{sensor}(px) }[/math]: The height of the sensor in pixels, in other words, the vertical resolution of the video file (e.g. for 1080p videos this would be 1080, for 480p this would be 480 and so on).
  6. [math]\displaystyle{ h_{sensor}(mm) }[/math]: The sensor size in millimeters, which is the pixel size in millimeters multiplied by [math]\displaystyle{ h_{obj}(px) }[/math].

All these parameters should be present on the spec-sheet of a well documented camera, such as the OnePlus 6 (which is a phone that was used for shooting our own footage).

When looking at the amount of error this method gives when implementing, we find that on scales less than 500mm, it is unfeasible. For a scale from 1000mm and up, it can be used. The results of our tests can be found in following table.

Distance Computation Evaluation
Scale (mm) Computed Distance (mm) Absolute Error (m, rounded)
100 7989.680302875845 7.89
200 2955.9793551540088 2.76
500 936.64000567 0.44
1000 687.851437808 0.31
2000 1692.22216373 0.30
3000 3183.46590645 0.18
4000 4143.14879882 0.14

If the drone is to be equipped with 2 cameras, computing the distance towards an object is possible using stereo vision as proposed by Manaf A. Mahammed, Amera I. Melhum and Faris A. Kochery, with an error rate of around 2% [36]. This does, however, increase the number of cameras that need to be used, which is a further constraint on power.

As the distance estimation described is to be used solely for user tracking purposes, and not for detecting and avoiding obstacles, we believe that an error rate of 3.5% ((0.14/4)*100) on a scale of 4m is acceptable.

Obstacle detection

In general there are two possible and readily accessible technologies for obstacle detection, namely, stereo vision and rangefinders. We will discuss how each of these hardware solutions recover or find depth information and then proceed on how to actually use the depth information.

Finding depth

Stereo vision works by having two cameras acting as human eyes, hence the name. It relies on the principle of triangulation and thus requires intimate mathematical knowledge. An illustration of triangulation now follows.

Triangulation.png

In above image, taken from the open CV documentation, one can see [math]\displaystyle{ O }[/math] and [math]\displaystyle{ O' }[/math], the two camera's. [math]\displaystyle{ B }[/math] denotes the baseline which is a vector from [math]\displaystyle{ O }[/math] to [math]\displaystyle{ O' }[/math], or simply put the distance between the two cameras. The focal length of both cameras (it is assumed to be the same type of lense used for a stereo vision camera) is denoted by [math]\displaystyle{ f }[/math]. [math]\displaystyle{ X }[/math] denotes the point where the rays of the two cameras coincide (which is a pixel). [math]\displaystyle{ Z }[/math] denotes the distance between the mounting point of both cameras and [math]\displaystyle{ X }[/math], or simply put it is the depth information that we are interested in. Finally, [math]\displaystyle{ x }[/math] and [math]\displaystyle{ x' }[/math] are the distance between points in the image plane corresponding to the scene point 3D and their camera center. With [math]\displaystyle{ x }[/math] and [math]\displaystyle{ x' }[/math], we can compute the disparity which simply is [math]\displaystyle{ x - x' }[/math] or just [math]\displaystyle{ \frac {B \cdot f} {Z} }[/math]. Rewriting this for [math]\displaystyle{ Z }[/math], the depth, is trivial [37] [38]. While it is mathematically advanced, it is very accessible thanks to the OpenCV framework [39].

Rangefinders work in a relatively simpler way, as it uses lasers, which is just a type of wave. Hence, the depth (or distance) can be computed using [math]\displaystyle{ D = \frac{c \cdot t_{\Delta}} {2} }[/math] where [math]\displaystyle{ c }[/math] is the speed of light and [math]\displaystyle{ t_{\Delta} }[/math] the time it takes in order for the light to return to the point from which it was sent [40].

Using depth

When it comes to representing a 3D environment for pretty much any computer task, one generally comes across the octomap framework. It is an efficient open source probabilistic 3D mapping framework (written in C++) based on octrees github.

An octree is a tree-like datastructure, as the name implies. (It is assumed that the reader is familiar with a tree data structure. If this is not the case, an explanation can be found here). The octree datastructure is structured according to the rule : "A node either has 8 children, or it has no children". These children are generally called voxels, and represent space contained in a cubic volume. Visually this makes more sense.

Octomap image.png

By having 8 children, and using an octree for 3D spatial representation, we are effectively recursively partitioning space into octants. As it is used in a multitude of applications dealing with 3 dimensions (collision detection, view frustum culling, rendering, ...) it makes sense to use the octree.

The OctoMap framework creates a full 3-dimensional model of the space around a point (e.g. sensor) in an efficient fashion. For the inner workings of the framework, we refer you to the original paper, from which we have also taken above figure. Important notions that need to be mentioned is that one can set the depth and resolution of the framework (as visible in the same figure). As it uses a depth-first-search traversal of nodes in an octree, by setting the maximal depth to be traversed one can influence the running time greatly. Moreover, the resolution of the resulted mapping can be changed by means of a parameter.

With the information presented in said paper, it is doable to convert any sort of sensory information (e.g. from a laser rangefinder or lidar) into an Octree representation ready for use [41], as long as it has following scheme.

NODE x y z roll pitch yaw
x y z
x y z
[...]
NODE x y z roll pitch yaw
x y z
[...]

An example of such a (converted) representation can be seen in the image below.

Octomap example map.png

Using an octree representation, the problem of flying a drone w.r.t. obstacles has been reduced to only obstacle avoidance, as detection is a simple check in the octree representation found by the octomap framework.

Obstacle avoidance

In order for a drone to navigate effectively through an urban environment, it must circumvent various obstacles that such an environment entails. To name some examples: traffic lights, trees and tall vehicles may form an obstruction for the drone to follow its route. To solve this problem, we will apply a computationally simple method utilizing a potential field [42].

The goal of this method is to continuously adjust the drone’s flightpath as it moves towards a target, so that it never runs in to any obstacles. For this subproblem, we assume the position and velocity of the drone, target and obstacles are known.

The potential field is constructed using attractive and repulsive potential equations, which pull the drone towards the target and push it away from obstacles. The attractive and repulsive potential fields can be summed together, to produce a field guides the drone past multiple obstacles and towards the target.

We consider potential forces to work in the x, y and z dimensions. To determine the drone’s velocity based on the attractive and repulsive potentials, we use the following functions:

[math]\displaystyle{ p_d^{att}(q_d, p_d) = \lambda_1 d(q_d, q_t) + p_t + \lambda_2 v(p_d, p_t)\\ p_d^{rep}(q_d, p_d) = -\eta_1 \dfrac{1}{d^3(q_o, q_d)} - \eta_2 v(p_o, p_d) }[/math]

where [math]\displaystyle{ \lambda_1, \lambda_2, \eta_1, \eta_2 }[/math] are positive scale factors, [math]\displaystyle{ d(q_d, q_t) }[/math] is the distance between the drone and the target and [math]\displaystyle{ v(p_d, p_t) }[/math] is the relative velocity of the drone and the target. Similarly, distance and velocity of an obstacle [math]\displaystyle{ o }[/math] are used.

There may be multiple obstacles and each one has its own potential field. To determine the velocity of the drone, we sum all attractive and repulsive velocities together:

[math]\displaystyle{ p_d(q_d, p_d) = p_d^{att} + \sum_o p_d^{rep} }[/math]

User Tracking

It should be clear that tracking the user is a large part of the software for the Follow-Me drone. We present python code (will be updated at the end of the project) that can be used with a webcam as well as normal video files, which can use various tracking algorithms.

Tracker Comparisons

In order to decide on which tracker to use, we have quite extensively compared all implemented trackers, except the Goturn tracker, in OpenCV (version 3.4.4 has been used). Both videos from an online source [43] as well as own footage (File:Video long.mp4 and File:Video short.mp4) has been used to compare the trackers.

We will first discuss all trackers w.r.t. the downloaded videos on a per-video basis, in the same order as has been tested. The discussion will include two major points, namely how good a tracker is dealing with occlusion in the sense that it should continue tracking the user and how good a tracker is dealing with occlusion in the sense that it should message that the user is occluded and afterwards resume tracking.

  • American Pharaoh: This video is from a horse race. For each tracker we select the horse in first position at the start of the video as object to be tracked. During the video, the horse racers take a turn, making the camera perspective change. There are also small obstacles, which are white poles at the side of the horse racing track.
    • BOOSTING tracker: The tracker is fast enough, but fails to deal with occlusion. The moment that there is a white pole in front of the object to be tracked, the BOOSTING tracker fatally loses track of its target. This tracker cannot be used to keep track of the target when there are obstacles. It can, however, be used in parallel with another tracker, sending a signal that the drone is no longer in line of sight with the user.
    • MIL tracker: The MIL tracker, standing for Multiple Instance Learning, uses a similar idea as the BOOSTING tracker, but implements it differently. Surprisingly, the MIL tracker manages to keep track of the target where BOOSTING fails. It is fairly accurate, but does not run fast with a measly 5 FPS average.
    • KCF tracker: The KCF tracker, standing for Kernelized Correlation Filters, builds on the idea behind BOOSTING and MIL. It reports tracking failure better than BOOSTING and MIL, but still cannot recover from full occlusion. As for speed, KCF runs fast enough at around 30 FPS.
    • TLD tracker: The TLD tracker, standing for tracking, learning and detection, uses a completely different approach to the previous three. Initial research showed that this would be well-performing, however, the amount of false positives is too high to use TLD in practice. To add to the negatives, TLD runs the slowest of the up to this point tested trackers with 2 to 4 FPS.
    • MedianFlow tracker: The MedianFlow tracker tracks the target in both forward and backward directions in time and measures the discrepancies between these two trajectories. Due to the way the tracker works internally, it follows that it cannot handle occlusion properly, which can be seen when using this tracker on this video. Similarly to the KCF tracker, the moment that there is slight occlusion, it fails to detect and continues to fail. A positive point of the MedianFlow tracker when compared with KCF is that its tracking failure reporting is better.
    • MOSSE tracker: The MOSSE tracker, standing for Minimum Output Sum of Squared Error, is a relative new tracker. It is less complex than previously discussed trackers and runs a considerable amount faster than the other trackers at a minimum of 450 FPS. The MOSSE tracker also can easily handle occlusion, and be paused and resumed without problems.
    • CSRT tracker: Finally, the CSRT tracker, standing for Discriminative Correlation Filter with Channel and Spatial Reliability (which thus is DCF-CSRT), is, similar to MOSSE, a relative new algorithm for tracking. It cannot handle occlusion and cannot recover. Furthermore, it seems that CSRT gives false positives after it lost track of the target, as it tries to recover but fails to do so.


At this point we concluded that, for actual tracking purposes (e.g. how well the tracker deals with occlusion in the sense that it continues tracking the user), only the MIL, TLD and MOSSE trackers could be used. For error reporting (e.g. how well the tracker deals with occlusion in the sense that it sends a message that the user is occluded and afterwards resumes tracking) it is seen that the KCF tracker is superior, which is also supported by online sources.


  • Dashcam Boston: This video is taken from a dashcam in a car, in snowy conditions. The car starts off behind a traffic light, then accelerates and takes a turn. The object to track has been selected to be the car in front.
    • MIL tracker: Starts off tracking correctly. Runs slow (4-5 FPS). Does not care about the fact that there is snow (visual impairment) or the fact that the camera angle is constantly changing in the turn. Can perfectly deal with the minimal obstacles occuring in the video.
    • TLD tracker: It runs both slower and inferior to MIL. There are many false positives. Even a few snowflakes were sometimes selected as object-to-track, making TLD impossible to use as tracker in our use case.
    • MOSSE tracker: Unsurprisingly, the MOSSE tracker runs at full speed and can track without problems. It does, however, make the bounding box slightly larger than initially indicated.

When looking at only the downloaded videos (there were more than 2), the conclusion remained the same. MOSSE is (in these cases) superior in terms of actual tracking, and KCF is (in these cases superior) in terms of error reporting. It is assumed that our own footage would conclude the same.

  • Real Life Scenario 1: The video, available File:Long video.mp4 here, has been shot by Noah. The user to track is Pim. The filming location is outside, near MetaForum on the TU/e campus. In the video, Noah is walking backwards while holding a phone in the air to emulate drone footage. During the course of this video, Nimo and Daniel will be occasionally walking next to Pim, as well as in front of him. There are 7 points of interest defined in the video (footage was around 2 and a half minutes long), which are as follows:
    1. Daniel walks in front of Pim.
    2. Nimo walks in front of Pim.
    3. Everyone makes a normal-paced 180 degrees turn around a pole.
    4. Pim almost gets out of view in the turn.
    5. Nimo walks in front of Pim.
    6. Daniel walks in front of Pim.
    7. Pim walks behind a pole.
    • BOOSTING tracker: Runs at a speed of around 20 - 25 FPS.
      1. This is not seen as an obstruction. The BOOSTING tracker continues to track Pim.
      2. The tracker follows Nimo until he walks out of the frame. Then, it stays at the place where Nimo would have been while the camera pans.
      3. As mentioned, the bbox (the part that shows what the tracker is tracking) stays at the place where Nimo was.
      4. At this point the tracker has finally started tracking Pim again, likely due to luck.
      5. see 2. and notes.
      6. see 2. and notes.
      7. see 2. and notes.
      • Notes: Does not handle shaking of the camera well. It's also bad at recognizing when the subject to track is out of view; the moment there is some obstruction (points 2, 5, 6, 7), it follows the obstruction and tries to keep following it even though.
    • MIL tracker: Runs at a speed of around : 10 - 15 FPS.
      1. The tracker follows Daniel for a second, then goes back to the side of Pim, where it was before.
      2. The tracker follows Nimo until he walks out of frame. Then it follows Daniel as he is closer to the bbox than Pim.
      3. The tracker follows Daniel.
      4. The tracker follows Daniel
      5. The tracker follows Nimo until he is out of view.
      6. The tracker follows Daniel until he is out of view.
      7. The tracker tracks the pole instead of Pim.
      • Notes: The bbox seems to be lagging behind in this video (it's too far off to the right). Similarly to BOOSTING, MIL cannot deal with obstruction and is bad at detection tracking failure.
    • KCF tracker: Runs at a speed of around: 10 - 15 FPS
      1. The tracker notices this for a split second and outputs tracking failure, but recovers easily afterwards.
      2. The tracker reports a tracking failure and recovers.
      3. no problem.
      4. no problem.
      5. The tracker does not report a tracking failure. This was not seen as an obstruction.
      6. The tracker does not report a tracking failure. This was not seen as obstruction.
      7. The tracker reports a tracking failure. Afterwards, it recovers.
      • Notes: none.
    • TLD tracker: Runs at a speed of around: 8 - 9 FPS.
      1. TLD does not care.
      2. TLD started tracking Daniel before Nimo could come in front of Pim, then switched back to Pim after Nimo was gone.
      3. TLD likes false positives such as water on the ground, random bushes and just anything except Pim.
      4. see notes.
      5. see notes.
      6. see notes.
      7. see notes.
      • Notes: The BBOX immediately focusses only on Pims face. Then it randomly likes Nimo's pants. It switches to Nimo's hand, Pim's cardigan and Pim's face continuously even before point 1. At points 4 through 7 the TLD tracker is all over the place, except for Pim.
    • Medianflow tracker: Runs at a speed of around: 70 - 90 FPS
      1. No probem.
      2. It tracks Nimo until he is out of view, then starts tracking Daniel as he is closer similar to MIL.
      3. It tracks Daniel.
      4. It tracks Daniel.
      5. It reports a tracking error only to return to tracking Daniel
      6. It keeps tracking Daniel until out of view, then reports tracking error for the rest of the video.
      7. Tracking error.
      • Notes: The bbox becomes considerably larger when the perspective changes (e.g. Pim walks closer to the camera). Then, it becomes smaller again but stops focussing on Pim and reports tracking failure.
    • MOSSE tracker: Runs at a speed of around: 80 - 100 FPS
      1. Doesn't care
      2. It follows Nimo.
      3. It stays on the side.
      4. It is still on the side.
      5. It follows Nimo
      6. It had stayed on the side and tracked Daniel for a split second.
      7. It tracks the pole.
      • Notes: Interestingly MOSSE doesn't perform well in a semi real-life scenario. It is bad at detecting when it is not actually tracking the subject to track.
    • CSRT tracker: Runs at a speed of around: 5 - 7 FPS
      1. It doesnt care.
      2. It follows Nimo, then goes to the entire opposite of the screen and tracks anything in the top left corner.
      3. Tracks anything in the top left corner.
      4. Tracks anything in the top left corner.
      5. Tracks Nimo, then keeps tracking the top left corner.
      6. was at top right
      7. Goes to top left
      • Notes: Deals badly with obstruction. Likes to track the top left and right corners of the screen.
  • Real Life Scenario 2: The video, available File:Short video.mp4 here, has been shot by Noah. The user to track is Pim. The filming location is outside, near MetaForum on the TU/e campus. In the short video, we tried to increase realism by having Noah stand on the stone benches walking backwards in a straight line.
    • BOOSTING tracker: Runs at a speed of around: 25 - 30 FPS Can surprisingly deal with having Pim's head cut off in the shot. Fails to follow when Nimo walks in front of Pim (e.g. it follows Nimo).
    • MIL tracker: Runs at a speed of around: 8 - 10 FPS Can somewhat deal with having Pim's head cut off in the shot, but doesn't reset bbox to keep tracking his head as well as his body. Fails to follow when Nimo walks in front of Pim (e.g. it follows Nimo).
    • KCF tracker: Runs at a speed of around: 12 - 13 FPS Can deal with having Pim's head cut off in the shot and nicely resets bbox to include head afterwards. Can deal with Nimo walking in front and reports tracking failure.
    • TLD tracker: Runs at a speed of around: 8 - 10 FPS Can not deal with having Pim's head cut off in the shot (it tracks the stone behind him). Can deal with Nimo walking in front, but does not report tracking failure.
    • Medianflow tracker: Runs at a speed of around: 60 - 70 FPS Can deal with having Pim's head cut off in the shot, but seems to be bad with tracking in general (again the bbox is off on the right side). Fails to follow when Nimo walks in front of Pim (e.g. it follows Nimo).
    • MOSSE tracker: Runs at a speed of around: 70 - 80 FPS Can deal with having Pim's head cut off in the shot and nicely resets bbox to include head afterwards. Fails to follow when Nimo walks in front of Pim (e.g. it follows Nimo).
    • CSRT tracker: Runs at a speed of around: 4 - 6 FPS Can deal with having Pim's head cut off in the shot and nicely resets bbox to include head afterwards. Shortly switches to Nimo when he walks in front of Pim, but then switches back.

Tracker Conclusions

During the testing phases, we came to mixed conclusions. On one hand, for the downloaded video sources (not all testing is mentioned as to not clutter the wiki too much), the MOSSE tracker consistently outperforms the other trackers for tracking whichever subject we had selected to be tracked. KCF consistently outperforms with respect to error reporting. For our own footage, however, it showed that MOSSE was outperformed by KCF on the tracking front as well.

That said, we have decided to not run only 1 tracker, but run 2 in parallel: MOSSE and KCF. The reasoning, if still unclear, is because KCF is superior on the error-reporting front regardless of which video source used (we have also tested with the webcam, which confirms this). MOSSE has shown e.g. in the downloaded video files that it can be good, and since it runs in near real-time, we can simply run it next to KCF without noticeable performance degradation. Furthermore, since KCF cannot properly deal with complete occlusion (which is when it reports that it cannot track the user anymore), and MOSSE simply doesn't care much about occlusion, they complement each other nicely.

User dependent positioning

Since a user should still be able to move around as freely as possible during being guided, there is a need for the drone to react appropriately to certain scenarios and movements of the user, while still having the goal of guiding the user to her destination. In the following section, various variables, formulas and conditions will be introduced, that together allow for smooth movements of the drone, while reacting to various scenarios. The definitions will all refer to and at the same time act as a legend for the following image.

ToleranceZone.png

Measured variables

D … Distance: this describes the continuously updated horizontal distance between the drone and the user.

α … User angle: The angle between the last significant location of the user (α0) and the current location. Continuously updated.

α0 … Last user angle: Set to the direction of the last significant position of the user and used to measure relative angles of user movement.

Parameters

Dd … Default distance. Parameter that can be varied by the user, defining the desired default horizontal distance that the drone is supposed to be apart from the user while guiding.

γ … Tolerance angle: This angle indirectly defines the size of the tolerance zone around the user. Specifically, how far sideways a user can move without the drone reacting.

Pn … Next waypoint: The next waypoint to navigate to, given by a navigation service.

Pp … Previous waypoint: The previous waypoint that was navigated to. (not relevant in calculations)

PD … Position of the drone. (not relevant in calculations)

PU … Position of the user. (not relevant in calculations)

PU … New position of the user (with respect to PU). (not relevant in calculations)

ts … Stabilization time: A dynamic variable with a range from 0 to 1, factoring the size of the tolerance zone, where 1 causes the tolerance zone to have its normal size. Allows for smoother drone movements, when the user is constantly moving.

Functions

move_rel ( distance, angle ) … Relative movement of the drone, given a distance and an angle (measured from α0)

dir ( coordinate ) … Returns the angle between α0 and a given point

dist ( coordinate ) … Returns the relative distance from the drone to the given a point

Conditions

User steps out of the tolerance zone into the direction of the drone

If the user crosses the green line of the tolerance zone, this causes the drone to move into the direction of the next coordinate until the default distance is reached again.

Formula UDP Cond1.png

User steps out of the tolerance zone sideways

If the user crosses one of the yellow lines, the drone has to adjust sideways. It will position itself on the line between the user and the next coordinate, taking the shortest way.

Formula UDP Cond2.png

User steps out of the tolerance zone away from the drone

If the user leaves the tolerance zone by crossing the red line, the drone follows the user, establishing the default distance again.

Formula UDP Cond3.png

Simulation

A simulation has been created with the following goals:

  1. Provide a concrete implementation of the drone controller that could be used for a real drone;
  2. Quantify the performance of our solution in an experimental scenario;
  3. Visualize the internal workings of the drone.

As a framework for this simulation, Unity [44] was used. The main advantages of Unity are real-time 3D visuals, easy-to-use physics engine and robust scripting capabilities.

Pathfinding

Before the guiding process starts, the drone needs to have a high-level idea of how to get from A to B. Given a map of the environment, the drone should find an efficient path from its current position to the destination. This path should also be traversible for the user. In a real-life scenario, this could be implemented using GPS and a streetmap routing API [45]. For this simulation, a map of the environment is built, which is then fed into a pathfinding algorithm to find the shortest path to the destination.

To build a map of the environment, navigation nodes are placed at intersections, where the drone can choose between multiple paths. These nodes, together with the starting point and destination, form the set of nodes of a graph. To find whether two nodes in this graph are connected, it is checked whether these nodes have line of sight to eachother. If a solid object such as a wall is blocking vision, it is assumed that the drone cannot move freely between these nodes.

Once this graph has been constructed, an A* algorithm is used to find an optimal path to the destination. The straight-line distance of a node to the destination is used as a heuristic to optimize the pathfinding algorithm.


Conclusion

TODO: Rewrite this into something useful. Generally, a phone app serves better than a follow-me drone except for the secluded hazardous environments mentioned above. On top of that, each of those special environments needs a specifically designed drone with sensors, programming and materials that are in accordance with the role and the environment. Lastly, the only situations where a drone serves better are not particular for DTD patients exclusively but rather for the general population.