SwarmPlay: Interactive Tic-tac-toe Board Game with Swarm of Nano-UAVs driven by Reinforcement Learning

08/03/2021 ∙ by Ekaterina Karmanova, et al. ∙ Skoltech 0

Reinforcement learning (RL) methods have been actively applied in the field of robotics, allowing the system itself to find a solution for a task otherwise requiring a complex decision-making algorithm. In this paper, we present a novel RL-based Tic-tac-toe scenario, i.e. SwarmPlay, where each playing component is presented by an individual drone that has its own mobility and swarm intelligence to win against a human player. Thus, the combination of challenging swarm strategy and human-drone collaboration aims to make the games with machines tangible and interactive. Although some research on AI for board games already exists, e.g., chess, the SwarmPlay technology has the potential to offer much more engagement and interaction with the user as it proposes a multi-agent swarm instead of a single interactive robot. We explore user's evaluation of RL-based swarm behavior in comparison with the game theory-based behavior. The preliminary user study revealed that participants were highly engaged in the game with drones (70 and found it less artificial compared to the regular computer-based systems (80 analyzed and put under discussion. User study revealed that SwarmPlay has the potential to be implemented in a wider range of games, significantly improving human-drone interactivity.



There are no comments yet.


page 1

page 3

page 4

page 5

page 6

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

Fig. 1:

Tic-tac-toe game with swarm of drones: (a) A human plays Tic-tac-toe board game against the SwarmPlay. (b) A closed view on the board when drones won the match. (c) Game board image processed for the Computer Vision (CV) analysis.

One of the potential domains for the human-robot interaction research is physical board games with an adjustable structure level. Perceiving the game components and the board, understanding human movements, reasoning about the state, and manipulating the game components to win against human players are integral steps in robot-centric board games [1, 2]. For the human player, on the other side, the interaction with a robot provides a fresh perspective on the well-known competitive games, e.g., robotic rock-paper-scissors with RASA presented by Ahmadi et al. [3].

Today much work is aimed to improve the AI in such games. On the one hand, the researchers achieve a high level of AI by improving the estimation of human behavior, e.g., tennis player’s movement prediction proposed by Wu et al.


. Another approach is to develop systems with a sufficient game strategy. RL is one of the most frequently used machine learning approaches in robotics, among other AI algorithms, due to its relative simplicity in comparison with Deep Learning algorithms and a variety of other Supervised learning algorithms, which require much more sophisticated and detailed instructions on problem-solving. Meanwhile, RL algorithms are the ones that react to an environment and finds a solution on their own, by trial-and-error, and do not require any pre-defined dataset to be trained. Therefore, RL is widely implemented in in-game scenarios, e.g., online-game strategy Dota 2

[5], gaming decision system developed for Go by Silver et al. [6], and curling robot with adaptive deep reinforcement learning framework proposed by Won et al. [7]. However, the system architecture in robot-centric applications has been relatively little investigated and is narrowed to the single robotic manipulators and mobile robots [8, 9, 10]. The research on multi-robot games, though, is mostly focused on coordination between robotic agents, such as soccer game strategies suggested by Reis et al. [11] and Liu et al. [12] that exclude humans from the gaming stage. There are several works introducing novel human-drone interaction (HDI) approaches in various scenarios. For example, the Flyables system by Knierim et al. [13] presents nano-quadrotors as levitating tangibles in 3D space which can be controlled by the user. A similar approach is presented in BitDrones by Gomes et al. [14], where nano-quadrotors are used as self-levitating tangible building blocks, forming an interactive 3D display with a touchscreen array. GridDrones is another multi-agent system where the user directly interacts with a volumetric mid-air grid of 15 cube-shaped nano-quadrotors [15]. A novel system SwarmCloak introduced by Tsykunov et al. [16]

proposes the landing of a swarm of four flying nano-quadrotors on the light-sensitive pads with vibrotactile feedback attached to the user’s arms. Another approach for drone interaction was suggested in DroneLight developed by Ibrahimov et al.

[17]. The proposed drone control system is based on the gesture recognition, followed by the drone light-painting of predefined patterns collocated with human gestures. Nitta et al. proposed a novel gaming approach HoverBall with a flying ball based on quadrotor technology, allowing to change the ball’s physical dynamics and behavior based on the context of the sports game [18].

To upgrade the level of engagement and interactivity of traditional games, we suggest a novel game paradigm where each game piece has its own intelligence and mobility, and behaves jointly with other agents to win against the opponent (see Fig. 1). The proposed SwarmPlay technology provides human-swarm interaction (HSI) driven by RL in board games. To our knowledge, our prototype is the first approach towards using a multi-UAV system in physical games that involves human presence. In this research, we focus on the system architecture and its validation by user study, followed by a discussion about future work and potential SwarmPlay game applications.

Ii System Overview

Ii-a System Architecture

The developed SwarmPlay system consists of Crazyflie drones, Vicon Tracking system with 12 IR cameras for drone localization, a CV camera for the game state evaluation, a drone landing table with a game board, and PCs with Mocap framework, drone-control framework, CV system, and decision-making system (Fig. 2). The game board is divided into 9 cells according to Tic-tac-toe rules. According to the specified algorithm, the drones play Crosses (Xs), landing on the game board’s cells. At the same time, a human plays Noughts (Ox), placing cards with printed circles on the game board.

To obtain pictures of the game board providing awareness of a current status of the game, we use a camera Logitech HD Pro Webcam C920 of @30FPS mounted on the ceiling of the room. The game board is placed right under the camera. The pictures are sent to the CV system to determine the human’s turn. After that, data on the human’s turn as a cell number is sent to the decision-making system to define a cell where the drones should make their next turn. CV and decision-making processing is performed on Intel® Core™ i7-9750HF CPU @ 2.60GHz × 12 threads of execution. The most recent cell data is sent to the drone-control framework. The framework obtains both the target cell, where a drone should be sent, and data from the motion capture system about current drone positions. To obtain the high-quality tracking of the drones, we applied Vicon motion capture system with 12 cameras (Vantage V5) covering a 27 m space. Drones are sent to the target cells with PID control parameters, i.e., the target position, speed, and acceleration. The Robot Operating System (ROS) Kinetic framework is applied to run the developed software and ROS stack [19] for Crazyflie 2.0. The position and altitude are updated at 100 Hz for all drones.

Fig. 2: Swarm play game layout. Human player performs their turn with cardboard game elements, then the system estimates the current state of the game with the developed CV system and calculates next command for the drones.

To detect to which cell the user puts circles, we developed a corresponding CV system. As its input, we use a picture taken by an RGB camera mounted on the ceiling of the room where the gaming board is located. At each step, the CV system takes a picture and converts it to the grayscale. Then, thresholding and erosion with a kernel 5x5 are applied. After that, the picture is cropped and divided into 9 small images, one per game cell. For each small image, a contour search is performed. When users make their turn, they put a circle on a cell, which is then detected as a contour by the CV system and filled with black pixels. At the end of each step, the CV system computes the density of black pixels per game cell. In this case, big colored circles show a great density value. Thus, using some threshold, makes it possible to separate game elements, drones, and empty areas from each other. After detecting a new circle on the playing field, the CV system sends a corresponding game cell number, as the latest human turn, to a decision-making system to solve how exactly drones should behave in the situation.

Iii Game Strategy

Iii-a Implementation

The Tic-tac-toe game is played on a three-by-three grid. Each player takes a turn to place a symbol on an open square. The drones play as an “X” player, and the user is playing as an “O” player. The game is over if one of the players has three identical elements in a row: horizontally, vertically, or diagonally. The game can end with a draw result if there is no possibility to achieve any winning combination. The board is represented by two-dimensional 3x3 matrix , where each cell was enumerated as 1, 2, 3, … 9. Each element of the matrix equals one of the following values: 0 : Unoccupied Cell; +1: Drone Symbol “X”; -1: Player symbol “O”.

Iii-B Reinforcement Learning Algorithm

Most game theory algorithms applied in human-robot interaction scenarios, e.g., Minmax algorithm [20], presume an opponent (human) and an agent (robot) to be ideal. Thus, during the game, each step of the opponent’s strategy aims to maximize their reward and minimize the agent’s reward, while the agent actions (next game steps) aim to avoid the loss outcome. However, in the Model-free RL algorithms, the agent does not operate with any model of an opponent, nevertheless achieving high results.

In our case, an agent is the swarm of drones, which interacts with an environment being the game board and the opponent, i.e., a human player. Depending on a certain state, i.e., a certain layout of the game board, the swarm takes a certain action by selecting a cell for the next turn. The general architecture of the RL algorithm is presented in Fig. 3, where , is the state at the time and the next time step ; , is the reward received by a swarm for achieving the states , ; is the action taken at time from state . The key goal of RL is to define the best sequence of decisions which may allow the swarm to solve a problem while maximizing a long-term reward.

Fig. 3: RL algorithm: SwarmPlay structure.

Temporal Difference (TD) learning is one of the most frequently encountered model-free learning algorithms in RL as it enables the agent to learn through every single action it takes. TD updates the agent’s knowledge on every time-step (action) rather than on every episode (reaching the goal or end state). In the basic scenario, the state-value (SV) function V(s) is initially evaluated. The agent always takes an action that leads to the state with the highest value. Another approach is to evaluate the action-state Q function, i.e., the value of an action in a particular state under a certain policy. Depending on how Q-function is updated after each action, the TD methods are subdivided into several approaches to policy learning:

  • Q-learning (QL): an off-policy method, where the RL agent learns about policy from another policy.

  • SARSA: an on-policy method, where the RL agent learns from both current and propagated by one step state and action.

All three approaches to the TD learning were implemented to Tic-tac-toe game scenario: QL, SARSA, and SV function evaluation. The algorithm was trained for 50 K episodes with parameters presented in Table I. The reward values are accumulated by each approach for the training period and presented in Fig. 4.

Reward OutcomeTurn First turn Second turn
Win 1
Lose -1
Draw 0.1 0.5
Learning rate 0.2
Discount rate 0.9
Epsilon-greedy policy 0.3
TABLE I: Parameters for the RL Learning
Fig. 4: Accumulated reward for 50 K episodes for Q-learning (QL), SARSA, and State-value (SV).

The results revealed that the QL algorithm performed faster than SARSA and SV approaches: QL performed in 38.8 sec, SARSA in 39.4 sec, SV in 87.6 sec on average per 10 K episodes. Additionally, the QL algorithm demonstrated on average 28% higher reward in every episode than the SV method and 33% higher reward than SARSA, winning 32% of the time while learning. Therefore the trained QL algorithm was implemented for further experiments.

Iii-C Game Theory Algorithm

To estimate and compare the user experience of interaction with a swarm of drones guided by the developed RL algorithm during the Tic-tac-toe game, we have adjusted a Basic Algorithm strategy [20], applying it for the human-drone interaction scenario.

Since the proposed drone-based scenario of the Tic-tac-toe requires more preparation time and complex actions from the swarm, in this research, we hypothesized that the high complexity of the game and deterministic strategy would not meet the player’s expectations. To provide a considerable challenge and excitement for the user, we propose an Improved Basic (IB) Algorithm (see Algorithm 1).

while game is not over do
       human turn;
       checking for winning condition if drones can win (row, column, diagonal) then
             insert “X” in the third cell and end game;
             if human can win (row, column, diagonal) then
                   playing defensive;
                   play to make 2 in a row, column, diagonal (50 %);
                   or random choice (50 %);
       drone moving command;
       checking for winning condition;
       if drones move first then
             making the first move (50 %);
             or random choice (50 %);
      if winning condition then
             end game;
Algorithm 1 Improved Basic Algorithm

The purpose of this research is to achieve an interactive and realistic game process with the swarm. To simulate the errors caused by the human factor, we proposed a 50% chance of a random move in the human winning scenario, when the IB algorithm can randomly place a drone into the empty cell at the start of the game, even if this is not going along its winning strategy. Therefore, the random factor increases the variability of the matches and allows the player to win in both game modes, providing an overall positive game experience.

Iv Experimental Evaluation

Iv-a Research Methodology


We invited 20 participants aged 22 to 43 years (mean = 25.6, std = 4.7, 4 females) to complete the survey. Two of them have never interacted with drones before, 5 of them regularly deal with drones, and 13 of them have participated in drone-based scenarios several times.


At the beginning of the experiment, the procedure and game equipment were introduced to each participant. Rules of the Tic-tac-toe were described for 1 participant who has never played the game before. Game elements, i.e., Noughts for human-player and Crosses for the SwarmPlay system, were represented by cardboard plates with printed black circles and cross shapes of the drones, respectively. Players placed the playing elements on the horizontally arranged game board, 1 m by 1.2 m white board with grid lines. All participants were randomly separated into two groups by 10 and played 2 matches with the SwarmPlay. The first group played the game with RL-based algorithm and the second with IB algorithm.

At the end of the game, all respondents were asked to evaluate the SwarmPlay game with a Questionnaire based on a 5-point Likert scale. Both algorithms were evaluated separately on 7 metrics proposed in previous research on augmented games [21] and human-robot interaction systems [22]: excitement, engagement, latency, challenge, tiredness, stress factor, and Turing test. A “Turing test” metric was proposed as an additional parameter to evaluate user’s perception of their opponent as a person and not as an artificial system. The participants responded to the following post-game questions:

  • Excitement: How did you enjoy playing the game? Would you play it again? (Never again - Definitely).

  • Engagement: To what extent did the game hold your attention? (Couldn’t concentrate - Enjoyed the game).

  • Latency: How appropriate did you find the responding time of the drones? (Unbearable - Unnoticeable delay).

  • Challenge: To what extent did you find the game challenging? (Too easy - Really challenging).

  • Tiredness: Did you feel tired after playing one game? (Not tired - Exhausted).

  • Stress factor: Did you feel confident playing in the same space with drones? (Relaxed - Really stressful).

  • Turing test: Did you consider your opponent being a person? (Completely artificial - It feels like playing with a human).

Iv-B Experimental Results

We conducted a chi-square analysis, based on the frequency of answers in each category. The results showed that the game parameters are all independent (min RL: = 0.12 0.05, min IB: = 0.16 0.05). Additionally the chi-square test of independence revealed that the participants’ experience with drones does not affect the evaluation of drone swarm perception criteria, such as tiredness (RL: = 1.88, = 0.76, IB: = 5.60, = 0.47), stress factor (RL: = 3.69, = 0.45, IB: = 2.38, = 0.88) and Turing test (RL: = 6.03, = 0.19, IB: = 14.0, = 0.12).

Fig. 5: Questionary feedback on 5-point Likert scale for the human-drone game with RL-based algorithm and IB game theory algorithm.

In summary, 70% of the participants who played the IB Algorithm found the game exciting and expressed their interest in playing such games more (Excitement 4.0), almost 90% of them did not feel any discomfort playing along with drones (mean score 4.6) and 60% of users found the SwarmPlay response fast enough ( 4.0) with game theory Algorithm compared to usual human-opponent move (mean score is of 3.6).

The participants were fully engaged in the game based on the game theory algorithm (60% put 5 score, 40% - 4 score, mean is of 4.6), and about 90% claimed they did not get tired playing with drones ( 4, mean score 4.6). However, for the Turing test IB algorithm shows high variation 95% CI (2.27, 4.6). Only 20% of participants considered that playing with a robotic opponent was much distinguishable from the real person (Turing test evaluated as 3). Almost 70% of respondents considered the game being challenging (challenge level evaluated as 4).

The results revealed the RL algorithm being more exciting for the participants, with 70% of respondents put 5 score for RL algorithm, and only 30% put 5 score for game theory algorithm (mean score IB: 3.9, RL: 4.6). The two-way ANOVA analysis results showed a statistically significant difference between the Excitement evaluation for RL-based algorithm and IB algorithm ( = 4.05, = 0.047 0.05). The evaluation of engagement level and tiredness for RL algorithm are close to IB (mean score Engagement RL: 4.7, IB: 4.6; mean score of Tiredness for RL: 4.6, for IB: 4.6). The results of the study is presented in Fig. 5. However, for the Turing test RL algorithm showed less variation than IB (RL: 95% CI(3.82, 4.78), IB: 95% CI(2.32, 4.68)). Being evaluated more evenly the RL algorithm is proved to be closer to a real person, making intelligent moves, with some unpredictable for the opponent strategy. According to the received data, SwarmPlay won almost equal number of matches when drones started the game for 2 algorithms (40% IB and 30% RL), while only once the SwarmPlay won with IB and almost 30% with RL-based algorithm when human-player moved first. SwarmPlay more often prevents human from winning the game with RL-based algorithm: 70% Draw result and 30% SwarmPlay won result with human player’s first move. With applied game theory algorithm Draw outcomes occurred more frequently when SwarmPlay started (60%) comparing to 30% of draw outcomes with human-player start. For RL algorithm Draw result appeared equally often (70%). The most interesting observation is that when human-players started, they won in 60% games with IB and never with RL algorithm.


The results revealed that the first move is important for the game outcome (Table II).

First move
First move
Average time
SwarmPlay won 3 4 3 1 60.3 65.6
Draw 7 6 7 3 76.5 71.0
Human won 0 0 0 6 0 29.5
Time (sec): 70.1 74.0 73.1 68.4 71.9 70.4
TABLE II: Relation between First Move and Game Results with RL-based and IB Algorithms

At the same time, with the applied IB algorithm, the first player won 46% cases, whilst with RL-based algorithm won only 18.2%. The RL-based algorithm was proved to be more successful in cases when the SwarmPlay takes the second turn. According to the received data, SwarmPlay won 35% of all matches when drones started the game, with only 7.7% of them won against game theory algorithm and 27.3% against RL algorithm. SwarmPlay more often prevents human players from winning the game with RL algorithm: 72.2% of the Draw result and 27.3% of the SwarmPlay winning result with human player’s first move.

[ ] [ ] [ ] [ ]

Fig. 6: Correlation between: (a) Excitement. (b) Challenge. (c) Stress. (d) Turing test on 5-point Likert scale and game outcome: human-player won, draw, or SwarmPlay won.

Surprisingly, we found a correlation between the game outcome (win-draw-lost) and game evaluation (Fig. 6). Firstly, the human losses affect the user’s excitement negatively (in average the excitement score is 12% lower with Draw result and 10% lower when SwarmPlay won than when a human player won). Nevertheless, users were much more focused on a game and not on system operation when SwarmPlay was leading the match. Secondly, it was discovered the more sophisticated strategy SwarmPlay performed and the more points it had the more human-like the behavior of the SwarmPlay participants mentioned. Meanwhile, results revealed that the participants evaluated the game challenge 34% more when he or SwarmPlay won. Moreover, the participants felt more confident playing with drones when the game resulted in SwarmPlay win or Draw result (stress factor in average is 10% less).

In the free comment space, four participants suggested implementing the external signal systems, e.g., sound alarms, either to indicate the start and the end of the game or to warn the player about swarm intentions.

V Conclusions and Future Work

We have developed the SwarmPlay, a system in which a human plays a Tic-tac-toe game against the swarm of drones. Our experimental results showed that 80% of the participants found the game exciting and expressed their interest in playing the game again, wherein the RL algorithm was more exciting for the participants: 70% of respondents put 5 scores for RL algorithm, and only 30% for IB algorithm (mean score IB: 3.9, RL: 4.6). Participants showed a high engagement with the proposed technology (engagement mean score of 4.7 out of 5.0 for RL algorithm and of 4.6 for IB algorithm). Therefore, SwarmPlay can potentially improve our way of interaction with game pieces. Machines can not only learn from a human’s winning strategy but also can teach humans how to achieve such a strategy throughout the interaction with an intelligent swarm.

The proposed system might be helpful in various scenarios, e.g., teaching the swarm of secure communication, where human shows the swarm places which can harm communication signals. Swarm then will establish the formations and communication nodes to achieve safe data transferring. Another approach could be open games with drones. The human-robot interaction via CV cameras can be implemented in training scenarios, to teach rescue personnel how to perform operations in a cluttered environment. There are a variety of home robot assistants, e.g., Kury, Buddy, Aido, etc. All these robots have a wheeled platform, therefore, they can move around the entire apartment. However, they can’t fly and bring humans any payload that is placed in cluttered and high locations. Home drone assistants can be used for that purpose. Humans can send the drones the target they need to reach, and they can do it in the safest way possible.

The future work will be devoted to more advanced board games, and we plan to apply ML techniques to learn the level of the player and adjust the difficulty level of the game in real-time.


  • [1] Y. Li, G. Carboni, F. Gonzalez, D. Campolo, and E. Burdet, “Differential game theory for versatile physical interaction,” Nature Machine Intelligence, vol. 1, pp. 36–43, 01 2019.
  • [2] C. Matuszek, B. Mayton, R. Aimi, M. P. Deisenroth, L. Bo, R. Chu, M. Kung, L. LeGrand, J. R. Smith, and D. Fox, “Gambit: An autonomous chess-playing robotic system,” in 2011 IEEE International Conference on Robotics and Automation.   IEEE, 2011, pp. 4291–4297.
  • [3] E. Ahmadi, A. Pour, A. Siamy, A. Taheri, and A. Meghdari, Playing Rock-Paper-Scissors with RASA: A Case Study on Intention Prediction in Human-Robot Interactive Games, 11 2019, pp. 347–357.
  • [4] E. Wu and H. Koike, “Futurepong: Real-time table tennis trajectory forecasting using pose prediction network,” in Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, ser. CHI EA ’20.   New York, NY, USA: Association for Computing Machinery, 2020, p. 1–8. [Online]. Available: https://doi.org/10.1145/3334480.3382853
  • [5] V. do Nascimento Silva and L. Chaimowicz, “Moba: a new arena for game ai,” 2017.
  • [6]

    D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre, G. van den Driessche, J. Schrittwieser, I. Antonoglou, V. Panneershelvam, M. Lanctot, S. Dieleman, D. Grewe, J. Nham, N. Kalchbrenner, I. Sutskever, T. Lillicrap, M. Leach, K. Kavukcuoglu, T. Graepel, and D. Hassabis, “Mastering the game of go with deep neural networks and tree search,”

    Nature, vol. 529, pp. 484–489, 2016.
  • [7] D.-O. Won, K.-R. Müller, and S.-W. Lee, “An adaptive deep reinforcement learning framework enables curling robots with human-like performance in real-world conditions,” Science Robotics, vol. 5, no. 46, 2020. [Online]. Available: https://robotics.sciencemag.org/content/5/46/eabb9764
  • [8] S. A. Nugroho, A. S. Prihatmanto, and A. S. Rohman, “Design and implementation of kinematics model and trajectory planning for nao humanoid robot in a tic-tac-toe board game,” in 2014 IEEE 4th International Conference on System Engineering and Technology (ICSET), vol. 4.   IEEE, 2014, pp. 1–7.
  • [9] A. Kyohei, N. Masamune, and Y. Satoshi, “The ping pong robot to return a ball precisely,” in Omron TECHNICS, vol. 51.016, 2020, p. 1–6.
  • [10] C. Becker-Asano, E. Meneses, N. Riesterer, J. Hué, C. Dornhege, and B. Nebel, “The hybrid agent marco: A multimodal autonomous robotic chess opponent,” in Proceedings of the Second International Conference on Human-Agent Interaction, ser. HAI ’14.   New York, NY, USA: Association for Computing Machinery, 2014, p. 173–176. [Online]. Available: https://doi.org/10.1145/2658861.2658915
  • [11] L. P. Reis, F. Almeida, L. Mota, and N. Lau, “Coordination in multi-robot systems: Applications in robotic soccer,” in

    Agents and Artificial Intelligence

    , J. Filipe and A. Fred, Eds.   Berlin, Heidelberg: Springer Berlin Heidelberg, 2013, pp. 3–21.
  • [12] X. Liu, “Research on decision-making strategy of soccer robot based on multi-agent reinforcement learning,” International Journal of Advanced Robotic Systems, vol. 17, p. 172988142091696, 05 2020.
  • [13] P. Knierim, T. Kosch, A. Achberger, and M. Funk, “Flyables: Exploring 3d interaction spaces for levitating tangibles,” in Proceedings of the Twelfth International Conference on Tangible, Embedded, and Embodied Interaction.   ACM, 2018, pp. 329–336.
  • [14] A. Gomes, C. Rubens, S. Braley, and R. Vertegaal, “Bitdrones: Towards using 3d nanocopter displays as interactive self-levitating programmable matter,” in Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems.   ACM, 2016, pp. 770–780.
  • [15] S. Braley, C. Rubens, T. Merritt, and R. Vertegaal, “Griddrones: A self-levitating physical voxel lattice for interactive 3d surface deformations.” in UIST, 2018, pp. 87–98.
  • [16] E. Tsykunov, R. Agishev, R. Ibrahimov, L. Labazanova, T. Moriyama, H. Kajimoto, and D. Tsetserukou, “Swarmcloak: Landing of a swarm of nano-quadrotors on human arms,” SIGGRAPH Asia 2019 Emerging Technologies, Nov 2019. [Online]. Available: http://dx.doi.org/10.1145/3355049.3360542
  • [17] R. Ibrahimov, N. Zherdev, and D. Tsetserukou, “Dronelight: Drone draws in the air using long exposure light painting and ml,” in Proceedings of the 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN 2020), 2020, pp. 2–3.
  • [18] K. Nitta, K. Higuchi, and J. Rekimoto, “Hoverball: Augmented sports with a flying ball,” in Proceedings of the 5th Augmented Human International Conference, ser. AH ’14.   New York, NY, USA: Association for Computing Machinery, 2014. [Online]. Available: https://doi.org/10.1145/2582051.2582064
  • [19] W. Hönig, C. Milanes, L. Scaria, T. Phan, M. Bolas, and N. Ayanian, “Mixed reality for robotics,” in IEEE/RSJ Intl Conf. Intelligent Robots and Systems, 2015, pp. 5382 – 5387.
  • [20] S. Karamchandani, P. Gandhi, O. Pawar, and S. Pawaskar, “A simple algorithm for designing an artificial intelligence based tic tac toe game,” in 2015 International Conference on Pervasive Computing (ICPC), 2015, pp. 1–4.
  • [21] A. Kulshreshth and J. J. LaViola, “Evaluating performance benefits of head tracking in modern video games,” in Proceedings of the 1st Symposium on Spatial User Interaction, ser. SUI ’13.   New York, NY, USA: Association for Computing Machinery, 2013, p. 53–60. [Online]. Available: https://doi.org/10.1145/2491367.2491376
  • [22] A. Moschetti, F. Cavallo, D. Esposito, J. Penders, and A. Di Nuovo, “Wearable sensors for human-robot walking together,” Robotics, vol. 8, no. 2, 2019. [Online]. Available: https://www.mdpi.com/2218-6581/8/2/38