1 Introduction and Related Work
Autonomous Vehicles (AVs) are robotic systems that can guide themselves without human operators. Such vehicles are equipped with Artificial Intelligence (AI) components and are expected to change dramatically the future of mobility, bringing a variety of benefits into everyday life, such as making driving easier, improving the capacity of road networks and reducing vehicle-related accidents.
Most likely, due to the lack of safety guarantees and legislation, as well as the missing scalability of Autonomous Driving (AD) systems, fully autonomous vehicles will not travel the streets in the near future. Nevertheless, over the past years, the progress achieved in the area of AI, as well as the commercial availability of Advanced Driver Assistance Systems (ADAS), has brought us closer to the goal of full driving autonomy [1, 2].
The Society of Automotive Engineers (SAE) has defined standard J3016 for different autonomy levels , which splits the concept of autonomous driving into 5 levels of automation. Levels 1 and 2 are represented by systems where the human driver is required to monitor the driving scene, whereas levels 3, 4 and 5, with 5 being fully autonomous, consider that the automated driving components are monitoring the environment.
An AV must be able to sense its own surroundings and form an environment model consisting of moving and stationary objects . Afterwards it uses this information in order to learn long term driving strategies. These driving policies govern the vehicle’s motion  and automatically output control signals for steering wheel, throttle and brake. At the highest level, the vehicle‘s decision-making system has to select an optimal route from the current position to the destination .
The main reason behind the human ability to drive cars is our capability to understand the driving environment, or driving context. In the followings, we will refer to the driving context as the scene and we will define it as linked patterns of objects. In this paper, we introduce AIBA (AI Behavior Arbitration), an algorithm designed to arbitrate between different driving strategies, or vehicle behaviors, based on AIBA’s understanding of the relations between the scene objects.
A driving scene consists of objects such as lanes, sidewalks, cars, pedestrians, bicycles, traffic signs, etc., all of them being connected to each other in a particular way (e.g. a traffic sign displays information for a driver). An example of such a driving scene, processed by AIBA within EB robinos, is depicted in Figure 1. EB robinos is a functional software architecture from Elektrobit Automotive GmbH, with open interfaces and software modules that manages the complexity of autonomous driving. The EB robinos reference architecture integrates the components following the sense, plan, act decomposition paradigm. Moreover, it also makes use of AI technology within its software modules in order to cope with the highly unstructured real-world driving environment.
Driving behavior prediction is a key part of an AV’s decision-making process. The task of identifying the future behavior of the scene’s objects is not a trivial one, since this type of information cannot be directly measured or communicated, being considered latent information . To be able to perform behavior prediction for the surrounding vehicles, an AV can use mathematical models which consider the variation in the objects’ movement and describe the driving scenario from the view point of the ego-car.
Such kind of models use several types of information, i.e. vehicle kinematics, the relationship between the ego-car and the surrounding entities, the interactions with other vehicles and a-priori knowledge. Vehicle kinematics and the relations with road entities were considered by almost all existing studies . For example, Lefevre et al. used the Time to Line Crossing (TTLC) to predict whether the vehicle will depart from the current lane or not .
Usually, most models for behavior arbitration are tailored for one specific scenario. Nevertheless, AVs must drive through dynamically changing environments in which a diversity of scenarios occur over time . Multiple scenario-specific models activate a corresponding model according to the characteristics of the scenario. In , the authors of this paper have proposed a three level Milcon type architecture. The second level, called the tactical level, includes a behavior collection and a manager program which trigger the appropriate AV behavior.
In this paper, we introduce AIBA, a system for behavior arbitration in AVs, which constructs a description and understanding model of the driving scene. Our idea is to model the human driver (HDr) understanding process of the driving scene, in order to achieve an optimal behavior arbitration solution. We describe the mechanism of the HDr thinking and transpose it to an approximate model.
The rest of the paper is organized as follows. The concept for describing the driving scene is given in Section II, followed by the understanding and modelling presented in Section III. The experiments showcasing AIBA are detailed in Section IV. Finally, conclusions are stated in Section V.
2 Driving Scene Description
The driving scene description is given from a human driver’s perspective, and it formulates properties derived from the definitions of classes, subclasses and objects which represent the core of an abstraction model, based on the authors’ previous work . The main idea behind AIBA is to model, or formalize, the HDr understanding process and afterwards transform it into a formal model for behavior arbitration in AV.
2.1 Driving Scene Analysis
A human driver is able to perceive the scene’s objects and observe them. This means that the HDr identifies the concepts and the different properties of the objects. In the end the driver can describe the scene in Natural Language (NL), as for the traffic example in Fig. 1: “the car is near exit E1; the main road continues straight and will reach the location L1, L2, L3; the traffic has a medium density and takes place under normal circumstances; the car is in the 3rd lane in safe vicinity to other cars”. Apparently, this message eludes a lot of information, but if it is shared with other human drivers, it proves to be a piece of very important and comprehensive knowledge.
The different scene objects are linked between them, as illustrated in the scene network from Fig. 2, built for the traffic scene shown in Fig. 1. Here, intuition assumes that the most important links are those established with the front car, or the main road (bold red lines in the image), while the less important links are those between the ego-car and the buildings (thin red lines in the image). In Fig. 2, each class of objects is linked with a rectangle. The symbols used are: AC for the ego-vehicle autonomous car, Bld for the buildings (i.e. side barriers), C1, C2, … , Cn for the other traffic cars, M, R, FR for lane identification, and RS for the road traffic signs.
The links definition is a first step in the knowledge process, which means it establishes the subjects of interest and also the importance level of each subject. The HDr scene understanding synthesis contains the following steps: identifying the link between the objects, allocating models to each link, running the models and afterwards finding a strategy to act. In fact, the HDr creates an implicit system and simulates it.
Two types of links can be observed: internal links, considered to be the set of links between the observer (the HDr) and all the observed objects, and external links, represented by the set of links between the objects present in the scene. From the HDr point of view, the links have different meanings and importance, or significance. Specifically, a human driver knows the traffic rules and how to get to his destination. These rules will determine traffic priorities, thus making the HDr to attach a greater importance to those links which are more important to his strategy.
In the next step, a description is established for each link. During the driving, the HDr adapts, or refines, the mentioned description by observation. The driver, by using the importance of the links, simulates a scene description. If we analyze the aim of driving scene understanding, we will observe that its origins are the stability in time and space. More precisely, the human driver has a driving task which can be accomplished if the possibility of locomotion is preserved in the current position and during a specific time. Intuitively, the stability is related to the objects in the scene and can be reached by understanding the scene. This understanding does not solve the driving problem, but offers the information upon which the human driver decides to act on a certain behavior.
2.2 Description proposal
Several concepts can be proposed at this point of analysis:
The understanding is related with the concept of intelligence which is defined by Piaget as “the human ability to obtain stability in time and space” . This means that the links and the models (behaviors) are correlated to the HDr intelligence.
The driving task has multiple attributes: determine a trajectory from start to end, ensure the comfort, minimize the energy consumption etc.,all which are time dependent. The task trajectory must be accomplished in safe conditions and should be included in the stability concept.
The driving scene is a complex environment, many entities being linked. In order to handle this complexity, the HDr establishes a network representing a holistic interpretation and priority interactions for each link of the network
Anticipating, the AD system is equipped with a sensor network, giving highly qualitative information about the scene.
Resuming, the previous analysis of the driving scene understanding can be described like a process which consists of the following steps:
Perceiving the objects (cars, traffic signs, lanes, pedestrians, etc.) and obtaining the object properties (the pedestrian intention is to cross the road).
Defining the links (the scene network) between the objects and their importance (i.e. the car in front is important, the pedestrian which intends to cross the road is very important, etc.).
Adding models to the mentioned links.
Simulating the model of the most significant links, and proposing driving behaviors which will be verified with the other significant links until the appropriate behavior (which allows the driving task) is found.
3 Driving Scene Modelling
Entities like objects, links, or networks, which have been introduced in the previous section, have correspondences in the modelling process. Our intention is to approximate the HD understanding process through a formal representation. More precisely, this assignment mimics an input/output process: using a perceived scene, the AIBA model must output a description which offers all the information needed within the AD system to arbitrate the driving behavior.
The block diagram representation of AIBA is illustrated in Fig. 3, where the information flow emulates the human driver scene understanding phenomenon.
Within AIBA, the first action is to transform the scene in to a collection of objects. This operation is accomplished by the following generative function:
where is the perceived scene; is the set of known object classes and is the set of objects.
Having an initial collection of classes, the generator function from Eq. 1 transforms the scene entities into a collection of objects. The following object classes are taken into consideration: traffic participants, pedestrians and buildings. The set of classes are a priori set within AIBA.
The complexity of this process, even for the set of road classes, is seen in many types of roads which exist around the world. In order to reduce the scene’s complexity, we split them in two major classes, static objects (lanes, traffic signs, buildings, etc.) and dynamic objects (cars, pedestrians, etc.).
Two kinds of definitions are here significant: the generic definition where the proximity and the properties are mentioned, and the extensive definition where the definition of the object (or a picture of it) is indicted or shown (similar to Fig. 1). This observation enables us to associate image recognition methods (which correspond to the extensive definition) with a generic definition collection.
Eq. 1 can be generalized when measurements (speed of the cars, distance between cars, size of traffic signs etc.) are associated with the scene description:
where is the set of measurements. The generated objects contain a structure of class specific properties and methods. The methods reflect the possible interactions of the objects with the ego-car.
The second step in AIBA’s workflow defines the links between a HDr and an object, as well as between the objects themselves, respectively:
where is the set of task trajectory properties and is the set of links’ significance.
Because the links are computed in term of stability around a particular point of the task trajectory, the significance is correlated with specific threats. Driving a car is subject to implicit negotiation based on traffic rules. The most important links are those related to agents which, according to these rules, have priority. Eq. 3 can be imagined as an expert system which will analyse all these links from the mentioned point of view, while outputting different marks:
Each recognized object provides methods which refer to the ego-car - object interactions. Information about possible threats on the task stability can be obtained by simulating the behaviors, and also how to select the appropriate driving behavior for avoiding these threats:
where is the set of important objects , which are linked with other objects with a significance , greater than a minimum (a priori imposed) significance . is the time horizon, is the current time, is the simulation time, is the set of the threat levels and is the set of recommended behaviors for the ego-car.
will simulate, for each important object , a collection of behaviors and a predicted threat level :
, or even neural networks.
In order to solve all links’ threats, several strategies can be chosen. If we adopt the HDr understanding description from the previous section, the behavior which solves the maximum threat is simulated for the other threats, obtaining the optimal behavior of the ego-car :
where are the threat levels.
The last step in AIBA’s modelling system is the transformation of the optimal behavior into a natural language explanation:
where is the set of descriptions, is the explanation of the significance associated to an object and is the number of significant objects.
Our experiments were conducted in our own autonomous vehicles prototyping simulator GridSim  and in Virtual Test Drive(VTD).
GridSim is a two dimensional birds’ eye view autonomous driving simulator engine, which uses a car-like robot architecture to generate occupancy grids from simulated sensors. It allows for multiple scenarios to be easily represented and loaded into the simulator as backgrounds, while the kinematic engine itself is based on the single-track kinematic model.
In such a model, a no-slip assumption for the wheels on the driving surface is considered. The vehicle obeys the ”non-holonomic” assumption, expressed as a differential constraint on the motion of the car. The non-holonomic constraint restricts the vehicle from making lateral displacements, without simultaneously moving forward.
Occupancy grids are often used for environment perception and navigation, applications which require techniques for data fusion and obstacles avoidance. We used such a representation in the previous work for driving context classification 
. We assume that the driving area should coincide with free space, while non-drivable areas may be represented by other traffic participants (dynamic obstacles), road boundaries, buildings, or other static obstacles. The virtual traffic participants inside GridSim are generated by sampling their trajectory from a uniform distribution which describes their steering angle’s rate of change, as well as their longitudinal velocity. GridSim determines the sensor’s freespace and occupied areas, where the participants are considered as obstacles. This representation shows free-space and occupied areas in a bird’s eye perspective. An example of such a representation can be seen in Fig.4.
VTD is a complete tool-chain for driving simulation applications. The tool-kit is used for the creation, configuration, presentation and evaluation of virtual environments in the scope of based simulations. It is used for the development of ADAS and automated driving systems as well as the core for training simulators. It covers the full range from the generation of 3D content to the simulation of complex traffic scenarios and, finally, to the simulation of either simplified or physically driven sensors.
For better understandability, we explain the behavior arbitration steps on a proposed simulation scenario. The sketch of the such a scenario can be found in Fig. 6. The car is driving in the middle lane, behind a car in the same lane. A car from the left side signals its intention to change the lane to the right, in front of the ego-vehicle.
In the first step the ego-vehicle identifies the three objects from the scene and measures the properties, also computing the possible impact times:
Car 1: in lane 3 (left, near), at distance d1, with speed v1, changing the lane
Car 2: in lane 2 (in front, far), at distance d2, with speed v2, following the 2nd lane
Car 3: in lane 1 (right, near), at distance d3, with speed v3, intending to stop
In the second step the ego-vehicle identifies links level:
Car 1 (left, near) = 0.3
Car 2 (in front, far) = 0.6
Car 3 (right, near) = 0.1
In the third step the threat time is computed, and compared to the maximum accepted level:
In the fourth and final step, the behavior of the ego-vehicle which solves the most probable threat is reducing speed. This is also simulated for the other threats which are above the maximum accepted level, and since it also solves the lane change to the right of the first car, it is chosen as the optimal behavior of the ego-car.
In summary, this paper proposes a novel approach to address the task of behavior arbitration. The method, coined AIBA, aims at formalizing and learning human-like driving behaviors from the description, understanding and modelling of a traffic scenario. The main advantage of the formal representation is that the functional safety of the automotive system can be analytically inferred, as opposed to a neural network black-box, for example.
As future work, we would like to extend the proposed method to more granular driving scenarios, such as driving in an indoor parking lot, or driving on uncharted roads. Additionally, we would like to add more objects in the scene, and to test the deployment on an embedded device.
We hereby acknowledge Elektrobit Automotive, Széchenyi István University, and the TAMOP - 4.2.2.C-11/1/KONV-2012-0012 Project ”Smarter Transport - IT for co-operative transport system” for providing the infrastructure and for support during research.
-  Katrakazas, C.; Quddus, M.; Chen, W.H.; Deka, L. Real-time motion planning methods for autonomous on-road driving: State-of-the-art and future research directions. Transp. Research Part C, 2015, 60, 416–442
-  Shai Shalev-Shwartz, Shaked Shammah, and Amnon Shashua. On a formal model of safe and scalable self-driving cars. arXiv preprint arXiv:1708.06374, 2017.
-  Litman, T. Autonomous vehicle implementation predictions (p. 28). Victoria, Canada: Victoria Transport Policy Institute. (2017)
-  Janai, J., Güney, F., Behl, A. and Geiger, A., 2017. Computer vision for autonomous vehicles: Problems, datasets and state-of-the-art. arXiv preprint arXiv:1704.05519.
-  B. Paden et. al “A survey of motion planning and control techniques for self-driving urban vehicles,” IEEE Trans. Intelligent Vehicles, vol. 1, no. 1, pp. 33–55, 2016.
-  J. Janai, F. G¨uney, A. Behl, and A. Geiger, “Computer vision for autonomous vehicles: Problems, datasets and state-of-the-art,” arXiv preprint arXiv:1704.05519, 2017.
-  Gindele, T.; Brechtel, S.; Dillmann, R. Learning driver behavior models from traffic observations for decision making and planning. IEEE Intell. Transp. Syst. Mag. 2015, 7, 69–79.
-  Lefèvre, S.; Dizan, V.; Christian, L. A survey on motion prediction and risk assessment for intelligent vehicles. Robomech J. 2014, 1, 1–14.
-  Lefevre, S.; Gao, Y.; Vasquez, D.; Tseng, H.E.; Bajcsy, R.; Borrelli, F. Lane keeping assistance with learning-based driver model and model predictive control. In Proceedings of the 12th ISAVC, Tykyo, Japan, 2014.
-  X. Geng, H. Liang, B. Yu, P. Zhao, L. He and R. Huang, ”A Scenario-Adaptive Driving Behavior Prediction Approach to Urban Autonomous Driving,” Applied Sciences, vol. 7, no. 4, p. 426, 2017.
-  Pozna, C., Troester, F.: Human Behavior Model Based Control Program for ACC Mobile Robot, Acta Polytechnica Hungarica Vol. 3, No. 3, pp. 59-70, 2006
-  Pozna, C., Precup, R. E., Minculete, N., Antonya, C., Dragos, C. A. ”Properties of classes, subclasses and objects in an abstraction model” In 19th International Workshop on Robotics in Alpe-Adria-Danube Region, pp. 291-296, 2010
-  Piaget, J., The Psychology of Intelligence, in Taylor & Francis, ISBN-10: 0415254019, 2005
-  XinyuWang, MengyinFu, Hongbin Ma,YiYang, Lateral control of autonomous vehicles based on fuzzy logic. In Control Engineering Practice Elsevier Volume 34, pp. 1-17, 2016
-  M. Bojarski et. al, “End to end learning for self-driving cars,” CoRR, vol. abs/1604.07316, 2016.
-  S. Grigorescu, Generative One-Shot Learning (GOL): A Semi-Parametric Approach for One-Shot Learning in Autonomous Vision, Int. Conf. on Robotics and Automation ICRA 2018, Brisbane, Australia, 21-25 May, 2018.
-  Trasnea, Bogdan, L.A. Marina, A. Vasilcoi, C. Pozna, and S. Grigorescu. ”GridSim: A Vehicle Kinematics Engine for Deep Neuroevolutionary Control in Autonomous Driving.” In 3rd IEEE International Conference on Robotic Computing (IRC), pp. 443-444, 2019.
Marina, L.A., B. Trasnea, T. Cocias, A. Vasilcoi, F. Moldoveanu, and S. Grigorescu. ”Deep Grid Net (DGN): A Deep Learning System for Real-Time Driving Context Understanding.” In 3rd IEEE International Conference on Robotic Computing (IRC), pp. 399-402, 2019.