Article citation information:

Ciuică, O., Dragomir, C., Pușcă, B. Safety culture model in military aviation organisation. Scientific Journal of Silesian University of Technology. Series Transport. 2020, 108, 15-25. ISSN: 0209-3324. DOI: https://doi.org/10.20858/sjsutst.2020.108.2.

 

 

Oliver CIUICĂ[1], Cristian DRAGOMIR[2], Bogdan PUȘCĂ[3]

 

 

 

SAFETY CULTURE MODEL IN MILITARY AVIATION ORGANISATION

 

Summary. The list of organisational factors that may constitute accident premises represents a good analysis instrument, equally preventive, but also retrospective, in investigation situations. Forming an effective safety culture is a vital element of acquiring and maintaining an appropriate level of safety and implicitly limiting aviation events. The safety culture “growth” process starts by successfully choosing the right model for the organisation, evaluating the safety culture with the help of the right tools and capitalising the results by taking improvement measures. Once acquired, the safety culture gives the organisation an increased level of security and confidence, thus: low accident rate, active involvement and responsibility from all members, initiative in operations, safety procedures, direct and effective feedback, careful and constant research of procedures, continuous and intense training, setting performance standards both indoors and outdoors, planning several scenarios to create the required variety, the desire to try new ideas, accepting the risk, the failure.

Keywords: safety culture, leadership, aviation, Air Force, military, training

 

 

1. INTRODUCTION

 

The essential attribute of an organisation in the safety field is to issue regulations and to monitor and control their compliance. The idea that regulations would be the absolute solution for accident avoidance is widespread. Although they are results of long experiences, some of them are “written with blood”; regulations are not an absolute guarantee of safety. Admittedly, regulations compliance has a big advantage: it represents an absolute parameter before the law. For example: Does that mean I can cross on green, even if I see a car coming, just because the law is on my side?

The fact that almost every accident has a list of violations of the rules is not a justification for the hope that, by absolute compliance with all regulations, the flight would be accident-free. Arguments:

a) Like aeroplanes, regulations are also made by people, so they are not perfect.

b) A regulation has a target (the field it is focused at) but also consequences on some adjacent fields. Even the best regulations can have unforeseen negative consequences. (Example: regulating the payment system of the air force staff).

c) If the rule does not fit the situation, then it is not good just for the fact that it exists. If we could imagine a fully regulated system, then man would have no role to play within it (it could be replaced by an automatic machine).

d) Many regulations respond to the moment’s need. After a while, they may no longer be effective. Delay in updating regulations is one of the most dangerous sources of risk in aviation. People have to face a changed situation based on old rules. This causes compromises (in order to make things work) or blockages (not to conflict with the rules).

e) There are several levels of regulation, which are not always concordant, and creates confusion and uncertain “improvisations” (the law is above an internal regulation).

f) At each hierarchical level, the pressure to comply with the rules is higher from top to bottom than from horizontal and even less from bottom to top. The mismatch between levels leads to normative practice inconsistency, with high risks for operational safety. Failure to follow the rules at a certain hierarchical level, when perceived or only suspected by subordinates, has a devastating effect on the general normative climate. This is, moreover, the essence of the mechanism of “authority dissolution” (understanding by authority a space regulated by norms).

g) Finally, it would not be excluded that if someone wanted to comply with maximum rigour to all existing regulations at any given time, no plane could be lifted from the ground.

 

The sum of all regulations defines an ideal professional space, as it can probably never be found in real life. This is truer in critical situations (in war, in times of economic crisis, transition, etc.). However, the arguments listed above are in no way a rebellion against rules and regulations.

Aviation has been and shall remain a highly normative institution. Most of the regulations are vital to maintaining the efficiency and safety of the system. However, it has no use to fetishise the compliance for norms, as a unique and absolute way of preventing all problems. Sometimes, blind implementation of an inappropriate rule can do more harm than ignoring it. However, we must keep in mind that people are usually inclined to comply with regulations because this gives them confidence and security (only some categories of psychopaths violate the rules in principle just because they exist). Therefore, even when a rule has been violated, there must be an explanation, which as a rule has a much deeper meaning than just finding the violation. When the analysis of an accident ends with the conclusion that certain norms have not been observed, without going further in deciphering the mechanism of this fact, we can say either that the investigation was not professional or that it is trying to hide some deeper truths. Example of “correct violation” of the rules and regulations: In case of loss of power supply on-board, at night, under heavy weather conditions, the regulation provides for “catapulting” of fighter jets equipped with a catapult seat. Yet, this “rule” is usually violated, in all known cases, the pilots succeeding in landing, usually with the damage of the aircraft, but without other consequences. Moreover, the “violation” has always been received positively, being highlighted by the driving factors.

In the end, however, the organisation shall have to resume this process of forming its own safety culture, to determine the impact or better still if the measures taken to improve the safety culture have had the desired impact or if they need to be deepened. In the end, we shall conclude with a statement by James Reason: “If you are convinced that the organisation you belong to has a good/efficient safety culture, you are more than likely wrong ... a good safety culture is something that you can tend to, but it's hard to get. Its value and result lie more in the struggle to obtain it than in effect.” [4]

 

 

2. SAFETY MANAGEMENT SYSTEM

 

Although air accidents or catastrophes are rare, less serious events and a whole range of incidents occur frequently. These negative manifestations of safety foresee the imminence of a disaster with a strong impact on the resources of the aeronautical organisation. Ignoring these indicators, with minimal influence on safety only increases unwanted events.

In support of diminishing security vulnerabilities, the International Civil Aeronautical Organization (ICAO), imperatively supports the implementation of a Safety Management System within all aeronautical organisations that sign the convention. Thus, starting from 2006, the Safety Management Manual, a manual that has the main role to outline this concept and not to present in detail the steps to be taken for its implementation was introduced.

SMS (Safety Management System) represents a well-defined process, centred at the level of the entire organisation that generates effective, viable decisions based on the potential risks identified during operations and services provided.

SMS promises lower loss rates, but safety culture is an essential condition for success and the key to achieving future goals.

The main pillar of aeronautical safety is the formation and development of a correct culture and attitude in the aspect of the safety of the activities carried out based on:

-       knowledge and discipline in compliance with the regulations, operational procedures and the correct exploitation of the technical means provided;

-       compliance with aeronautical safety rules in carrying out activities;

-       encouraging free and honest reporting and information of any potential factor or hazard that could affect the level of aviation safety or which has generated events, including by distinguishing between the occurrence of events due to unintentional errors and those causing voluntary violations of normative acts in the field of aviation safety.

 

Through education, training and action, staff must (re)know, identify and raise awareness of acceptable and unacceptable actions/attitudes in military aeronautical activities. Therefore, at all levels of the organisation, it should be understood that in the case of situations considered unacceptable (indiscipline, as an intentional action outside the limits of the normative acts in force), the commanders can and must establish certain disciplinary or administrative measures, proportional with the situation and the consequences manifested and in accordance with the regulations in force.

The systemic approach that encompasses all the actions carried out to improve the aeronautical safety, as presented in the concept of safety management, brings novelty elements by combining all the key elements that compete in carrying out the activities in good conditions.

The key elements of a Safety Management System are represented by:

-            the identification of hazards – recognition method of distinct  vulnerabilities of each organisation;

-            reporting of events that occurred - process of data acquisition and preparation of statistics on safety indicators (incidents, accidents, recurrence of events, etc.), but also voluntary reporting of “minor” incidents with a negative effect on safety;

-            risk management - standard approach for assessing risks and vulnerabilities for controlling and eliminating them;

-            measuring performance in terms of meeting the objectives - management tool for analysing the safety objectives imposed within the organisation;

-            safety quality assurance (auditing) - process based on the quality of managerial principles that support the improvement of the organisation’s performance to maintain the required safety standard.

 

 

3. HAZARDS IDENTIFICATION

 

One of the most important and applied models in the aviation community for determining causes that lead to the occurrence of aviation events is one proposed and developed by Professor James Reason[5], which can be found in speciality literature as the “Reason Model” or the “Swiss Chees” Model (cheese). This model (Fig. 1) describes the causality of events as a chain of successive failures of the security defence system and the successive activation/manifestation of latent factors that have adverse effects on the performance of the mission (equipment failures, organisational factors, errors, procedural violations, weather phenomena/conditions, deficiencies of regulations, deficiencies in training/evaluation, etc.).

Rarely, a single failure in this defensive chain can lead to the production of an event or its severe effects. The Air Force aeronautical system is based on the generation of defensive elements for any of its components. Interrupting the defensive chain can be prevented by including decision making at different levels of the organisation.

In certain situations, the interruption of the defence may be caused by a combination of active failures or latent conditions that, according to the Reason model, can be chronologically successive until the generation of a dangerous situation or even the production of aviation events with serious effects.

Active failures are actions or inactions that usually have an immediate aggravating effect. This category includes errors and violations of normative acts. These are attributed to the staff who actually carry out aeronautical activities (pilots, crew members, technical-engineering staff, traffic controllers, etc.) and the effect can be severe.

 

 

Fig. 1. Causal model of accidents

 

Latent conditions exist in the system before an event occurs. They have the following particularities:

-            they do not produce consequences as they are inactive, as a result, they are often difficult to perceive and identify, and risks are rarely associated with them;

-            there is a long period of time from occurrence to manifestation;

-            in most cases they are generated by people;

-            they are highlighted when the defensive means fail in aviation safety (occurrence of events).

 

These latent conditions can be generated by:

-            weak aeronautical safety culture;

-            equipment/material with deficiencies, which have been designed, manufactured or exploited improperly;

-            conflicting objectives or states in the organisation;

-            erroneous decisions;

-            organisation malfunctions.

 

The Reason model is thus constituted as a model of the “organisational” aviation events, bringing the human factor to the centre of attention of aeronautical events investigation by identifying its interaction:

-            at the individual level (the role of the person who leads and the one who enforces) and

-            collectively (the role of the organisation), both by the organisation way and the decision making.


Similarly, the model presents a series of elements regarding both the causality and ways of preventing the events, highlighting the following:

-            people make mistakes and it is important that these errors are made known, especially by the crews during the flight, and is important that there is time, space, resources and ways to prevent, mitigate or cancel the effects;

-            the aeronautical system has defensive means that can compensate for the fluctuation of human performance or the quality of decisions;

-            the system can generate latent conditions that can be activated in certain situations or under the influence of certain factors.

 

By analysing in detail, the causality of the accidents proposed by this model, we identify a mechanism of correlation with the processes that are carried out at the level of the organisations, which by identifying the process progression during the activities, determine both the logic of the occurrence of the aviation events and the modalities of prevention (the defensive layers of the model) that the aeronautical system can have available in different phases.

A particular path is represented by the generation and then the manifestation of latent conditions. They may be represented, in addition to those deficiencies in design and construction of equipment and incomplete or inadequate standard operating procedures, or deficiencies in staff selection and training. Thus, we can identify two major components that can generate latent conditions:

-            deficiencies in identifying the dangers/threats as well as ways to reduce the risks associated with them; these threats remain in the system and can be activated by factors or operational conditions in a particular situation;

-            normalisation of situations considered as an exception to the rule regarding, in particular, the allocation of resources. This situation forces the system to adapt so as to continue the execution of the assigned missions until the critical limit is reached or exceeded (exceptions are established that become rules and this process risks becoming a habit as the resources are continually maintained or reduced).

 

The latent conditions limit the defence capacity of the aeronautical system against threats/dangers, especially of the crews who carry out missions in flight and increase the risk of producing events by also restricting the possible solutions to prevent and correct errors of any kind:

-            the level of training, experience and continuity of the training are limited, affecting the maintenance, acquisition or improvement of flying skills;

-            decision-making and application skills are affected (increases the time, space and resources required, decreases the quality in their application);

-            the performance, characteristics and operating particularities of aircraft and technical systems remain in line with the requirements for mission execution and offer limited possibilities to prevent dangerous situations or limit their severity;

-            changes to normative acts, assigning new missions or increasing their complexity are more difficult to control in order to identify specific dangers or safety measures;

-            the frequent changes of structure or classification, the lack of qualified or experienced staff in the field generated by resources or by bad management of staff lead to difficulties in the sense of identifying specific threats or efficient and effective security measures.


The training and experience in aeronautical activities, the provisions of the normative acts and the technology are means that finally offer the possibilities of error prevention and equally the main tools for the enforcement of the missions in suitable conditions of performance, as well as efficiency and effectiveness.

A second path is represented by the working conditions, as follows:

-            personal factors: stability at the workplace, qualification and experience, moral and atmosphere within the organisation or within the crew, the attitude of the management, the remuneration;

-            environmental factors at work: ergonomics, lighting, temperature, cabin pressure, vibration and last but not the least, the associated occupational diseases.

 

Poor working conditions directly influence human performance during activities, being of major importance to the crews in flight, especially through their contribution to the manifestation of errors and voluntary violation of procedures or regulations. The difference between error and violation of rules is motivational and intentional.

From the viewpoint of prevention, the aeronautical safety activities should allow the intervention on the two paths presented above, thus:

-            supervises, analyses and evaluates the processes carried out at the level of organisations (internally and hierarchically: respectively of the Air Force base of the General Staff of the Air Force and of the Flight Group and of the squads from the Air Force base level), identifying in real and objective terms, the latent conditions and regenerating/strengthening defensive instruments;

-            supervises, analyses and evaluates the working conditions for identifying those factors that affect their performance and generates actions that will develop better working conditions and tools for avoiding errors, limiting or correcting their effects;

-            develops specific risk management processes for identifying security threats as well as identifying and applying risk reduction or elimination measures.

 

 

4. REPORTING EVENTS

 

The safety management system involves the reactive and proactive identification of the dangers manifested to safety during aeronautical activities. It is fully accepted that accidents or catastrophic investigations are much more detailed and thorough in research compared to investigations on-premises or incidents. Thus, when the safety optimisation measures are taken only following the conclusions of the investigations on accidents and disasters, the scenarios underlying them are limited. This way, wrong conclusions regarding the level of safety can be drawn, and moreover, inadequate corrective actions can be established.

The statistics show that the number of premises and incidents is much higher compared to the number of accidents and disasters (Classification of aviation events in the Romanian Air Force according to Annex no. 13). The causes and contributing factors associated with the premises or incidents can escalate into accidents or disasters. Often, only luck causes a minor event not to turn into a disaster. Unfortunately, these minor events are not always known by those in charge of implementing risk reduction and elimination measures. This may be due to the lack of a reporting system or the lack of motivation of the staff in not reporting the events or dangers discovered.


The lessons learned, the conclusions drawn from the incidents offer important scenarios for analysing and avoiding future events during activities. Therefore, there is a need for a database that provides the nature, cause and remediation of unwanted situations.

Equally valuable, such as information on the production of events, are information on unsafe, dangerous conditions that are yet to cause any incident.

The information on reported incidents facilitates the discovery of the risks associated with them, helps to implement intervention strategies and provides feedback regarding the evaluation of the effectiveness of the intervention. Events with minimal impact on safety also provide a first-hand understanding of the actions taken at the incident site regarding the conditions and actors involved. They can provide important details regarding the relationship between existing stimuli and their actions during the event, reactions that can affect their performance based on multiple factors such as fatigue, interpersonal interactions or distraction. Moreover, the members of the organisation involved in conducting events can offer solutions to increase the level of safety depending on the type of events. Data on-premises and incidents, even accidents, can be used to improve operating procedures, control the design of the technique used, and provide a better perspective on human performance in aircraft operation, air traffic control or technical service of the aerodrome.

Civil aviation regulations [6] propose three types of reporting systems:

1.        Mandatory reporting system - requires the persons responsible for reporting the events (or those involved in the event) to report on a hierarchical scale that a particular event has occurred. For this, a regulation that stipulates who shall report, to whom shall report and what should be reported[7] is needed. Although, such regulations may not cover all types of events to be reported given the large number of operational variants, the basic rule in informing about their production should be: “Report if you are unsure about reporting”.

2.        Voluntary reporting system - involves the reporting on the initiative of any member of the organisation of an event or of any inappropriate behaviour that endangers future activities.

Mandatory reporting involves informing about the events produced with the equipment provided (the hardware part of the organisation), with the need to collect data regarding the technical failures and its implications. To prevent these unwanted events, we propose the introduction of the voluntary reporting system meant to provide more information on the role of the human factor in the production and development of aviation events.

A good example of a voluntary reporting system is the US Aviation Safety Action Program (ASAP)[8]. Designed to increase aviation safety by preventing accidents and incidents, this is a system that protects the identity of the person reporting, it was created based on a model used by many airlines. This program, based on the transmission of network information, encourages the voluntary reporting of safety problems during the operation or maintenance of the technique, critical safety information, which might otherwise remain unknown.

The program is specially designed to detect the dangers and errors observed by crews, technical staff or air traffic controllers and their dissemination throughout the organisation so that everyone can have access to safety information. It additionally gives the organisation management examples of risk that might otherwise be “invisible” so those risk management decisions can increase the security of operations.

The challenge in implementing this reporting system is represented by the lack of punitive actions against those who subscribe to the information about the security threats. Being a non-punitive system, it will encourage the reporting of this much-needed data in the process of increasing the safety level.

3.        Confidential reporting system - aims to protect the identity of the information provider. Confidential reporting is not stored in a database or recorded. Usually, it is verbal information, mainly about the errors produced by the human factor in the activities of the organisation. This should be initiated without fear of reprisals or embarrassment, the main purpose of the information is to learn from the mistakes of others.

It is understandable that man is reluctant to report his own mistakes. Many times, following an aviation event, the commissions of investigation find that many of those present in the organisation were aware and knew the latent conditions of the event production before it happened. The non-reporting of perceived threats can be due to several reasons: embarrassment feeling before the interlocutor, self-accusation (if they were the ones generating the risk conditions), reprisals or sanctions from the hierarchy.

For a reporting system to be valid, the organisation must avoid reasons safety issues are not shared.

Trust and avoidance of sanctions are the basic principles in promoting a positive safety culture.

Persons reporting incidents, behaviours or errors that impact on the safety of their activities should be convinced that the organisation (management of the organisation) will not use the information received against them in any way or for any reason. Without this certainty, staff will avoid reporting errors or other observed hazards.

A positive safety culture within the organisation will generate the level of confidence required in reporting the observed inconsistencies. Specifically, the organisation must have a good tolerance for errors inherent in human activity of any kind, and the reporting system should be perceived as being correct in terms of its handling of errors (unintended errors). One must not misunderstand that in this way deliberate acts of violation of the regulations will remain unpunished. This is an example of just culture, an integral part of the safety culture.

To avoid anonymous reporting, which may leave interpretations in the information transmitted, and may have other purposes other than those related to security, the reporting system must be non-sanctioning, non-punitive and be based on confidentiality.

 

 

5. CONCLUSIONS

 

Recommendations on aviation safety:

-          analyse all aviation events that occur in your unit and other units to identify malfunctions that may lead to other aviation incidents or accidents;

-          approach rationally and functionally, the decision-making process for flight, starting from the interactions between human, technical or environmental factors that can ultimately lead to the occurrence of aviation events;

-          uniformly distribute effort according to the stage of training objectives and tasks to eliminate overwork at work, inadequate perception of danger or motivational dysfunctions;

-          identify the social factors that impact on in-flight training:

       social problems;

       financial problems;

       family relationships;

       relationships within the group;

-          eliminate tensions within the system and harmonise the activity of the compartments based on the idea that pilots, technical staff, insurance staff and navigators are the main preventers of aviation events.

 

During fight missions, mission accomplishment becomes paramount to all other considerations. No state, throughout history, has allowed the resources available for the accomplishment of the mission to be diminished by facts that are not attributed to the actions of the opponent. The accidents, beyond the costs they incur, reduce the operational capacity and, implicitly, the successful accomplishment of the assigned missions and create an image crisis of the Air Force, the most important category of an army.

 

 

Aknowledgement

 

This work was supported by a grant of the Ministry of National Education and Romanian Space Agency, RDI Programe for Space Technology and Advanced Research - STAR, project number 174/20.07.2017.

 

 

References

 

1.        Prentiss L., B.J. Brownfield. ”Aeronautical Decision Making”. Available at: http://www.avhf.com/html/Library/Aeronautical_Decision_Making.ppt.

2.        AFI91-204 12 FEB 2014 - Air Force Instruction , Department of the Air Force Headquarters Air Force Safety Center, Washington, 2016.

3.        AOPA. Available at: http://flighttraining.aopa.org/magazine/1999/September/199909_Features_Hazardous_Attitudes.html.

4.        Reason J. 1997. Managing the Risks of Organizational Accident. Ashgate, England: Aldershot.

5.        Reason J. 1990. Human Error, Cambridge. University Press, New York.

6.        Williamson A., A. Feyer, D. Cairns, D. Biancotti. 1997. „The development of a measure of safety climate: the role of safety perceptions and attitudes”. Safety Science 25.

7.        JAR-OPS 1 Subpart N Section 2 AMC OPS 1.943/1.945(a)(9)/1.965(b) (6)/1.965(e) - Crew Resource Management (CRM).

8.        CANSO The Civil Air Navigation Services Organization. Safety Culture Definition and Enhancement Process. 11April 2013.

9.        ICAO Doc 9856. „Safety Management Manual”. 2006.

10.    Instructions on the technical investigation of aviation events produced with military aircraft. Bucharestt, General Staff of the Air Force. 2007.

11.    Air Force Safety Center. Available at: http://www.afsec.af.mil/proactiveaviationsafety/asap/index.asp.

 


Received 20.02.2020; accepted in revised form 30.05.2020

 

 

Scientific Journal of Silesian University of Technology. Series Transport is licensed under a Creative Commons Attribution 4.0 International License



[1] “Henri Coandă” Air Force Academy, Mihai Viteazul 160 Street, Brasov, Romania. Email: oliverciuica@yahoo.com

[2] “Aurel Vlaicu” Air Force Training School, Boboc, Buzău county, Romania. Email: dragomir.cristian.safa@gmail.com

[3] Romanian Air Force Headquarter, București-Ploiești Street, 10,5 Km, Bucharest, Romania. Email: puscabogdan@yahoo.com

[4] Reason J. 1997. Managing the risks of organisational accidents. Aldershot, 1997, Ashgate Publishing Limited

[5] Reason J. Human Error ed. Cambridge press-New-York, 1990, p. 54

[6] ICAO Doc 9856, „Safety Management Manual”, first edition, 2006, p. 89

[7] Instructions on the technical investigation of aviation events produced with military aircraft, Bucharestt, General Staff of the Air Force, 2007

[8] Available at: http://www.afsec.af.mil/proactiveaviationsafety/asap/index.asp