Category Archives: Uncategorized
The 5th Review Conference of the Convention on Conventional Weapons (CCW) wrapped up today in Geneva and we’re very pleased that states agreed to hold two weeks of formal meetings in 2017 to discuss autonomous weapons. This Group of Governmental Experts (GGE) is the next step towards new international law about autonomous weapons. The international Campaign to Stop Killer Robots has a comment on the GGE decision online.
It’s been a busy week at CCW, Mines Action Canada delivered a statement in the General Debate and then we worked with our campaign colleagues to shore up support for the GGE.
So you didn’t miss out on any of the week’s events, we’ve created daily recaps in both Storify and video format. This week marks the start of a whole new phase of our efforts to ban killer robots. Donate today to support our work.
Thank you Chair. I appreciate the opportunity to speak on behalf of Mines Action Canada.
Although today we are starting the 5th Review Conference of the Convention on Conventional Weapons, we must spend our time looking forward. We are entrusted with preventing humanitarian harm from existing weapons like incendiary weapons and from future weapons that will require new legal instruments to avoid catastrophes to come.
CCW has spent three years holding informal meetings about autonomous weapons systems. At times during those discussions, we have felt that some have underestimated the skills, knowledge, intelligence, training, experience, humanity and morality that women and men in uniform combine with situational awareness and IHL to make decisions during conflict. We work closely with roboticists, and engineers, but despite their expertise and the high quality of their work, we do not believe an algorithm could replicate this complex and very human decision making process. Robotics should only be used to inform and supplement human decision making.
In the CCW’s work on autonomous weapons systems, we have learned more about Article 36 reviews but it is clear that states need to be more transparent, systemic and rigorous in their weapons review processes. Mines Action Canada believes that Article 36 weapons reviews should be a topic of discussion at the international level to strengthen both policy and practice around the world.
However, better weapons reviews will not solve the problems associated with autonomous weapons systems. For example, Article 36 reviews are not obligated to cover weapons used for domestic purposes outside of armed conflict such as policing, border control, or crowd control. Most importantly, weapons reviews cannot answer moral, ethical, technical and political questions. An Article 36 review cannot tell us if it is acceptable to the public conscience for a machine to kill without meaningful human control.
It is time for a separate effort to strengthen the standards and transparency around weapons reviews. That effort must neither distract from nor overtake our work here to deal with the real moral, legal, ethical and security problems associated with autonomous weapons systems. Weapons reviews must be grounded in new and robust international law that clearly and deliberately puts meaningful human control at the centre of all weapons development.
The concerns raised by autonomous weapons are urgent and must take priority. If we wait until everyone has a clear understanding of every aspect of the issue to start a Group of Governmental Experts the window of opportunity to prevent humanitarian harm from autonomous weapons will close. A GGE will allow high contracting parties to develop their understanding of the issue and to pursue effective outcomes.
In Canada, particularly, this year’s defence review offered an opportunity for the government to hear from a number of experts on autonomous weapons systems. A GGE next year would give Canada the opportunity to share the results of that process and to contribute our collective understanding of the issue.
Mines Action Canada, as a co-founder of the Campaign to Stop Killer Robots, believes that the way forward must lead to a pre-emptive ban on autonomous weapons systems as a tool to prevent humanitarian harm without damaging research and development on autonomy and robotics for military or civilian purposes. Earlier this year, a Canadian robotics expert made it clear there are no other applications for an autonomous system which can make a “kill or not kill” decision. The function providing an autonomous weapon the ability to make the “kill decision” does not have an equivalent civilian use therefore, pre-emptive ban on autonomous weapons systems would have no impact on the funding of research and development for artificial intelligence.
As experts at the meeting in April made clear our window of opportunity to prevent future humanitarian harm from autonomous weapons will not stay open long so we need to be moving forward at this Review Conference. Therefore, we urge states to accept the recommendation for an open-ended Group of Governmental Experts next year.
For the third year the Convention on Conventional Weapons (CCW) met in Geneva to work on autonomous weapon systems. From April 11th to 15th 2016, the informal experts meeting held at the United Nations addressed a wide range of issues and concerns. The meeting was attended by 94 member and observer States to the 1980 Convention on Conventional Weapons. This marked the highest turnout yet reflecting the increasing awareness and growing concern about these future weapons.
The CCW is a framework treaty that prohibits or restricts certain conventional weapons deemed to be excessively injurious or to have indiscriminate effects. The meeting included members of UN agencies, including the UN Institute for Disarmament Research, and the International Committee of the Red Cross (ICRC). The Campaign to Stop Killer Robots, with a diverse and passionate group of 40 campaigners from 10 countries, was active and provided very substantive input and expert analysis into discussions on issues such as meaningful human control and the weaknesses of Article 36 weapon reviews among other topics.
As a co-founder of the Campaign to Stop Killer Robots, Mines Action Canada (MAC) supports a pre-emptive ban on lethal autonomous weapons systems before they are ever developed or used. MAC was well represented at the meetings and gave statements about the repercussions of allowing such weapons to be developed. Executive Director Paul Hannon’s statement talked about the need to implement a ban on these systems before a humanitarian catastrophe occurs as well as the growing support for a pre-emptive ban. Project Coordinator Erin Hunt’s statement detailed the limitations of weapon reviews for lethal autonomous weapons systems and the importance of prioritizing a ban on these systems. Project Ploughshares, a new Canadian member of the Campaign to Stop Killer Robots, also participated actively.
Encouragingly, five nations Algeria, Chile, Costa Rica, Mexico, and Nicaragua called for a pre-emptive ban on lethal autonomous weapons systems bringing the total to 14 states that now support this goal. Throughout the meeting, Cuba, Ecuador, the Holy See and Pakistan as well as all NGO speakers reiterated the need for a ban on lethal autonomous weapons systems.
Although many nations, including Canada, were not willing at this time to support a ban on lethal autonomous weapons systems, the importance of meaningful human control was underlined many times during the 5-day meeting. The Netherlands announced new policy in support of meaningful human control over deployed weapon systems. As well, Austria made note of the recommendation by two UN Special Rapporteurs in their February 2016 report to prohibit lethal autonomous weapons systems that require no meaningful human control. The report can be found here with reference to lethal autonomous weapons systems on page 17.
The common area of concern regarding lethal autonomous weapons systems is the lack of ability to properly target and apply force in a strike. Without human control over key points of identification of targets and approval of a strike, innocent civilians could be inappropriately targeted and killed. It is important to maintain meaningful human control over critical functions of lethal autonomous weapons systems to ensure proper application of international humanitarian law as well as accountability were a target later to be understood as inappropriate. These qualifiers cannot be guaranteed with lethal autonomous weapons and thus threaten international security and stability.
The proliferation of lethal autonomous weapons systems risks the development of an arms race and a weakening of global security. Prohibitions and tight restrictions have helped to calm arms races and promote international peace. As well, lethal autonomous weapons systems threaten the protection of civilians in conflict zones. Autonomous robotic systems are most successful in predictable and reliable environments. However, autonomous weapons systems would be in highly unpredictable conflict zones and thus risk performing unreliably and endangering civilians.
NGOs remain concerned over the lack of concrete action taken on this issue. Strong and substantive statements were made by Mines Action Canada, Nobel Women’s Initiative, Human Rights Watch and others at CCW in April, urging states to establish an open-ended Group of Governmental Experts (GGE). A GGE would see in-depth study into the issues surrounding lethal autonomous weapons and submitting their findings to the UN. A GGE would be the first step in understanding issues of how lethal autonomous weapons systems would comply with international humanitarian law, the effects on regional or global security, create instability and the risk of an arms race. In addition, the GGE would give the CCW time to explore the blurred nature between soldier and weapon that fully autonomous weapon systems present. A short video compiled by the Campaign to Stop Killer Robots at the meeting in April can be seen here and daily updates can be found online for Monday, Tuesday, Wednesday, Thursday and Friday.
The meeting ended with some recommendations for the future and states ‘may decide to establish’ a GGE at the 2016 Fifth Review Conference of the High Contracting Parties to the Convention on Prohibition or Restrictions on the Use of Certain Conventional Weapons in December. In addition, the recommendations state that the GGE should work to identify characteristics of lethal autonomous weapons systems and create a working definition.
Although most interventions supported the formation of a GGE as the next step in this process, the final decision will not be made until the Review Conference in December. While Mines Action Canada is cautiously optimistic that a Group of Governmental Experts will be formalized at the meeting this December, we remain concerned that the pace of these discussions is not keeping up to the speed of technological change. Of course, the specific mandate of the GGE will be all important.
Much work remains over the coming months.
The Campaign to Stop Killer Robots released their report on the activities undertaken at the 3rd meeting of the Convention on Conventional Weapons and can be found here.
A student post by our summer student Miranda Barclay.
With the third and hopefully final Convention on Conventional Weapons (CCW) informal experts meeting coming up in a couple days, it is important to remind ourselves of what was discussed last year and what work still needs to be done.
The gathering of the CCW member states and organisations in Geneva in April 2015 was designed as a forum at which states could discuss the important technical, legal, moral and ethical issues surrounding autonomous weapons, otherwise known as ‘killer robots’.
At the 2015 meetings, almost all states that spoke agreed that further work is necessary and desirable and many expressed that no autonomous weapons should be allowed to operate without meaningful human control. Nor with human control that is ‘devoid of meaning.’ There were however a small number of states who were more reserved regarding the eventual achievement of a pre-emptive ban on autonomous weapons. The US and Israel implied that they plan to leave the door open for the future acquisition of these weapons. While France and the UK stated that they would not pursue killer robots but still neither indicated support for the logical conclusion of a pre-emptive ban.
Another important notion that arose from the CCW 2015 meetings was the fact that autonomous weapons or killer robots are not an inevitable piece of weaponry and should never be allowed to become an inevitable piece of weaponry. This notion was a useful counterpoint to some interventions that seemed to under-estimate to value and importance of human soldiers.
Further, the CCW focused heavily on norm creation, with members emphasising the need to establish norms in order to efficiently discuss and articulate what is most disturbing and threatening about the possibility of autonomous weapons use. Once these norms are clearly established and accepted by a majority of states, hopefully there will be a more concerted effort to transform these norms into fully ratified international laws.
Finally, multiple countries and organisations identified the need to define what exactly some of the key terms commonly used at the conference meant. For example, what exactly is meant by ‘meaningful human control’? Further explorations of this principle could be a key component of a Group of Governmental Experts in 2017 leading to a process to prevent the use of fully autonomous weapons through law.
Hopefully, this year some more solid definitions can be agreed upon and a Group of Governmental Experts will be called for next year so the process of banning autonomous weapons through international law can be accelerated leading to a pre-emptive ban.
Claudia Pearson is an undergraduate student at the University of Leeds, currently studying abroad at the University of Ottawa.
Executive Director Paul Hannon delivered our closing statement at the Convention on Conventional Weapons today. Download the statement here or read it below.
The Way Forward
Thank you Mr. Chair and your team for the strong foundation to move forward with the urgency and focus this issue requires. This week we have seen wide-ranging discussions on autonomous weapons systems. The CCW does not often enough deal with issues of morality, human rights and ethics. We welcome all states who have asserted the necessity of maintaining meaningful human control over the use of force. These conversations should continue and deepen.
There is one issue we would like to raise as food for thought. At times during the week, we have felt that some have underestimated the skills, knowledge, intelligence, training, experience, humanity and morality that men and women in uniform combine with situational awareness and IHL to make decisions during conflict. We work closely with roboticists, engineers, and technical experts and despite their expertise and the high quality of their work we do not believe an algorithm could replicate this complex decision making process. Robotics should only be used to inform and supplement human decision making. To go further than that risks “dehumanizing those we expose to harm” as RCW’s CCW Report’s editorial stated yesterday.
Allow me to conclude with the assertion that the international response to the possibility of autonomous weapons systems must not be limited to transparency alone. The expert presentations and the debates this week have strengthened our belief that autonomous weapons systems are not a typical new weapon and our current IHL and weapons review processes will not be sufficient. A mandate for a group of governmental experts next year is an appropriate and obvious next step. We look forward to working with the high contracting parties to ensure that meaningful human control remains at the centre of all decisions to use violent force.
Today at the Convention on Conventional Weapons meeting about lethal autonomous weapons systems, Mines Action Canada released a new memo to delegates on the impact of autonomous weapons systems on public trust in robotics. In this memo we discuss how the creation and use of autonomous weapons systems could change public perception of robotics more generally. Read the memo here and let us know what you think!
Will the use of killer robots make you more or less likely to want other autonomous robots in your life?
With the Convention on Conventional Weapons high contracting parties meeting this week to discuss lethal autonomous weapons systems, Mines Action Canada has released an updated memo to delegates on CCW Protocol IV which pre-emptively banned blinding laser weapons.
Please down load the Updated Protocol IV Memo.
Mines Action Canada delivered an opening statement at the Convention on Conventional this afternoon. The text of the statement is available online here.
Opening Statement – Convention on Conventional Weapons 13 April 2015
Thank you Mr.Chair. I appreciate the opportunity to speak on behalf of Mines Action Canada. Mines Action Canada is a Canadian disarmament organization that has been working to reduce the impact of indiscriminate weapons for over twenty years. For years we have worked with partners around the world including here at the CCW to respond to the global crisis caused by landmines and cluster munitions. We have seen that the international community can come together to respond to a humanitarian catastrophe and can create international humanitarian law to protect civilians often after the fact due to the changing nature of conflict and technological advances. However, we are here today in an attempt to look forward. We are looking at future weapons that will require new legal instruments to prevent future catastrophes. Throughout this week I hope we will keep my grandmother’s advice in mind: an ounce of prevention is worth a pound of cure.
As a co-founder of the Campaign to Stop Killer Robots, Mines Action Canada is very conscious of public opinion concerning autonomous weapons systems. Since last year’s discussions here at the CCW, opposition to autonomous weapons systems has grown in Canada. In addition to our member organizations, academics, parliamentarians, industry, faith communities and members of the general public have expressed concern about the potential humanitarian impacts of autonomous weapons systems. The widespread opposition to this technology indicates that there may be negative consequences for robotics more generally should autonomous weapons systems be used in armed conflict or in other circumstances. The erosion of public trust in robotic systems and autonomy as a result of the use of autonomous weapons systems could severely limit our ability to harness the good that robotics could do for humanity.
In addition to these concerns about the impact on public trust in robotics, we have numerous legal, moral, ethical, technical, military, political and humanitarian concerns about autonomous weapons systems which have led to the conclusion that new international humanitarian law is needed to ensure meaningful human control over these and other weapons. There is a moral imperative to consider the long term effects of the development and deployment of autonomous weapons systems on human society. Proponents of these technologies cite possible battlefield benefits and yet a discussion only dealing with short term or battlefield effects is not enough. We must ask the difficult questions: is it acceptable to cede decisions over life and death in conflict to machines? Who would be accountable for autonomous weapons systems? How can IHL adapt when new technology blurs the line between combatant and weapon?
IHL has demonstrated an ability to adapt and evolve to prevent the development and deployment of new and unnecessarily harmful technology. CCW Protocol IV banning blinding laser weapons is a good example which demonstrates that not only is there a need to add to IHL to address new technology, but also that we can prevent the development and use of weapons before their unacceptable humanitarian consequences create a catastrophe. We have published a memo to delegates which further explores the lessons learned from Protocol IV.
Autonomous weapons systems are not your average new weapon; they have the potential to fundamentally alter the nature of conflict. As a “game-changer” autonomous weapons deserve a serious and in-depth discussion. We hope that this week will see attempts to define meaningful human control and will foster a strong desire to pursue discussions towards a new legal instrument that places meaningful human control at the centre all decisions to use violent force.
Mines Action Canada is offering the Keep Killer Robots Fiction t-shirt for a limited time only. Campaign supporters can purchase this Keep Killer Robots Fiction t-shirt through TeeSpring until March 30, 2015. The Keep Killer Robots Fiction t-shirt is available in three styles: unisex (S-5XL), women’s fitted (S-2XL) and children’s (S-XL). These t-shirts can be shipped almost anywhere in the world.
The proceeds from the sales of this limited edition t-shirt supports Mines Action Canada’s advocacy work on autonomous weapons systems. You can purchase your t-shirt by visiting TeeSpring’s safe and secure site: http://teespring.com/keepkillerrobotsfiction.
The Keep Killer Robots Fiction t-shirt is available for a limited time only. Order before March 30 to ensure you don’t miss out!
Help us keep autonomous weapons systems off the battlefield and keep killer robots fiction!