According to NY Times, 737 MAX 8 pilots were trained on (their own?) iPads. 

What's next? Bring your own cockpit? Like suggested for car sharing interfaces 
by excited UX students all over the world these days. 

Such news scare me more than all the AI horror stories together. This 
banalization or "desktopization"* of high responsibily jobs should be seriously 
questioned. Even when it is technically possible, even if “magic pane of glass” 
has more processing power when onboard computer, even if flight deck software 
is written in Java Script, these are not sufficient reasons for a pilot to have 
it open in one of her browser tabs, even for training. 

may complex systems stay complex in the eyes of their operators

https://www.nytimes.com/2019/03/16/business/boeing-max-flight-simulator-ethiopia-lion-air.html?action=click&module=Top%20Stories&pgtype=Homepage

"For many new airplane models, pilots train for hours on giant, 
multimillion-dollar machines, on-the-ground versions of cockpits that mimic the 
flying experience and teach them new features. But in the case of the Max, many 
pilots with 737 experience learned about the plane on an iPad."

" But Boeing isn’t planning to overhaul its training procedures. And neither 
the F.A.A., nor the European Union Aviation Safety Agency, are proposing 
additional simulator training for pilots, according to a person familiar with 
the deliberations. Instead, the regulators and Boeing agree that the best way 
to inform pilots about the new software is through additional computer-based 
training, which can be done on their personal computers."

*in 2014 i wrote about desktopization of remote piloted aircrafts for Interface 
Critique http://contemporary-home-computing.org/RUE/

---- Olia Lialina wrote ----

>i was rereading today this 5 y. o. article about a decade old accident
>
>https://www.vanityfair.com/news/business/2014/10/air-france-flight-447-crash/amp
>
> following are parts of  IV. Flying Robots and the article's final statement
>
>It takes an airplane to bring out the worst in a pilot.
>[... ] 
>Wiener pointed out that the effect of automation is to reduce the cockpit 
>workload when the workload is low and to increase it when the workload is 
>high. Nadine Sarter, an industrial engineer at the University of Michigan, and 
>one of the pre-eminent researchers in the field, made the same point to me in 
>a different way: “Look, as automation level goes up, the help provided goes 
>up, workload is lowered, and all the expected benefits are achieved. But then 
>if the automation in some way fails, there is a significant price to pay. We 
>need to think about whether there is a level where you get considerable 
>benefits from the automation but if something goes wrong the pilot can still 
>handle it.”
>
>
>Sarter has been questioning this for years and recently participated in a 
>major F.A.A. study of automation usage, released in the fall of 2013, that 
>came to similar conclusions. The problem is that beneath the surface 
>simplicity of glass cockpits, and the ease of fly-by-wire control, the designs 
>are in fact bewilderingly baroque—all the more so because most functions lie 
>beyond view. Pilots can get confused to an extent they never would have in 
>more basic airplanes. When I mentioned the inherent complexity to Delmar 
>Fadden, a former chief of cockpit technology at Boeing, he emphatically denied 
>that it posed a problem, as did the engineers I spoke to at Airbus. Airplane 
>manufacturers cannot admit to serious issues with their machines, because of 
>the liability involved, but I did not doubt their sincerity. Fadden did say 
>that once capabilities are added to an aircraft system, particularly to the 
>flight-management computer, because of certification requirements they become 
>impossibly expensive to remove. And yes, if neither removed nor used, they 
>lurk in the depths unseen. But that was as far as he would go.
>
>
>Sarter has written extensively about “automation surprises,” often related to 
>control modes that the pilot does not fully understand or that the airplane 
>may have switched into autonomously, perhaps with an annunciation but without 
>the pilot’s awareness. Such surprises certainly added to the confusion aboard 
>Air France 447. One of the more common questions asked in cockpits today is 
>“What’s it doing now?” Robert’s “We don’t understand anything!” was an extreme 
>version of the same. Sarter said, “We now have this systemic problem with 
>complexity, and it does not involve just one manufacturer. I could easily list 
>10 or more incidents from either manufacturer where the problem was related to 
>automation and confusion. Complexity means you have a large number of 
>subcomponents and they interact in sometimes unexpected ways. Pilots don’t 
>know, because they haven’t experienced the fringe conditions that are built 
>into the system. 
>
>[... ] 
> At a time when accidents are extremely rare, each one becomes a one-off 
>event, unlikely to be repeated in detail. Next time it will be some other 
>airline, some other culture, and some other failure—but it will almost 
>certainly involve automation and will perplex us when it occurs. Over time the 
>automation will expand to handle in-flight failures and emergencies, and as 
>the safety record improves, pilots will gradually be squeezed from the cockpit 
>altogether. The dynamic has become inevitable. There will still be accidents, 
>but at some point we will have only the machines to blame.
>
>---- Morlock Elloi wrote ----
>
>>Handling of the recent B737 Max 8 disaster is somewhat revealing.
>>
>>What seems to have happened (for the 2nd time) is that computing machine 
>>fought the pilot, and the machine won.
>>
>>It looks like some cretin in Boeing that drank too much of AI Kool Aid 
>>(probably a middle manager) decided to install trained logic circuit 
>>that was supposed to make new aircraft behave (to pilots) like the older 
>>one. As its operation was far too complicated (ie. even Boeing didn't 
>>quite understand it) they decided not to inform pilots about it, as it 
>>could disturb the poor things with too much information.
>>
>>One part of the unknown operation appears to be the insistence of ML 
>>black box on crashing the airplane during ascent. As it had full control 
>>of the trim surfaces there was nothing pilots could do (I guess using 
>>fire axe to kill the circuit would work, if pilots knew where the damn 
>>thing was.)
>>
>>That's what the best available info right now is on what was the cause.
>>
>>What is interesting is how this was handled, particularly in the US:
>>
>>- There were documented complaints about this circuit for long time;
>>- FAA ignored them;
>>- After the second disaster most of the world grounded this type of 
>>aircraft;
>>- FAA said there is nothing wrong with it;
>>- It seems that intervention from White House made FAA see the light and 
>>ground the planes.
>>
>>Why? What was so special about this bug? FAA previously had no problem 
>>grounding planes on less evidence and fewer complaints.
>>
>>It may have to do with the first critical application of the new deity 
>>in commercial jets. The deity is called "AI", and its main function is 
>>to deflect the rage against rulers towards machines (it's the 2nd 
>>generation of the concept, the first one was simply "computer says ...".)
>>
>>FAA's hesitation may make sense. After several hundred people have been 
>>killed, someone will dig into the deity, and eventually the manager 
>>idiot and its minions will be declared (not publicly, of course) the 
>>guilty party. This could be a fatal blow to the main purpose of the deity.
>>
>>
>>(BTW, 'rage' is also a verb)
>>
>>
>>
>>
>>#  distributed via <nettime>: no commercial use without permission
>>#  <nettime>  is a moderated mailing list for net criticism,
>>#  collaborative text filtering and cultural politics of the nets
>>#  more info: http://mx.kein.org/mailman/listinfo/nettime-l
>>#  archive: http://www.nettime.org contact: nett...@kein.org
>>#  @nettime_bot tweets mail w/ sender unless #ANON is in Subject:
>
>#  distributed via <nettime>: no commercial use without permission
>#  <nettime>  is a moderated mailing list for net criticism,
>#  collaborative text filtering and cultural politics of the nets
>#  more info: http://mx.kein.org/mailman/listinfo/nettime-l
>#  archive: http://www.nettime.org contact: nett...@kein.org
>#  @nettime_bot tweets mail w/ sender unless #ANON is in Subject:
#  distributed via <nettime>: no commercial use without permission
#  <nettime>  is a moderated mailing list for net criticism,
#  collaborative text filtering and cultural politics of the nets
#  more info: http://mx.kein.org/mailman/listinfo/nettime-l
#  archive: http://www.nettime.org contact: nett...@kein.org
#  @nettime_bot tweets mail w/ sender unless #ANON is in Subject:

Reply via email to