This doesn’t seem that complex to me. If there is a pedestrian in front of your car when the light turns green, you wait. Pretty fucking simple. This isn’t some offshoot of the trolley problem where an incident was unavoidable. The car made the active decision to proceed when it was not safe to do so.
Why have we programmed our self-driving cars to emulate the psychotic behaviour of a typical road ragin’ car-brained human? Isn’t that the problem these projects should be trying to solve?
I’m going to inject some unpopular nuance here, so I’ll preface by admitting that I haven’t looked further into this event than the information provided in the linked article, which isn’t much. Nevertheless, a few points:
No system is perfect, including exclusively human drivers. Obviously zero accidents is ideal, but as you said, road ragin’ car-brained behavior is typical. How many people are killed every year by human drivers?
Obviously driverless system development should aspire to dynamic reactivity comparable to the best human driver. But when running a cost-benefit analysis for driverless adoption it’s worth considering if, normalizing each by their respective total hours-on-the-road, the mistakes made by driverless cars due to rigid adhesion to traffic laws outnumber the mistakes made by drivers due to their own flagrant disobedience.
I would respond that in court, the traffic laws are the traffic laws. It looks like the pedestrian is the accountable party here. But from a pedestrian perspective, cars that are only dangerous if you’re jaywalking are objectively an upgrade.
Reminder that paint isn’t infrastructure so if you want people to use the “crosswalk” then it needs to be raised up and it needs a protected middle if you have to cross multiple lanes. If you make the crosswalk actually safer to use then people will use it.
I’d argue that human drivers are absolutely not held accountable in the US. When my buddy was killed by a driver texting in a giant SUV they gave the driver a small fine and called it an “accident”.
That’s an interesting comparison and something I’ve wondered about quite a bit. I would be surprised if machine drivers were not categorically safer than human ones, and if safety is (rightly) a priority in the cost-benefit analysis of driverless car adoption, then it’s hard to imagine not concluding that we ought to proceed in that direction.
But I think this specific incident illustrates very well that the human vs. machine driver debate is tragically myopic. If an infallible machine driver adhering perfectly to traffic laws is empowered to accelerate from a standstill directly into a violent collision with a pedestrian, then maybe it doesn’t matter how “safe” the driver is. I take it as evidence that car travel the way we have it set up is inherently unsafe. Our traffic laws emphasize the convenience of car traffic above everything else – including safety – and only really serve to shift blame when something goes wrong. Despite its certainty, there is very little builtin allowance for human error aside from the begrudging mercy of other parties.
To be fair, human drivers are an unmitigated disaster which we really need to do something about, but I think if we’re going to go through the messy process of reforming how we think about cars, we might as well go farther than a marginal improvement. We could solve the underlying problem and abolish the institution of car dependency altogether, for instance. Otherwise it just amounts to slapping a futuristic band-aid on a set of social and economic issues that will continue to cause unimaginable harm.
This doesn’t seem that complex to me. If there is a pedestrian in front of your car when the light turns green, you wait. Pretty fucking simple. This isn’t some offshoot of the trolley problem where an incident was unavoidable. The car made the active decision to proceed when it was not safe to do so.
Why have we programmed our self-driving cars to emulate the psychotic behaviour of a typical road ragin’ car-brained human? Isn’t that the problem these projects should be trying to solve?
Because Elon Musk was involved at some point
I’m going to inject some unpopular nuance here, so I’ll preface by admitting that I haven’t looked further into this event than the information provided in the linked article, which isn’t much. Nevertheless, a few points:
No system is perfect, including exclusively human drivers. Obviously zero accidents is ideal, but as you said, road ragin’ car-brained behavior is typical. How many people are killed every year by human drivers?
Obviously driverless system development should aspire to dynamic reactivity comparable to the best human driver. But when running a cost-benefit analysis for driverless adoption it’s worth considering if, normalizing each by their respective total hours-on-the-road, the mistakes made by driverless cars due to rigid adhesion to traffic laws outnumber the mistakes made by drivers due to their own flagrant disobedience.
Removed by mod
I would respond that in court, the traffic laws are the traffic laws. It looks like the pedestrian is the accountable party here. But from a pedestrian perspective, cars that are only dangerous if you’re jaywalking are objectively an upgrade.
Reminder that paint isn’t infrastructure so if you want people to use the “crosswalk” then it needs to be raised up and it needs a protected middle if you have to cross multiple lanes. If you make the crosswalk actually safer to use then people will use it.
I’d argue that human drivers are absolutely not held accountable in the US. When my buddy was killed by a driver texting in a giant SUV they gave the driver a small fine and called it an “accident”.
That’s an interesting comparison and something I’ve wondered about quite a bit. I would be surprised if machine drivers were not categorically safer than human ones, and if safety is (rightly) a priority in the cost-benefit analysis of driverless car adoption, then it’s hard to imagine not concluding that we ought to proceed in that direction.
But I think this specific incident illustrates very well that the human vs. machine driver debate is tragically myopic. If an infallible machine driver adhering perfectly to traffic laws is empowered to accelerate from a standstill directly into a violent collision with a pedestrian, then maybe it doesn’t matter how “safe” the driver is. I take it as evidence that car travel the way we have it set up is inherently unsafe. Our traffic laws emphasize the convenience of car traffic above everything else – including safety – and only really serve to shift blame when something goes wrong. Despite its certainty, there is very little builtin allowance for human error aside from the begrudging mercy of other parties.
To be fair, human drivers are an unmitigated disaster which we really need to do something about, but I think if we’re going to go through the messy process of reforming how we think about cars, we might as well go farther than a marginal improvement. We could solve the underlying problem and abolish the institution of car dependency altogether, for instance. Otherwise it just amounts to slapping a futuristic band-aid on a set of social and economic issues that will continue to cause unimaginable harm.
How do you intend to inject nuance of all things when you haven’t even bothered to read the article. Honestly.
Uh, I did read the article.
Hold on now… you can’t expect everyone to just read your whole comment…
Won’t it be fun if the car failed to see the person because it’s ai was trained on white Americans and there were no Chinese in the data set