Приказ основних података о документу
When something goes wrong: Who is responsible for errors in ML decision-making?
dc.creator | Berber, Andrea | |
dc.creator | Srećković, Sanja | |
dc.date.accessioned | 2023-12-26T10:49:45Z | |
dc.date.available | 2023-12-26T10:49:45Z | |
dc.date.issued | 2023 | |
dc.identifier.issn | 0951-5666 | |
dc.identifier.uri | http://reff.f.bg.ac.rs/handle/123456789/5887 | |
dc.description.abstract | Because of its practical advantages, machine learning (ML) is increasingly used for decision-making in numerous sectors. This paper demonstrates that the integral characteristics of ML, such as semi-autonomy, complexity, and non-deterministic modeling have important ethical implications. In particular, these characteristics lead to a lack of insight and lack of comprehensibility, and ultimately to the loss of human control over decision-making. Errors, which are bound to occur in any decision-making process, may lead to great harm and human rights violations. It is important to have a principled way of assigning responsibility for such errors. The integral characteristics of ML, however, pose serious difficulties in defining responsibility and regulating ML decision-making. First, we elaborate on these characteristics and their epistemic and ethical implications. We then analyze possible general strategies for resolving the assignment of moral responsibility and show that, due to the specific way in which ML functions, each potential solution is problematic, whether we assign responsibility to humans, machines, or using hybrid models. Then, we shift focus on an alternative approach that bypasses moral responsibility and attempts to define legal liability independently through solutions such as informed consent and the no-fault compensation system. Both of these solutions prove unsatisfactory because they leave too much room for potential abuses of ML decision-making. We conclude that both ethical and legal solutions are fraught with serious difficulties. These difficulties prompt us to re-weigh the costs and benefits of using ML for high-stake decisions. | sr |
dc.language.iso | en | sr |
dc.publisher | Springer | sr |
dc.rights | closedAccess | sr |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | |
dc.source | AI & SOCIETY | sr |
dc.subject | Machine learning | |
dc.subject | Algorithmic decision-making | |
dc.subject | Opacity | |
dc.subject | Responsibility | |
dc.subject | Liability | |
dc.subject | Hybrid responsibility | |
dc.subject | Machine responsibility | |
dc.title | When something goes wrong: Who is responsible for errors in ML decision-making? | sr |
dc.type | article | sr |
dc.rights.license | BY | sr |
dc.identifier.doi | 10.1007/s00146-023-01640-1 | |
dc.type.version | publishedVersion | sr |
Документи
Датотеке | Величина | Формат | Преглед |
---|---|---|---|
Уз овај запис нема датотека. |