{"id":4756,"date":"2023-07-20T15:45:09","date_gmt":"2023-07-20T13:45:09","guid":{"rendered":"https:\/\/innovatrics.com\/trustreport\/?p=4756"},"modified":"2023-08-09T11:17:12","modified_gmt":"2023-08-09T09:17:12","slug":"ai-researcher-martin-tamajka","status":"publish","type":"post","link":"https:\/\/innovatrics.com\/trustreport\/ai-researcher-martin-tamajka\/","title":{"rendered":"AI researcher Martin Tamajka: \u201cIf we are to trust AI in courtrooms, it needs to justify its decisions.\u201d"},"content":{"rendered":"\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" width=\"2560\" height=\"1707\" src=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04166-scaled.jpg\" alt=\"\" class=\"wp-image-4760\" srcset=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04166-scaled.jpg 2560w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04166-300x200.jpg 300w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04166-1024x683.jpg 1024w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04166-1536x1024.jpg 1536w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04166-2048x1365.jpg 2048w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04166-300x200@2x.jpg 600w\" sizes=\"(max-width: 2560px) 100vw, 2560px\" \/><\/figure>\n\n\n\n<div class=\"wp-block-columns bottom0\">\n<div class=\"wp-block-column\" style=\"flex-basis:33.33%\"><\/div>\n\n\n\n<div class=\"wp-block-column\" style=\"flex-basis:66.66%\">\n<p class=\"has-medium-font-size\"><strong>AI is transforming jobs across a wide range of industries. However, there are still concerns about using it extensively when people&#8217;s lives or futures are at risk, such as in medicine or law. In these cases, it&#8217;s not enough for AI to just produce an answer \u2013 it also needs to be able to explain how it came up with that answer.<\/strong><\/p>\n<\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-columns top0 bottom0\">\n<div class=\"wp-block-column\" style=\"flex-basis:33.33%\"><\/div>\n\n\n\n<div class=\"wp-block-column\" style=\"flex-basis:66.66%\">\n<p>\u201cIf the algorithm can identify a patient as having dementia, but can&#8217;t explain what led to the diagnosis, the doctor will be unlikely to act upon it,\u201d claims Martin Tamajka, a lead research engineer at the Kempelen Institute of Intelligent Technologies (KInIT).&nbsp;<\/p>\n\n\n\n<p>To tackle this problem, he and other researchers at KInIT are conducting research in&nbsp; explainable AI (or \u201cXAI\u201d for short) \u2013 a branch of AI research that aims to provide an explanation of models\u2019 \u201cthought processes\u201d. But how do they make sure a model\u2019s explanations are understandable? Can explainable AI teach us more about the biases inherent in our society? And ultimately, can we always trust its explanations?<\/p>\n<\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-columns top0\">\n<div class=\"wp-block-column\" style=\"flex-basis:33.33%\"><\/div>\n\n\n\n<div class=\"wp-block-column\" style=\"flex-basis:66.66%\">\n<h4>Martin, can you tell us what explainable AI is and how we can imagine it at work? Is it essentially a \u201cChatGPT\u201d that can give you resources and reasoning?<\/h4>\n\n\n\n<p>First of all, let\u2019s look at the ordinary \u201cblack box AI\u201d. In this model you feed it data and train it for a specific task, for example: \u201ctell me whether there is a dog or a fish in the picture\u201d. We can see the input (the data you trained it with) and the output (the prediction), but we don\u2019t exactly know what happened in between to generate the prediction.<\/p>\n\n\n\n<p>Of course, we know the principle: in artificial neural networks, hundreds of thousands or even millions of nodes add or multiply the individual pixels to find structures characteristic of an image of a dog or a fish. In fact, they act as if they subsequently filtered the image \u2013 first, they detect simple structures like edges or blobs of different colours, and later they combine these structures into more abstract ones until they reach the final prediction. &nbsp;<\/p>\n\n\n\n<p>However, we don\u2019t know exactly what calculations and node combinations have led to that specific response \u2013 even though we can understand how multiplication and addition work, there are so many of them in this decision-making process that humans are not able to mentally contain them.<\/p>\n\n\n\n<h4>And this is where explainable AI could step in and show us its decision process?<\/h4>\n\n\n\n<p>Yes, explainable AI can show us what influences its decision-making. For example, in the dog\/fish-image scenario, explainable AI could highlight the sets of pixels that have influenced its decisions the most. So, if the picture had a rectangular shape and a lot of water, it\u2019s most likely a fish. Whereas if the picture had three big black blobs and big ears, it\u2019s most likely a dog.&nbsp;<\/p>\n\n\n\n<p>In this way, we could potentially verify whether the prediction of an AI system was based on relevant parts of the input data.<\/p>\n\n\n\n<p>So, to answer your question about explainable AI being able to reason, then the answer is no \u2013 no reasoning, or at least not for now. Even though ChatGPT might in many cases provide a certain level of reasoning, it\u2019s still far from being reliable. Many other AI models provide no reasoning at all. Hopefully, one day we will have mechanisms able to provide reliable and human-understandable explanations for at least the majority of models used in practice \u2013 that would be the holy grail of our work.<\/p>\n<\/div>\n<\/div>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" width=\"2508\" height=\"1484\" src=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz-1.png\" alt=\"\" class=\"wp-image-4850\" srcset=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz-1.png 2508w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz-1-300x178.png 300w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz-1-1024x606.png 1024w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz-1-1536x909.png 1536w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz-1-2048x1212.png 2048w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz-1-300x178@2x.png 600w\" sizes=\"(max-width: 2508px) 100vw, 2508px\" \/><\/figure>\n\n\n\n<div style=\"height:20px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-columns bottom0\">\n<div class=\"wp-block-column\" style=\"flex-basis:33.33%\"><\/div>\n\n\n\n<div class=\"wp-block-column\" style=\"flex-basis:66.66%\">\n<h4>How is explainable AI used today, and when is it preferred to a black box approach?<\/h4>\n\n\n\n<p>Explainable AI is a cornerstone for a new field of applications, mostly in areas that have a high impact on the lives of individuals.<\/p>\n\n\n\n<p>In my research, I have focused a lot on the use of artificial intelligence in medicine, and here, explainability is crucial. If our algorithm can identify a patient as having dementia, but can\u2019t explain what led to the diagnosis, the doctor will be unlikely to act upon it. The tests necessary to confirm the diagnosis might be inconvenient or even harmful, as could be the potential treatment \u2013 so without a clear explanation of why the algorithm suggests the diagnosis, the whole thing has little value.<\/p>\n\n\n\n<p>But certainly, XAI has a role to play in a variety of fields \u2013 such as self-driving cars, where it\u2019s possible to track the decision-making process that led to an accident; finance, where AI can aid with loan approvals; and even in automated weapon systems, where it can explain the reasoning behind launching a strike (which can also prevent the strike from being performed, if the human operator detects faulty reasoning).<\/p>\n\n\n\n<h4>What about the justice system? Is AI actively used there?<\/h4>\n\n\n\n<p>AI \u201chelpers\u201d have been used in adjudication for decades. Since the early 2000s, the COMPAS (Correctional Offender Management Profiling for Alternative Sanctions) algorithm developed by Northpointe, and many like it, have been used in the USA, advising judges by giving the defendant a \u201cgrade\u201d of how likely they are to re-offend.<\/p>\n\n\n\n<blockquote class=\"wp-block-quote\"><p>\u201cWe need to find the sweet spot between providing all the information that the AI system used to make its decision, and in just enough detail to reflect the accuracy of that decision, but without overwhelming the end-user receiving the explanation.\u201d<\/p><\/blockquote>\n<\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-columns top0\">\n<div class=\"wp-block-column margin\" style=\"flex-basis:33.33%\">\n<hr class=\"wp-block-separator\"\/>\n\n\n\n<h6>Wisconsin vs. Loomis<\/h6>\n\n\n\n<p class=\"has-small-font-size\">In 2013, Eric Loomis was found driving a car that had been involved in a shooting. To determine his sentence, the judge used both his previous criminal record as well as an output from the COMPAS algorithm. He was classified as being at high risk of re-offending, and sentenced to six years.&nbsp;<\/p>\n\n\n\n<p class=\"has-small-font-size\">COMPAS itself has faced controversy multiple times. Based on <a href=\"https:\/\/www.science.org\/doi\/10.1126\/sciadv.aao5580\">a study<\/a> by Julia Dressel and Hany Farid from Dartmouth College, it was found to be working with a level of precision similar to the judgement of random strangers on the internet.<\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column\" style=\"flex-basis:66.66%\">\n<p>Here we can see the problems with black box AI directly, most prominently in the case of Wisconsin vs. Loomis, where the defendant appealed the ruling because he felt that the judge made the decision based on an algorithm whose workings were secretive and could not be checked or validated. He argued that this was unfair and went against his right to have a fair trial.<\/p>\n\n\n\n<p>The appeal went up to the Wisconsin Supreme Court, who ultimately ruled against Loomis. Their ruling, however, urged caution in the use of algorithms. We mustn\u2019t forget that even felons are still people with rights \u2013 and the right to an explanation of a ruling is one of them. You must be able to show what you have based your ruling on, and here, explainable AI can help considerably.<\/p>\n<\/div>\n<\/div>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" width=\"2560\" height=\"1707\" src=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04323-scaled.jpg\" alt=\"\" class=\"wp-image-4841\" srcset=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04323-scaled.jpg 2560w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04323-300x200.jpg 300w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04323-1024x683.jpg 1024w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04323-1536x1024.jpg 1536w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04323-2048x1365.jpg 2048w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04323-300x200@2x.jpg 600w\" sizes=\"(max-width: 2560px) 100vw, 2560px\" \/><\/figure>\n\n\n\n<div style=\"height:5px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-columns bottom0\">\n<div class=\"wp-block-column\" style=\"flex-basis:33.33%\"><\/div>\n\n\n\n<div class=\"wp-block-column\" style=\"flex-basis:66.66%\">\n<h4>Explainability doesn\u2019t automatically have to mean understanding. With different levels of education, age and familiarity, how do you make sure the algorithm\u2019s explanations are understood by the recipient?<\/h4>\n\n\n\n<p>Explainability is all about walking the line between understandability and faithfulness.<\/p>\n\n\n\n<p>Let\u2019s take, for example, an explainable model similar to COMPAS. Like COMPAS, it would provide a recommendation on a prudent ruling, but it would also provide an explanation of why. But what should that explanation look like?<\/p>\n\n\n\n<p>It could be complete and faithful, and simply provide a long text file with all the data of the convict, with each word being assigned a \u201cweight\u201d of how much it affected the final decision. That would be open and transparent, but not really helpful, as it would be very difficult for the judge to understand.<\/p>\n\n\n\n<p>On the other hand, AI could \u201cexplain\u201d via an understandable justification presented as a shorter text, e.g. I advised parole because it was a minor charge and the defendant showed remorse. The risk here is that even though this might be partly true, there might have been a lot of other important details that led to the final decision, which you still don\u2019t know, so how \u201ccomplete and faithful\u201d is it really?<\/p>\n\n\n\n<p>In other words, we need to find the sweet spot between providing all the information that the AI system used to make its decision, and in just enough detail to reflect the accuracy of that decision, but without overwhelming the end-user receiving the explanation.<\/p>\n\n\n\n<h4>Your research at KInIT is trying to tackle this part of explainability. Can you tell us more about how you approach it?<\/h4>\n\n\n\n<p>We are looking for ways to give you the best of both worlds. One branch of our research is creating a meta-algorithm that applies multiple explainability algorithms to your problem and finds the one that gives the most satisfactory explanation.&nbsp;<\/p>\n\n\n\n<p>We call it AutoXAI \u2013 similar to how AutoML (Automated Machine Learning) tries to find the best machine learning model for the problem at hand, we try to find the best explanation for the problem with XAI algorithms.<\/p>\n\n\n\n<p>Through optimisation, we want to find an XAI algorithm that balances the two components mentioned above \u2013 understandability and faithfulness. Basically, we define a set of explainability algorithms and their parameters and we generate explanations of predictions generated by the machine learning model. Then, the question is \u201cWhich XAI algorithm is better\u201d?<\/p>\n\n\n\n<blockquote class=\"wp-block-quote\"><p>\u201cThe algorithm that produces better explanations is the one we want to use for our machine-learning model and data. Therefore, we need to define a set of metrics that measure the extent to which the predictions are understandable and faithful. This is actually the most difficult part of the whole process, which leads to the ultimate question \u2013 what is actually a good explanation?\u201d<\/p><\/blockquote>\n\n\n\n<p>To answer that question, we need to know how to measure the quality of explanations. Once we have that defined, we know that the algorithm that produces better explanations is the one we want to use for our machine-learning model and data. Therefore, we need to define a set of metrics that measure the extent to which the predictions are understandable and faithful. This is actually the most difficult part of the whole process, which leads to the ultimate question \u2013 what is actually a good explanation?&nbsp;<\/p>\n\n\n\n<p>The answer is usually not simple and it depends on the problem. For example, in some situations, such as a picture of a dog, an explanation in the form of highlighted pixels that most influenced the prediction might be the right one. In other cases, a textual explanation is better.<\/p>\n<\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-columns top0\">\n<div class=\"wp-block-column\" style=\"flex-basis:33.33%\"><\/div>\n\n\n\n<div class=\"wp-block-column\" style=\"flex-basis:66.66%\">\n<h4>Let\u2019s go back to the courtroom. You mentioned that we need a balance between not being overwhelmed, but also getting a detailed enough explanation of the decision. Is there any way that judges could ask for either a more or less detailed explanation?<\/h4>\n\n\n\n<p>One way how this can be tackled is something similar to a \u201csensitivity slider\u201d. A judge working on a case could ask the AI to give them a list of the ten most decisive factors for the suggested ruling. They can consider them, and if they need more granularity, just increase the sensitivity and see even more factors, which individually hold less weight but as a whole give a broader picture.&nbsp;<\/p>\n\n\n\n<p>In other words, the judge can first look at the most significant argument for or against sending someone to jail, and then they can look at an even broader picture if needed.<\/p>\n\n\n<div class=\"small-takeout-box\">\n    <div class=\"row\">\n        <div class=\"col-lg-5 col-4\">\n            <img width=\"227\" height=\"227\" src=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/innovatrics_detektiv_1-227x227.png\" class=\"attachment-small-takeout size-small-takeout\" alt=\"\" loading=\"lazy\" srcset=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/innovatrics_detektiv_1-227x227.png 227w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/innovatrics_detektiv_1-150x150.png 150w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/innovatrics_detektiv_1-186x186.png 186w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/innovatrics_detektiv_1-414x414.png 414w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/innovatrics_detektiv_1-227x227@2x.png 454w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/innovatrics_detektiv_1-150x150@2x.png 300w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/innovatrics_detektiv_1-186x186@2x.png 372w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/innovatrics_detektiv_1-414x414@2x.png 828w\" sizes=\"(max-width: 227px) 100vw, 227px\" \/>\n\n        <\/div>\n        <div class=\"col-lg-7 col-8\">\n            <div class=\"small-takeout-box__content\">\n                <a href=\"https:\/\/innovatrics.com\/trustreport\/former-nypd-inspector-joseph-courtesis\/\" class=\"small-takeout-box__read-more\" target=\"_blank\">Read more<\/a>\n                <h2 class=\"small-takeout-box__title\">\n                    <a href=\"https:\/\/innovatrics.com\/trustreport\/former-nypd-inspector-joseph-courtesis\/\" target=\"_blank\">\n                        Former NYPD inspector Joseph Courtesis: \u201cWe create less bias when we use facial recognition algorithms in our work.\u201d                    <\/a>\n                <\/h2>\n            <\/div>\n\n\n        <\/div>\n    <\/div>\n<\/div>\n\n\n\n<p>This also applies to different levels of understanding. An explanation for a skilled judge might consist of 50 different factors, uncovering a high level of detail and nuance. While the same system can explain the case to a first-year law student by using just the three most important factors, summing up the gist of the case with the most hard-hitting evidence.<\/p>\n\n\n\n<h4>While on the subject of court rulings, let\u2019s get to AI biases. There is mounting evidence that court rulings are biased toward certain groups of people. Can explainable AI help find those biases?<\/h4>\n\n\n\n<p>Of course, but it\u2019s not that the AI will directly tell you there is something wrong. But by examining which attributes matter more than others in decision making, you can see the bias of the training data that was projected into the model\u2026 or indeed, the world.<\/p>\n\n\n\n<p>There was a famous case where explainable AI was used to study how a machine-learning model classified different types of objects in pictures, including horses. They thought the algorithm worked brilliantly until they found out that instead of looking at important parts of the picture that should be relevant for the prediction, the algorithm just focused on the watermark. This led the model to implicitly create a rule like this \u2013 \u201cif there is a watermark, say \u2018horse\u2019\u201d.<\/p>\n\n\n\n<blockquote class=\"wp-block-quote\"><p>\u201cIt\u2019s not that the AI will directly tell you there is something wrong. But by examining which attributes matter more than others in decision making, you can see the bias of the training data\u2026 or indeed, the world.\u201d<\/p><\/blockquote>\n\n\n\n<p>If you use an XAI algorithm on thousands of court-case final verdicts and the predictions of the AI model that served as recommendations for the judge, you can get a nice image of how the underlying AI model actually creates its decision. In this way, you can potentially discover biases present in the model.&nbsp;<\/p>\n\n\n\n<p>To fix such biases you can clean your training data and retrain the model, adjust training strategy to explicitly punish the model when it makes a prediction based on some bias, or use another \u201cde-biasing\u201d procedure, such as deliberately including more of some specific types of rulings, or giving them more weight in the input.<\/p>\n<\/div>\n<\/div>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" width=\"2508\" height=\"1484\" src=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz2.png\" alt=\"\" class=\"wp-image-4847\" srcset=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz2.png 2508w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz2-300x178.png 300w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz2-1024x606.png 1024w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz2-1536x909.png 1536w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz2-2048x1212.png 2048w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-kolaz2-300x178@2x.png 600w\" sizes=\"(max-width: 2508px) 100vw, 2508px\" \/><\/figure>\n\n\n\n<div class=\"wp-block-columns bottom0\">\n<div class=\"wp-block-column\" style=\"flex-basis:33.33%\"><\/div>\n\n\n\n<div class=\"wp-block-column\" style=\"flex-basis:66.66%\">\n<h4>Knowing is the first part, but we might also actively do something against the biases. Could XAI help fight that, and is that something you are aiming for?<\/h4>\n\n\n\n<p>There are, of course, many ethical problems, on which I am not an expert. But to simplify it, you can aim for two things: equality of opportunity, or equality of outcome.<\/p>\n\n\n\n<p>Let\u2019s steer away from justice for a moment, and consider an AI system admitting children to a school. If you aim for equality of opportunity, AI will choose the 100 best-scoring students to be admitted.&nbsp;<\/p>\n\n\n\n<p>The upside is that you get the cleverest and most hard-working students \u2013 while everyone had the same opportunity to apply, only those with the best score were accepted. But if you look below the surface, you might find that most of those children are from wealthy families, as those children can more easily dedicate their time to studying, have the best tutors, etc.<\/p>\n\n\n\n<p>If you find that unfair, you might be in the camp of equality of outcome. This means that instead of just looking at the scores, you include measures like gender, race, wealth, social strata, etc., and try to admit the best people from different groups.<\/p>\n\n\n\n<p>What is great about this is that people who need education the most, have a shot at a better life. On the other hand, they might not be the best students for the opportunity. And other, more suitable students from other groups might have been left behind because there are only 100 admissions available.<\/p>\n\n\n\n<blockquote class=\"wp-block-quote\"><p>\u201cPeople have to learn to question AI. I believe that questioning it and not believing its outputs outright is what will, in the end, add to the acceptance of the technology.\u201d<\/p><\/blockquote>\n\n\n\n<p>Which option is correct? I can\u2019t tell you. I personally prefer equality of opportunity in most cases, but I also clearly see its downsides. If we want to change some pathology in our society, like increasing the education level of poor minorities in order to elevate these communities as a whole, the equality of outcome might be the right choice.<\/p>\n\n\n\n<p>To make it even more complicated, we don\u2019t always want to get rid of all the biases as some of them are wanted. For example, when sending out invitations for breast cancer screenings, we want AI to be biased towards women over 50. Because that is not a bad bias \u2013 it\u2019s a statistical fact, that incidentally acts as a bias.<\/p>\n\n\n\n<p>Recognising \u201cright\u201d and \u201cwrong\u201d biases, while right and wrong are painted with very broad brushstrokes, is a difficult task for ethicists, and its importance in the future will be immense.<\/p>\n<\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-columns top0\">\n<div class=\"wp-block-column\" style=\"flex-basis:33.33%\"><\/div>\n\n\n\n<div class=\"wp-block-column\" style=\"flex-basis:66.66%\">\n<h4>Let\u2019s finish up on the most important subject \u2013 responsibility. Who is responsible when the AI makes the decision? How do you think the future of responsibility for decisions will evolve?<\/h4>\n\n\n\n<p>I don\u2019t believe we should let any kind of AI do important decision making in the foreseeable future. I see AI and humans as more like co-workers.<\/p>\n\n\n\n<p>There are many things I would let AI do solely on its own. I will let it recommend me the best phone. The same goes for maybe 80\u201390% of the decision making in many professions. Let the AI do the tedious and repetitive work, so you have time and mental capacity to focus on the 10% of difficult cases.<\/p>\n\n\n\n<p>Especially in critical decision making, a human should always have the last word. AI can be an advisor to a doctor, judge or soldier, but they must make the final call and weigh out all the pros and cons.<\/p>\n\n\n\n<p>Let\u2019s not forget that AI does make mistakes. It\u2019s far from infallible. And it will never be!<\/p>\n\n\n\n<p>There will always be errors, and people have to learn to question AI. I believe that questioning it and not believing its outputs outright is what will, in the end, add to the acceptance of the technology.&nbsp;<\/p>\n\n\n\n<p>In these times, when we\u2019re observing the growth of very powerful generative models such as ChatGPT or Midjourney, we will have to learn how to check the trustworthiness of many things on the web, in the news or on TV, as creating things like Deepfake will be easy.<\/p>\n\n\n\n<p>So yes, we will always need humans to be making the final call and thinking critically about the outputs they get from the AI. Just as you don\u2019t believe everything you find on Google (or I would like to believe so), you will learn to use your common sense when working with AI. In this way, it can help us in many areas while keeping the risks minimal.<\/p>\n<\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-cover has-black-background-color has-background-dim top0\"><div class=\"wp-block-cover__inner-container\">\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" width=\"1024\" height=\"760\" src=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-profile-1024x760.png\" alt=\"\" class=\"wp-image-4859\" srcset=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-profile-1024x760.png 1024w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-profile-300x223.png 300w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-profile-1536x1140.png 1536w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-profile.png 2000w, https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/Tamajka-profile-300x223@2x.png 600w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<div style=\"height:20px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<p><a href=\"https:\/\/kinit.sk\/member\/martin-tamajka\/\">Martin Tamajka<\/a> has been working with different forms of AI in high-responsibility domains since his student years. In the past, he focused on the analysis of multidimensional medical images and image analysis in general. He now works on deep learning and computer vision and aims to increase the transparency and reliability of neural networks through methods of explainability and interpretability.<\/p>\n<\/div><\/div>\n\n\n\n<hr class=\"wp-block-separator\"\/>\n\n\n\n<p class=\"has-text-color has-small-font-size\" style=\"color:#7c878e\"><strong>AUTHOR<\/strong>: Andrej Kras<br><strong>PHOTOS: <\/strong>Dominika Beh\u00falov\u00e1<\/p>\n","protected":false},"excerpt":{"rendered":"<p>AI is transforming jobs across a wide range of industries. However, there are still concerns about using it extensively when people&#8217;s lives or futures are at risk, such as in medicine or law. In these cases, it&#8217;s not enough for AI to just produce an answer \u2013 it also needs to be able to explain how it came up with that answer.<\/p>\n","protected":false},"author":6,"featured_media":4760,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[1],"tags":[33,66,51],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v15.6.2 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>AI researcher Martin Tamajka: \u201cIf we are to trust AI in courtrooms, it needs to justify its decisions.\u201d - Innovatrics<\/title>\n<meta name=\"description\" content=\"AI is transforming jobs across a wide range of industries. However, there are still concerns about using it extensively when people&#039;s lives or futures are at risk, such as in medicine or law. In these cases, it&#039;s not enough for AI to just produce an answer \u2013 it also needs to be able to explain how it came up with that answer.\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/innovatrics.com\/trustreport\/ai-researcher-martin-tamajka\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"AI researcher Martin Tamajka: \u201cIf we are to trust AI in courtrooms, it needs to justify its decisions.\u201d - Innovatrics\" \/>\n<meta property=\"og:description\" content=\"AI is transforming jobs across a wide range of industries. However, there are still concerns about using it extensively when people&#039;s lives or futures are at risk, such as in medicine or law. In these cases, it&#039;s not enough for AI to just produce an answer \u2013 it also needs to be able to explain how it came up with that answer.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/innovatrics.com\/trustreport\/ai-researcher-martin-tamajka\/\" \/>\n<meta property=\"og:site_name\" content=\"Innovatrics\" \/>\n<meta property=\"article:published_time\" content=\"2023-07-20T13:45:09+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2023-08-09T09:17:12+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/OG-image-Tamajka.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1200\" \/>\n\t<meta property=\"og:image:height\" content=\"631\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Est. reading time\">\n\t<meta name=\"twitter:data1\" content=\"15 minutes\">\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebSite\",\"@id\":\"https:\/\/innovatrics.com\/trustreport\/#website\",\"url\":\"https:\/\/innovatrics.com\/trustreport\/\",\"name\":\"Innovatrics\",\"description\":\"Len \\u010fal\\u0161ia WordPress str\\u00e1nka\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":\"https:\/\/innovatrics.com\/trustreport\/?s={search_term_string}\",\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"en-US\"},{\"@type\":\"ImageObject\",\"@id\":\"https:\/\/innovatrics.com\/trustreport\/ai-researcher-martin-tamajka\/#primaryimage\",\"inLanguage\":\"en-US\",\"url\":\"https:\/\/innovatrics.com\/trustreport\/wp-content\/uploads\/2023\/07\/DSC04166-scaled.jpg\",\"width\":2560,\"height\":1707},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/innovatrics.com\/trustreport\/ai-researcher-martin-tamajka\/#webpage\",\"url\":\"https:\/\/innovatrics.com\/trustreport\/ai-researcher-martin-tamajka\/\",\"name\":\"AI researcher Martin Tamajka: \\u201cIf we are to trust AI in courtrooms, it needs to justify its decisions.\\u201d - Innovatrics\",\"isPartOf\":{\"@id\":\"https:\/\/innovatrics.com\/trustreport\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/innovatrics.com\/trustreport\/ai-researcher-martin-tamajka\/#primaryimage\"},\"datePublished\":\"2023-07-20T13:45:09+00:00\",\"dateModified\":\"2023-08-09T09:17:12+00:00\",\"author\":{\"@id\":\"https:\/\/innovatrics.com\/trustreport\/#\/schema\/person\/aa8ac119593dd2e3e7e0e0fe172a40fc\"},\"description\":\"AI is transforming jobs across a wide range of industries. However, there are still concerns about using it extensively when people's lives or futures are at risk, such as in medicine or law. In these cases, it's not enough for AI to just produce an answer \\u2013 it also needs to be able to explain how it came up with that answer.\",\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/innovatrics.com\/trustreport\/ai-researcher-martin-tamajka\/\"]}]},{\"@type\":\"Person\",\"@id\":\"https:\/\/innovatrics.com\/trustreport\/#\/schema\/person\/aa8ac119593dd2e3e7e0e0fe172a40fc\",\"name\":\"Kristina Sebejova\",\"image\":{\"@type\":\"ImageObject\",\"@id\":\"https:\/\/innovatrics.com\/trustreport\/#personlogo\",\"inLanguage\":\"en-US\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/3f3afee03f3588bdfd390a79fbeea359?s=96&d=mm&r=g\",\"caption\":\"Kristina Sebejova\"}}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","_links":{"self":[{"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/posts\/4756"}],"collection":[{"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/users\/6"}],"replies":[{"embeddable":true,"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/comments?post=4756"}],"version-history":[{"count":44,"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/posts\/4756\/revisions"}],"predecessor-version":[{"id":5449,"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/posts\/4756\/revisions\/5449"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/media\/4760"}],"wp:attachment":[{"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/media?parent=4756"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/categories?post=4756"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/innovatrics.com\/trustreport\/wp-json\/wp\/v2\/tags?post=4756"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}