Cantitate/Preț
Produs

Accountable and Explainable Methods for Complex Reasoning over Text

Autor Pepa Atanasova
en Limba Engleză Paperback – 6 apr 2024
This thesis presents research that expands the collective knowledge in the areas of accountability and transparency of machine learning (ML) models developed for complex reasoning tasks over text. In particular, the presented results facilitate the analysis of the reasons behind the outputs of ML models and assist in detecting and correcting for potential harms. It presents two new methods for accountable ML models; advances the state of the art with methods generating textual explanations that are further improved to be fluent, easy to read, and to contain logically connected multi-chain arguments; and makes substantial contributions in the area of diagnostics for explainability approaches. All results are empirically tested on complex reasoning tasks over text, including fact checking, question answering, and natural language inference.
This book is a revised version of the PhD dissertation written by the author to receive her PhD from the Faculty of Science, University ofCopenhagen, Denmark. In 2023, it won the Informatics Europe Best Dissertation Award, granted to the most outstanding European PhD thesis in the field of computer science.

Citește tot Restrânge

Preț: 51273 lei

Preț vechi: 64091 lei
-20% Nou

Puncte Express: 769

Preț estimativ în valută:
9814 10228$ 8169£

Carte tipărită la comandă

Livrare economică 06-20 ianuarie 25

Preluare comenzi: 021 569.72.76

Specificații

ISBN-13: 9783031515170
ISBN-10: 303151517X
Ilustrații: XVIII, 199 p. 24 illus. in color.
Dimensiuni: 155 x 235 mm
Greutate: 0.31 kg
Ediția:2024
Editura: Springer Nature Switzerland
Colecția Springer
Locul publicării:Cham, Switzerland

Cuprins

1. Executive Summary.- Part I: Accountability for Complex Reasoning Tasks over Text.- 2. Fact Checking with Insufficient Evidence.- 3. Generating Label Cohesive and Well-Formed Adversarial Claims.- Part II: Explainability for Complex Reasoning Tasks over Text.- 4. Generating Fact Checking Explanations.- 5. Generating Fluent Fact Checking Explanations with Unsupervised Post-Editing.- 6. Multi-Hop Fact Checking of Political Claims.- Part III: Diagnostic Explainability Methods.- 7. A Diagnostic Study of Explainability Techniques for Text Classification.- 8. Diagnostics-Guided Explanation Generation.- 9. Recent Developments on Accountability and Explainability for Complex Reasoning Tasks.

Notă biografică

Pepa Atanasova is a postdoctoral researcher at the University of Copenhagen. She has received her PhD degree at the University of Copenhagen receiving the Best Dissertation Award of Informatics Europe in 2023. Her current research focuses on explainability for machine learning models, encompassing natural language explanations, post-hoc explainability methods, and adversarial attacks as well as the principled evaluation of existing explainability techniques.


Textul de pe ultima copertă

This thesis presents research that expands the collective knowledge in the areas of accountability and transparency of machine learning (ML) models developed for complex reasoning tasks over text. In particular, the presented results facilitate the analysis of the reasons behind the outputs of ML models and assist in detecting and correcting for potential harms. It presents two new methods for accountable ML models; advances the state of the art with methods generating textual explanations that are further improved to be fluent, easy to read, and to contain logically connected multi-chain arguments; and makes substantial contributions in the area of diagnostics for explainability approaches. All results are empirically tested on complex reasoning tasks over text, including fact checking, question answering, and natural language inference. This book is a revised version of the PhD dissertation written by the author to receive her PhD from the Faculty of Science, University of Copenhagen, Denmark. In 2023, it won the Informatics Europe Best Dissertation Award, granted to the most outstanding European PhD thesis in the field of computer science.

Caracteristici

Won the Informatics Europe Best Dissertation Award in 2023 for an outstanding thesis in the field of informatics Presents research that facilitates the analysis of the reasons behind the outputs of machine learning models Results are applicable to complex reasoning tasks like fact checking, question answering or natural language inference