Sarcouncil Journal of Engineering and Computer Sciences

Sarcouncil Journal of Engineering and Computer Sciences
An Open access peer reviewed international Journal
Publication Frequency- Monthly
Publisher Name-SARC Publisher
ISSN Online- 2945-3585
Country of origin-PHILIPPINES
Impact Factor- 3.7
Language- English
Keywords
- Engineering and Technologies like- Civil Engineering, Construction Engineering, Structural Engineering, Electrical Engineering, Mechanical Engineering, Computer Engineering, Software Engineering, Electromechanical Engineering, Telecommunication Engineering, Communication Engineering, Chemical Engineering
Editors

Dr Hazim Abdul-Rahman
Associate Editor
Sarcouncil Journal of Applied Sciences

Entessar Al Jbawi
Associate Editor
Sarcouncil Journal of Multidisciplinary

Rishabh Rajesh Shanbhag
Associate Editor
Sarcouncil Journal of Engineering and Computer Sciences

Dr Md. Rezowan ur Rahman
Associate Editor
Sarcouncil Journal of Biomedical Sciences

Dr Ifeoma Christy
Associate Editor
Sarcouncil Journal of Entrepreneurship And Business Management
Explainable AI: Building Trust and Transparency in Machine Learning Models
Keywords: Explainable artificial intelligence, interpretable machine learning, algorithmic transparency, user trust, model interpretability.
Abstract: Artificial intelligence systems increasingly influence critical decisions across healthcare, finance, and judicial domains, yet their complex architectures often operate as opaque black boxes. This creates significant barriers to user acceptance and regulatory compliance in high-stakes applications where understanding decision rationale proves essential. Explainable AI emerges as a fundamental requirement for building trustworthy systems that balance predictive performance with interpretability. Various techniques address interpretability challenges, including model-agnostic methods like LIME and SHAP, intrinsically interpretable architectures, and post-hoc explanation frameworks. However, significant technical challenges persist, including the accuracy-interpretability trade-off, computational overhead, and ensuring explanation fidelity. Human factors play crucial roles in determining explanation effectiveness, with user cognitive limitations and domain expertise influencing trust calibration. Successful implementation requires addressing both technical capabilities and human-centered design principles. The regulatory landscape increasingly demands algorithmic accountability, making explainability not merely beneficial but mandatory for many applications. Transparent AI systems demonstrate potential for enhanced user adoption, improved decision quality, and ethical compliance. Future developments must prioritize standardized evaluation metrics, scalable explanation methods, and domain-specific interpretability frameworks to realize the full potential of trustworthy artificial intelligence in society.
Author
- Amit Taneja
- UMB bank USA