Leading the way in innovation for over 55 years, we build greater futures for businesses across multiple industries and 55 countries.
Our expert, committed team put our shared beliefs into action – every day. Together, we combine innovation and collective knowledge to create the extraordinary.
We share news, insights, analysis and research – tailored to your unique interests – to help you deepen your knowledge and impact.
At TCS, we believe exceptional work begins with hiring, celebrating and nurturing the best people — from all walks of life.
Get access to a catalog of the latest news stories from across TCS. Discover our press releases, reports, and company announcements.
Blog
Dr. P.H. Anantha Desik
You have these already downloaded
We have sent you a copy of the report to your email again.
Today, the use of artificial intelligence (AI) is so widespread that it encapsulates various domains, directly and indirectly, making a significant difference in almost all fields of work. AI applications are used in wide-ranging domains such as security, finance, healthcare, medicine, and criminal justice. Successful AI-based models have become increasingly complex and are difficult to express in easily understandable human terms, adversely affecting the adoption of AI models. AI model explainability depends on the quality of the data. If the data is not qualitatively rich, it will result in inaccurate insights, leading to unpredictable decisions in human context models.
Accuracy and interpretability in AI
AI model algorithms are classified based on the extent to which they can be interpreted or on the degree of accuracy. While linear and smooth relationship models such as classification, linear regression, decision trees, and K-Nearest Neighborhood (KNN), are easily interpretable, they are less accurate. On the other hand, support vector machines, random forest, and boosting algorithms are less interpretable but highly accurate. Neural network models are highly accurate but less interpretable. It has been observed that medium and less interpretable models require more explainability and typically use more explainable tools and frameworks. However, much depends on the data quality of the model being used or exposed.
AI explainability models
This has two approaches, model-based (where interpretable ML models are built), and post-hoc (where derived explanation is used for complex models). The model-based approach is linear, based more on input and output explainability. The post-hoc approach is either not known or a known design, based on data and its relationships. Model explainability approaches can further be categorized as model-agnostic, applicable to all model types, and model-specific, applicable only to specific models. Explainable AI depends on data types such as image, text, numeric, and audio and on explanation types such as visual, data points, and feature importance. These parameters for explainability can be mapped, derived, and processed from the original data types, parameters, distributions, and relations of the model approach. All depends on the quality and trustworthiness of the data, since without data, there is no explainability.
Explainability and interpretability are interchangeably used in the basic models of AI explainability. In order to ensure fair explainability for complex models, both interpretability and fidelity are required. Fidelity refers to accurate, elaborate descriptions of the models, while interpretability provides a single rationale of input and output relationships in a compact model form. Fidelity provides sufficient and truthful information for completeness of model explainability.
AI dependence on data quality
Both the model and post-hoc approaches of explainability depend on data inputs, feature importance, predictions, and business rules. Explainability depends on the extent to which the data is qualitative, quantitative, or biased for the AI model. Data patterns are based on data quality and relations. Low data quality leads to a skewed model and incorrect AI model explainability, adversely impacting human belief in the models.
AI models adhere to data quality of different dimensions such as accuracy, relevance, completeness, and constituency in data standardization. There are separately defined metrics to measure trustworthiness before considering data for the AI model data bias plays an important role for obtaining good data quality and for model explainability to remain unaffected in both model and post-hoc approaches. If it is not maintained, the impact is high which skews the model’s outcomes.
AI model development generally has a sequential approach for data processing, involving data operations ranging from treatment and wrangling to model building. It is important to capture the logs for data changes and distribution since they are vital for post transformation explainability, after model building. Data engineering logs play a significant role in the explainability of the data type, relations, data patterns, and changes in the post-hoc model. If data quality issues are not addressed initially, they can snowball into larger issues during the successive processes, impacting explainability, human decisions, and society at large.
AI model explainability is important for humans to understand and take decisions based on data while data quality is crucial in providing fair explainability of solutions. Research is ongoing to identify new explainability tools for complex AI models, as this will directly impact data quality, associated challenges and the way forward in overcoming these hurdles in AI model adoptability for human application.
The Role of AI in HRMS Industry
Cybersecurity: The new frontier in the digital age
Transforming Space Tech Operations Using STAR and ORBIT Frameworks
Understanding the EU CSRD Regulation and How TCS Can Help