TAB: Transformer Attention Bottlenecks enable User Intervention and Debugging in Vision-Language Models
Multi-head self-attention (MHSA) is a key component of Transformers, a widely popular architecture in both language and vision. Multiple heads...
Multi-head self-attention (MHSA) is a key component of Transformers, a widely popular architecture in both language and vision. Multiple heads...
Interpretability for Table Question Answering (Table QA) is critical, particularly in high-stakes industries like finance or healthcare. Although recent approaches...
An Achilles heel of Large Language Models (LLMs) is their tendency to hallucinate nonfactual statements. A response mixed of factual...
Most face identification approaches employ a Siamese neural network to compare two images at the image embedding level. Yet, this...
Nearest neighbors (NN) are traditionally used to compute final decisions, e.g., in Support Vector Machines or k-NN classifiers, and to...
* Equal contribution. CLIP-based classifiers rely on the prompt containing a {class name} that is known to the text encoder....
Face identification (FI) is ubiquitous and drives many high-stake decisions made by law enforcement. State-of-the-art FI approaches compare two images...
Explaining how important each input feature is to a classifier’s decision is critical in high-stake applications. An underlying principle behind...
Large-scale, multimodal models trained on web data such as OpenAI’s CLIP are becoming the foundation of many applications. Yet, they...
Explaining artificial intelligence (AI) predictions is increasingly important and even imperative in many high-stakes applications where humans are the ultimate...
Three important criteria of existing convolutional neural networks (CNNs) are (1) test-set accuracy; (2) out-of-distribution accuracy; and (3) explainability. While...
Explaining the decisions of an Artificial Intelligence (AI) model is increasingly critical in many real-world, high-stake applications. Hundreds of papers...
Interpretability methods often measure the contribution of an input feature to an image classifier’s decisions by heuristically removing it via...
* All authors contributed equally. Adversarial training has been the topic of dozens of studies and a leading method for...
* Equal contributions. Attribution methods can provide powerful insights into the reasons for a classifier’s decision. We argue that a...
Deep neural networks (DNNs) have demonstrated state-of-the-art results on many pattern recognition tasks, especially vision classification problems. Understanding the inner...
We can better understand deep neural networks by identifying which features each of their neurons have learned to detect. To...