From f37019a12bbecbcf68120cf517b3e45bbc5f4807 Mon Sep 17 00:00:00 2001
From: Amit Parekh <7276308+amitkparekh@users.noreply.github.com>
Date: Wed, 3 Jul 2024 19:42:39 +0100
Subject: [PATCH] docs: more formatting and pruning
---
README.md | 13 +------------
1 file changed, 1 insertion(+), 12 deletions(-)
diff --git a/README.md b/README.md
index 3a0d663..c77fe3d 100644
--- a/README.md
+++ b/README.md
@@ -18,8 +18,6 @@
----
-
Evaluating the generalisation capabilities of multimodal models based solely on their performance on out-of-distribution data fails to capture their true robustness. This work introduces a comprehensive evaluation framework that systematically examines the role of instructions and inputs in the generalisation abilities of such models, considering architectural design, input perturbations across language and vision modalities, and increased task complexity. The proposed framework uncovers the resilience of multimodal models to extreme instruction perturbations and their vulnerability to observational changes, raising concerns about overfitting to spurious correlations. By employing this evaluation framework on current Transformer-based multimodal models for robotic manipulation tasks, we uncover limitations and suggest future advancements should focus on architectural and training innovations that better integrate multimodal inputs, enhancing a model's generalisation prowess by prioritising sensitivity to input content over incidental correlations.
@@ -30,7 +28,7 @@ Evaluating the generalisation capabilities of multimodal models based solely on
![Table of perturbations from the paper](docs/PERT%20Table.png)