README.md

July 31, 2025 ยท View on GitHub

Feather the Throttle: Revisiting Visual Token Pruning for Vision-Language Model Acceleration ArXiv PDF Project Page

Mark Endo , Xiaohan Wang , Serena Yeung-Levy

In International Conference on Computer Vision (ICCV) 2025

Our repository is split into two sections, one for model code and the other for evaluation.

Setup

  1. Clone the repository
git clone https://github.com/markendo/FEATHER
cd FEATHER
  1. Install packages
conda create -n feather python=3.10 -y
conda activate feather
cd prismatic-vlms
pip install -e .
cd ../vlm-evaluation
pip install -e .
cd ..
  1. Prepare data

The script for preparing evaluation datasets is at vlm-evaluation/scripts/datasets/prepare.py. More information about dataset preparation is available in the original codebase. Lastly, copy your HuggingFace token to vlm-evaluation/.hf_token.

Inference and Evaluation

We provide code for our experiments on evaluating various criteria for token pruning such as FastV, our modified version removing RoPE from the criteria, and our final FEATHER approach.

export DATASET_ROOT_DIR=/path/to/dataset/directory/
cd vlm-evaluation
bash scripts/eval_fastv.sh
bash scripts/eval_fastv_norope.sh
bash scripts/eval_feather.sh

Below are the results for RefCOCO and OCID-Ref.

CriteriaOCID-RefRefCOCOgRefCOCO+RefCOCO
FastV5.85.06.67.6
FastV w/o RoPE23.215.013.815.4
FEATHER32.538.738.743.4

The main implementation of FEATHER is provided in prismatic-vlms/prismatic/models/backbones/llm/llama2_models.py. Note that results can vary slightly based on attention implementation.

Acknowledgments

This repository is built on top of the prismatic-vlms and vlm-evaluation codebases.

Citation

@article{endo2025feather,
  author    = {Endo, Mark and Wang, Xiaohan and Yeung-Levy, Serena},
  title     = {Feather the Throttle: Revisiting Visual Token Pruning for Vision-Language Model Acceleration},
  journal   = {ICCV},
  year      = {2025},
}