Repository for the paper: Revisiting LLM Evaluation through Mechanism Interpretability: a New Metric and Model Utility Law
-
Updated
Apr 11, 2025 - Python
Repository for the paper: Revisiting LLM Evaluation through Mechanism Interpretability: a New Metric and Model Utility Law
A set of scripts to generate full attention-head heatmaps for transformer-based LLMs
A set of auxiliary systems designed to provide a measure of estimated confidence for the outputs generated by Large Language Models.
A set of scripts to generate full attention-head heatmaps for transformer-based LLMs
Add a description, image, and links to the llms-evalution topic page so that developers can more easily learn about it.
To associate your repository with the llms-evalution topic, visit your repo's landing page and select "manage topics."