Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LLM offsets logic consolidate w/ checks and test case fix #1422

Closed
wants to merge 9 commits into from

Conversation

craymichael
Copy link
Contributor

Summary: Consolidate offsets logic with extra checks to one function. May be used to later group data in gradient LLM attribution. Test case fixed as a result of checks.

Differential Revision: D65010820

@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D65010820

Summary:

Slight bug in the retry GitHub workflow causes failing workflows to rerun indefinitely. This _should_ fix it

Differential Revision: D65008676
Summary:

Add __call__ to TokenizerLike for transformers compatibility

Differential Revision: D64998805
Summary:

Use the __call__ method of tokenizers that returns a BatchEncoding with offsets. This allows us to grab text from the fully decoded string and not make assumptions about how many tokens correspond to a single string.

Differential Revision: D64998804
Summary:

visualize_image_attr_multiple can return a List[Axes], adds proper annotations to satisfy mypy


Why casting is necessary: numpy/numpy#24738

https://github.com/matplotlib/matplotlib/blob/v3.9.2/lib/matplotlib/pyplot.py#L1583C41-L1584C1

Differential Revision: D64998799
Summary:

Adds enough typing to get rid of
`captum/metrics/_core/infidelity.py:498: note: By default the bodies of untyped functions are not checked, consider using --check-untyped-defs  [annotation-unchecked]`

Differential Revision: D64998800
Summary:

Fix pyre/mypy errors in infidelity.py. Introduce new BaselineTupleType

Differential Revision: D64998803
Summary:

Fix incorrect **kwargs annotation

Differential Revision: D65001879
Summary:

Add base class BaseLLMAttribution to consolidate repeat logic between perturbation/gradient-based LLM attr classes

Differential Revision: D65008854
Summary:

Consolidate offsets logic with extra checks to one function. May be used to later group data in gradient LLM attribution. Test case fixed as a result of checks.

Differential Revision: D65010820
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D65010820

@facebook-github-bot
Copy link
Contributor

This pull request has been merged in 492ae0e.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants