Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Confusion about the Riemann Integral of expected gradients. #7

Open
LaZzyMan opened this issue Jan 12, 2021 · 0 comments
Open

Confusion about the Riemann Integral of expected gradients. #7

LaZzyMan opened this issue Jan 12, 2021 · 0 comments

Comments

@LaZzyMan
Copy link

Hello,
In the Section of Averaging Over Multiple Baselines the expected gradients is originally defined as the first integral as an expectation over D, and the second integral as an expectation over the path between x’ and x which is double integral. Then as the EG is belong to "the path atrribution methods", it becomes the Riemann Integral of one dimension.

In the origin paper the path methods are defined as integrating the gradients along a smooth path from x' to x. Expected gradient is the average of multiple path methods but these path methods have different origin point x', then why they can be considered as the " integrate gradients over one or more paths between two valid inputs"(there are more than two inputs). Then why it can be computed as the Riemann Integral of one dimension with a single path.

The second confusion stems from the figure(4) in Using the Training Distribution. I want to know what the blue line's value means in this picture. In A Better Understanding of Integrated Gradients the blue line in figure(4) represents the |f(x)-f(x')|, but in the train dataset expected gradient formula, the x' is not a constant value so why is the blue line represents a constant value. If it is the average of f(x') , it is reasonable that the blue line value changes along with the increasement of k.

Moreover, the gradient function is obviously not a linear function. In the Riemann Integral the differential part converge to x-E(x') while the function part never converge to gradient(x-E(x')), so I'm not sure that the expected gradient can really converge a constant value or a curve.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant