Options

"there any reference to get more understanding the

bniyaseenbniyaseen Member Posts: 4 Contributor I
edited May 2019 in Help
my conclusion about this algorithm is : 
The algorithm starts with using dataset instances around each instance that have to explain, then use the model to create a prediction for those neighbors. After that, the correlation calculated between each instance attributes and the predictions to get the local weight for each attribute, the importance value of each attribute for current prediction is determined based on this weight. 

and after that, the weights aggregate for all attribute in the same prediction to get the global importance.

is that right??
thank you in advance:smile:
Tagged:

Best Answers

  • Options
    IngoRMIngoRM Administrator, Moderator, Employee, RapidMiner Certified Analyst, RapidMiner Certified Expert, Community Manager, RMResearcher, Member, University Professor Posts: 1,751 RM Founder
    edited December 2018 Solution Accepted
    Hi,
    Almost :smile:
    The key sentence is this here: "For each Example in an ExampleSet, this operator generates a neigboring set of data points, and uses correlation to identify the local attribute weights in that neighborhood."
    More details on your comments now:
    1. "The algorithm starts with using dataset instances around each instance that have to explain" - nope, we actually generate artificial data points around the point to be explained.  See above.
    2. "...then use the model to create a prediction for those neighbors." - correct.
    3. "...the correlation calculated between each instance attributes and the predictions to get the local weight for each attribute" - correct.
    4. "and after that, the weights aggregate for all attribute in the same prediction to get the global importance." - nope, the global importance is independent of the local importances.  In Auto Model we use again correlation for this but there are 30+ possibilities in RapidMiner (all the "Weight by..." operators).
    The algorithm is an improved variant of the well-known LIME algorithm.  The difference is that our implementation can deal with all types of data (categorical, numerical) for both the inputs and the labels.  We also calculate the weights in linear time which makes our algorithm much faster than the original LIME.
    Hope this helps,
    Ingo
Sign In or Register to comment.