Rule extraction from black box models is critical in domains that require model validation before implementation, as can be the case in credit scoring and medical diagnosis. Though already a challenging problem in statistical learning in general, the difficulty is even greater when highly nonlinear, recursive models, such as recurrent neural networks (RNNs), are fit to data. Here, we study the extraction of rules from second-order RNNs trained to recognize the Tomita grammars. We show that production rules can be stably extracted from trained RNNs and that in certain cases, the rules outperform the trained RNNs.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1162/neco_a_01111 | DOI Listing |
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!