Below you can see the output intrinsic attention of the model using the current code for different class_id's:
learn.intrinsic_attention(text=text,class_id=9,cmap=cm.RdYlGn_r)
TensorText([0.9443, 0.7151, 0.9200, 0.3891, 0.7568, 0.6479, 0.3891, 0.6266, 1.0000,
0.3891, 0.6479, 0.7568, 0.6417, 0.9947, 0.6417, 0.6417, 0.6417, 0.7568,
0.6479, 0.9313, 0.3891, 0.3891, 0.7151, 0.3891, 0.3891, 0.9313, 0.9200,
0.7151, 0.9255, 0.6417, 0.6417, 0.9947, 0.9255, 0.3891, 0.7568, 0.9200,
0.7151, 0.9255, 0.7151, 1.0000, 0.9255, 0.6417, 1.0000, 0.6417, 0.6417,
0.6417, 0.7568, 0.7754, 0.9255, 0.9291, 1.0000],
learn.intrinsic_attention(text=text,class_id=0,cmap=cm.RdYlGn_r)
ensorText([0.9443, 0.7151, 0.9200, 0.3891, 0.7568, 0.6479, 0.3891, 0.6266, 1.0000,
0.3891, 0.6479, 0.7568, 0.6417, 0.9947, 0.6417, 0.6417, 0.6417, 0.7568,
0.6479, 0.9313, 0.3891, 0.3891, 0.7151, 0.3891, 0.3891, 0.9313, 0.9200,
0.7151, 0.9255, 0.6417, 0.6417, 0.9947, 0.9255, 0.3891, 0.7568, 0.9200,
0.7151, 0.9255, 0.7151, 1.0000, 0.9255, 0.6417, 1.0000, 0.6417, 0.6417,
0.6417, 0.7568, 0.7754, 0.9255, 0.9291, 1.0000],
grad_fn=<AliasBackward>)
learn.intrinsic_attention(text=text,class_id=-1,cmap=cm.RdYlGn_r)
TensorText([0.9443, 0.7151, 0.9200, 0.3891, 0.7568, 0.6479, 0.3891, 0.6266, 1.0000,
0.3891, 0.6479, 0.7568, 0.6417, 0.9947, 0.6417, 0.6417, 0.6417, 0.7568,
0.6479, 0.9313, 0.3891, 0.3891, 0.7151, 0.3891, 0.3891, 0.9313, 0.9200,
0.7151, 0.9255, 0.6417, 0.6417, 0.9947, 0.9255, 0.3891, 0.7568, 0.9200,
0.7151, 0.9255, 0.7151, 1.0000, 0.9255, 0.6417, 1.0000, 0.6417, 0.6417,
0.6417, 0.7568, 0.7754, 0.9255, 0.9291, 1.0000],
grad_fn=<AliasBackward>)