Entropy
Last updated
Last updated
Entropy is a way to calculate the amount of "disorder" in a non-numeric column. Lower entropy indicates less disorder, while higher entropy indicates more.
The calculation for Shannon's entropy is: H = -Sum[ P(xi) * log2( P(xi)) ]
Name | Type | Description | Is Optional |
---|---|---|---|
group_by | column_list | Columns to group by | |
columns | column_list | Columns to calculate entropy on. Must be non-numeric. |