🍋
Menu
Math

Entropy (Information)

Shannon Information Entropy

A measure of uncertainty or information content in a dataset, calculated as the negative sum of probabilities times their logarithms.

รายละเอียดทางเทคนิค

Entropy (Information) answers 'to what power must the base be raised to produce this value?' Common bases: natural log (ln, base e ≈ 2.718), common log (log10), and binary log (log2, used in computing for bit-depth and information theory). Logarithmic scales are used for pH (chemistry), decibels (audio), Richter magnitude (seismology), and algorithm complexity analysis (O(log n)). The slide rule — the engineer's calculator before electronics — operated entirely on logarithmic principles.

ตัวอย่าง

```
Logarithm identities:

  log_b(x) = y  means  b^y = x

  log₁₀(1000) = 3   (10³ = 1000)
  log₂(256)   = 8   (2⁸ = 256)
  ln(e)       = 1   (e¹ = e ≈ 2.718)

  log(a × b) = log(a) + log(b)
  log(a / b) = log(a) - log(b)
  log(a^n)   = n × log(a)
```

เครื่องมือที่เกี่ยวข้อง

คำศัพท์ที่เกี่ยวข้อง