๋‘ ํ™•๋ฅ  ๋ถ„ํฌ ์™€ ์˜ ์ฐจ์ด๋ฅผ ์ธก์ •ํ•˜๋Š” ๋ฐฉ๋ฒ•

H(P, Q) = H(P) + D\_{KL}(P||Q) = E\[\\log {Q(x)}\] = \sum\_{x \in \mathcal X} P(x) \log {Q(x)}
  • Information Entropy์™€ ์œ ์‚ฌํ•˜์ง€๋งŒ, ์˜ ์ •๋ณด๋Ÿ‰์„ ์ธก์ •ํ•˜๋Š” ๊ฒƒ์ด ์•„๋‹Œ, ์˜ ์ •๋ณด๋Ÿ‰์„ ์ธก์ •ํ•œ๋‹ค.
  • Kullback-Leibler Divergence์™€ ๋น„์Šทํ•˜๊ฒŒ ๋‘ ํ™•๋ฅ ๋ถ„ํฌ์˜ ์ฐจ์ด๋ฅผ ์ธก์ •ํ•œ๋‹ค.
  • ๋กœ ํ‘œํ˜„๋˜๋ฏ€๋กœ, Cross Entropy๋ฅผ ์ตœ์†Œํ™” ํ•˜๋Š” ๊ฒƒ์€ ๋ฅผ ์ตœ์†Œํ™” ํ•˜๋Š” ๊ฒƒ๊ณผ ๊ฐ™๋‹ค.

๋ฐฐ์šฐ๋Š” ์ด์œ 

  • KL Divergence๋Š” ๊ธฐ๋ณธ์ ์œผ๋กœ ๋ถ„์ˆ˜์˜ ํ˜•ํƒœ๋ฅผ ๋„๊ณ  ์žˆ์–ด ๊ณ„์‚ฐ์—์„œ ์–ด๋ ค์›€์ด ์žˆ๋‹ค.
    • p = 0์ผ ๋•Œ, log(0)์€ ๋ฌดํ•œ๋Œ€๋กœ ๋ฐœ์‚ฐํ•œ๋‹ค.
  • ํ•ด๋‹น ํ•จ์ˆ˜๊ฐ€ ๋ฏธ๋ถ„์ด ์‰ฝ๊ณ  ๋งค๋„๋Ÿฌ์›Œ ํ•™์Šต์— ์šฉ์ดํ•˜๋‹ค.