TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Evidence for entropy maximisation in human free choice behaviour

103 pointsby Gaessakiover 4 years ago

14 comments

bArrayover 4 years ago
There is some work that explores this from an AI perspective with relative success [1]. It turns out that you can create quite intelligent agents if they look to maximize future entropy, with interesting results [2]. It&#x27;s still quite computationally expensive, but your normal search reduction tricks apply and you can get something computationally feasible.<p>[1] <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1310.1863" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1310.1863</a><p>[2] <a href="https:&#x2F;&#x2F;www.mdpi.com&#x2F;1099-4300&#x2F;16&#x2F;5&#x2F;2789" rel="nofollow">https:&#x2F;&#x2F;www.mdpi.com&#x2F;1099-4300&#x2F;16&#x2F;5&#x2F;2789</a>
评论 #25644192 未加载
评论 #25644285 未加载
FlyingSaucerover 4 years ago
This topic is extremely interesting, and good to see that experiments support it.<p>In AI lens:<p>In a way, you can compare this to novelty seeking and intelligent exploration which is quite an active field in Artificial Life and game AI[1]. If you find this interesting: Jeff Clune, Kenneth Stanley and Joel Lehman conducted interesting related research.<p>Also, isn&#x27;t this somehow related to the Free Energy principle by Karl Friston? If you look at entropy maximization as a way to minimize surprises.<p>[1] : <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1901.10995" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;1901.10995</a>
评论 #25653024 未加载
unabstover 4 years ago
Are we sure the availability of options equals entropy? It doesn&#x27;t appear as though we all act to simply increase our options. Preferring options over reward may also constitute delayed gratification and sacrifice, which is another interesting can of worms, but can it be predicated in terms of just preference of options over reward when those are your only two artificial options?<p>Human behavior appears to point towards the maximization of current order as an investment in power&#x2F;potential to drive future entropy, as opposed to simply maximizing entropy. This is the difference between building a nuclear bomb and keeping it, as opposed to building the bomb to use it. When one was used, it was meant to end a war, not start one. And success in life may as well be defined by hoarding order, be it technologically, financially, socially, or just objects. The pyramids were a feat in lowering entropy, not increasing it. And we love our diamonds.<p>This is also an extrapolation from the evidence in biology that energy entering a system increases order and contributes to the orderly structuring of matter and hence life [1].<p>[1] <a href="https:&#x2F;&#x2F;www.quantamagazine.org&#x2F;a-new-thermodynamics-theory-of-the-origin-of-life-20140122&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.quantamagazine.org&#x2F;a-new-thermodynamics-theory-o...</a>
yottaloveover 4 years ago
When my son has asked me what he should do about courses or employment or even vacation choices, I answer that he ought to choose the path that gives him the most choices.<p>He has thanked me many time for that advice, which has resulted in a high-value path for him.
hktover 4 years ago
On an intuitive level, this makes perfect sense. Assuming that entropy is, roughly speaking, novelty, that makes the calculation one about exploring new options for utility gains.<p>I recall seeing a study (although not where) suggesting novelty seeking was a key hallmark of intelligence. Maybe this means the entropy-utility calibration drives their intelligence? (Alongside their actual material circumstances)
评论 #25644437 未加载
gglonover 4 years ago
Also see Casual Entropic Forces [1] by A. D. Wissner-Gross [2] and C. E. Freer<p>[1] <a href="https:&#x2F;&#x2F;www.alexwg.org&#x2F;publications&#x2F;PhysRevLett_110-168702.pdf" rel="nofollow">https:&#x2F;&#x2F;www.alexwg.org&#x2F;publications&#x2F;PhysRevLett_110-168702.p...</a><p>[2] <a href="https:&#x2F;&#x2F;www.alexwg.org&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.alexwg.org&#x2F;</a>
dmichulkeover 4 years ago
aka Mobility heuristics<p>A very strong heuristics that works well in many games (exceptions are usually very interesting games) and is the root to other heuristic concepts such as piece value, central positioning, &quot;protected king&quot;, ... in Chess and similar concepts in, e.g., Starcraft.<p>Also very easy to implement, for discrete turn-based games it&#x27;s just the number of moves in a given state.<p><a href="http:&#x2F;&#x2F;ggp.stanford.edu&#x2F;lectures&#x2F;heuristics.pdf" rel="nofollow">http:&#x2F;&#x2F;ggp.stanford.edu&#x2F;lectures&#x2F;heuristics.pdf</a>
PeterisPover 4 years ago
Isn&#x27;t it pretty much the optimal behavior as evidenced e.g. by multi-armed bandit algorithms and explore-exploit balance in reinforcement learning?
评论 #25645288 未加载
_nhynesover 4 years ago
Of course. That&#x27;s why the first two versions of the Matrix failed.
r-zipover 4 years ago
Their definition of entropy is missing a negative sign...I understand that this is a preprint but come on.
johnsmith4739over 4 years ago
&quot;several studies have shown that individuals demonstrate a preference for choice, or the availability of multiple options, over and above utilitarian value.&quot; -&gt; yes, it is called the need for orientation&#x2F;control and &quot;utilitarian value&quot; has nothing to do with it -&gt; Index Funds vs. Actively-Managed Funds -&gt; people prefer the latter even if the returns are consistently lower. [1]<p>&quot;Yet we lack a decision-making framework that integrates preference for choice with traditional utility maximisation in free choice behaviour.&quot; -&gt; utility maximisation &quot;has charm for economists, but it rests on the shaky foundation of an implausible and untestable assumption&quot; - Daniel Kahneman [2] -&gt; TL;DR the author of &quot;Thinking Fast and Slow&quot; proves it false<p>&quot;We found that participants were biased towards states that kept their options open, even when both states were balanced in the total number of goal locations. This bias was evident not only when both contexts were equally valuable but throughout all value conditions...&quot; AND &quot;Participants were not informed of the precise values ...&quot; -&gt; seeing the utilitarian variable being forced upon conclusions is disheartening<p>[1] <a href="https:&#x2F;&#x2F;www.thebalance.com&#x2F;index-funds-vs-actively-managed-funds-2466445" rel="nofollow">https:&#x2F;&#x2F;www.thebalance.com&#x2F;index-funds-vs-actively-managed-f...</a> [2] <a href="https:&#x2F;&#x2F;papers.ssrn.com&#x2F;sol3&#x2F;papers.cfm?abstract_id=870494" rel="nofollow">https:&#x2F;&#x2F;papers.ssrn.com&#x2F;sol3&#x2F;papers.cfm?abstract_id=870494</a>
nemoniacover 4 years ago
Interesting paper. The exploration Vs exploitation payoff measured.
juskreyover 4 years ago
Look for Kelly criterion for the whole field on how to cook entropy
johndoe42377over 4 years ago
There is no such thing as evidence for abstract concepts.
评论 #25644669 未加载
评论 #25646925 未加载