TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Social learning: Collaborative learning with large language models

79 pointsby t3naryabout 1 year ago

7 comments

dr_kiszonkaabout 1 year ago
<a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;The_blind_leading_the_blind" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;The_blind_leading_the_blind</a>
falcor84about 1 year ago
Really interesting work there, and I particularly liked the gif-based story telling - <a href="https:&#x2F;&#x2F;blogger.googleusercontent.com&#x2F;img&#x2F;b&#x2F;R29vZ2xl&#x2F;AVvXsEgAndq_MjAVBs4j3lmxEX71nMrCLpAasklndZyE8F7yj3slyafRsNauzW4yRxI_Ncg7Sp5jllAXpItsjA-BOmdB2O1jP3Awu09-DVRHBE_Urf58yzm5tDBBpM-aibZxmgA9O6CySCCRdSMMqG7vj-OU07jHa0OU0YixCxRB0Q3APMQbn8Vz5rEBp70ZNogH&#x2F;s900&#x2F;image3.gif" rel="nofollow">https:&#x2F;&#x2F;blogger.googleusercontent.com&#x2F;img&#x2F;b&#x2F;R29vZ2xl&#x2F;AVvXsEg...</a>
评论 #39635268 未加载
评论 #39635951 未加载
omezeabout 1 year ago
So the main idea is that teachers can use examples with PII&#x2F;user data from gmail etc, but distill task-specific capabilities to students that don’t contain PII&#x2F;user data. Seems kinda useful for companies that want to utilize private information safely.<p>Edit: misunderstood the paper on first skim, they don’t actually want smaller students, they might actually want bigger ones. So this is more a privacy thing
评论 #39638573 未加载
v4dokabout 1 year ago
I wonder where this would be practical. How big would the teacher model need to be?
评论 #39638521 未加载
评论 #39637538 未加载
mediumsmartabout 1 year ago
When is the Large Google Machine going to learn to be social and offer me a <i>decline all non collaborative cookies</i> button?
acituanabout 1 year ago
Sounds like a stretch of the concept of social learning, and more like vanilla model distillation.<p>Social learning exist to transcend the limited processing power and limited training data exposure of individual agents through multimodal transfer of their own individual models (distilled down from an individual&#x27;s entire worldview, sense-of-self, perspectives, skills, semantic memory etc)<p>LLMs already exploit the propositional transfer of human models over language, and they abuse their massive compute capacity to compress them all in a giant model to simulate-them-all. For sure internally it does have some notion of distribution - as it at least has to distribute the compute at train time - but this is not an agent level distribution - not to confuse with the weaker metaphor of an &quot;agent&quot; used in model architectures -, and the end product presents itself as a singular &quot;agent&quot; with <i>all</i> of the processing power and <i>all</i> the training data that is infinitely copyable.<p>&gt; &quot;A teacher model provides instructions or few-shot examples to a student model without sharing its private data.&quot;<p>So the real concern is not utilizing social learning to transcend compute and training data limitations, it is about creating inferior models that can be distributed back into the world without giving up all of the secret sauce.<p>For sure this could work, one could create inferior &quot;agents&quot; from stronger &quot;agents&quot;, but we cannot create an even stronger &quot;agent&quot; through the dialogue of two strong &quot;agent&quot;s, because everything to be shared is already perfectly encoded in the model&amp;architecture and perfectly copyable. Therefore this is not social learning at all.<p>To abuse back their anthropomorphization, they are trying to create a deliberately stupid kid to send out to the world so that the kid doesn&#x27;t tell all the things mommy and daddy already knows and could have perfectly taught. Because one can make more money from selling&#x2F;renting a bundle of differently stupid agents than a singular state-of-the-art one I guess?
llmzeroabout 1 year ago
Think about the following scenario: I write a calculus book and the agents of this model just modify every example and every definition and change a little the ordering of the material to teach students. Now they are using my book but it seems they are not using my book. Are they trying to copy without copying?