TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Socratic Models – Composing Zero-Shot Multimodal Reasoning with Language

115 pointsby parsadotshabout 3 years ago

4 comments

nynxabout 3 years ago
This is super impressive. Transformers have consistently done better than almost anyone thought.<p>I still hold the opinion that we’re going to need to move to spiking neuron (SNN) models in the future to keep growing the networks. Spiking networks require lots of storage, but a lot, lot less compute. They also propagate additional information in the _timing_ of the spikes, not just the values. There are a lot of low-hanging fruit in SNNs and I think people are still trying to copy biological systems too much.<p>Unfortunately, the main issue with SNNs is that no one has figured out a way to train them as effectively as ANNs.
评论 #30980643 未加载
评论 #30984915 未加载
评论 #30980576 未加载
axg11about 3 years ago
At this point I&#x27;m comfortable in putting transformers as one of the top three developments in machine learning history. The way things are headed, they may turn out to be one of the most important &quot;discoveries&quot; ever made by humankind.<p>I&#x27;m extremely optimistic about how transformers can recursively speed up progress in multiple areas of science. Transformers are reaching a point where they can demonstrate reasoning abilities within the ballpark of what you might expect from a human. For certain qualities, they far exceed what any human is capable of. One of those areas being depth of knowledge. Transformers (e.g. RETRO) can incorporate a library of knowledge far larger than any human can. Soon we will improve and harness this ability to the point where it may be pointless to create a scientific hypothesis without first &quot;consulting&quot; a large language model that is able to process the entire library of scientific publications.
评论 #30984882 未加载
mountainriverabout 3 years ago
This is really awesome, multimodal is definitely where transformers are headed and holds the promise of solving a lot of the grounding issues we see with the current sota
评论 #30982369 未加载
arjvikabout 3 years ago
We&#x27;ve come to the consensus that large language models are just stochastic parrots... What makes us think that we can achieve a higher level of intelligence by putting them in conversation?<p>I think the next step in NLP will be a drastic innovation on today&#x27;s learning model.
评论 #30979935 未加载
评论 #30982416 未加载
评论 #30980044 未加载
评论 #30980915 未加载
评论 #30982813 未加载
评论 #30981134 未加载