TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Kolmogorov-Arnold Networks

568 点作者 sumo43大约 1 年前

33 条评论

GistNoesis大约 1 年前
I quickly skimmed the paper, got inspired to simplify it, and created some Pytorch Layer :<p><a href="https:&#x2F;&#x2F;github.com&#x2F;GistNoesis&#x2F;FourierKAN&#x2F;">https:&#x2F;&#x2F;github.com&#x2F;GistNoesis&#x2F;FourierKAN&#x2F;</a><p>The core is really just a few lines.<p>In the paper they use some spline interpolation to represent 1d function that they sum. Their code seemed aimed at smaller sizes. Instead I chose a different representation, aka fourier coefficients that are used to interpolate the functions of individual coordinates.<p>It should give an idea of Kolmogorov-Arnold networks representation power, it should probably converge easier than their spline version but spline version have less operations.<p>Of course, if my code doesn&#x27;t work, it doesn&#x27;t mean theirs doesn&#x27;t.<p>Feel free to experiment and publish paper if you want.
评论 #40235456 未加载
评论 #40247231 未加载
评论 #40233557 未加载
评论 #40228108 未加载
评论 #40224332 未加载
krasin大约 1 年前
I&#x27;ve spent some time playing with their Jupyter notebooks. The most useful (to me, anyway) is their Example_3_classfication.ipynb ([1]).<p>It works as advertised with the parameters selected by the authors, but if we modified the network shape in the second half of the tutorial (Classification formulation) from (2, 2) to (2, 2, 2), it fails to generalize. The training loss gets down to 1e-9, while test loss stays around 3e-1. Getting to larger network sizes does not help either.<p>I would really like to see a bigger example with many more parameters and more data complexity and if it could be trained at all. MNIST would be a good start.<p>Update: I increased the training dataset size 100x, and that helps with the overfitting, but now I can&#x27;t get training loss below 1e-2. Still iterating on it; a GPU acceleration would really help - right now, my progress is limited by the speed of my CPU.<p>1. <a href="https:&#x2F;&#x2F;github.com&#x2F;KindXiaoming&#x2F;pykan&#x2F;blob&#x2F;master&#x2F;tutorials&#x2F;Example_3_classfication.ipynb">https:&#x2F;&#x2F;github.com&#x2F;KindXiaoming&#x2F;pykan&#x2F;blob&#x2F;master&#x2F;tutorials&#x2F;...</a>
评论 #40220925 未加载
评论 #40230067 未加载
评论 #40233633 未加载
esafak大约 1 年前
There exists a Kolmogorov-Arnold inspired model in classical statistics called GAMs (<a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Generalized_additive_model" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Generalized_additive_model</a>), developed by Hastie and Tibshirani as an extension of GLMs (<a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Generalized_linear_model" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Generalized_linear_model</a>).<p>GLMs in turn generalize logistic-, linear and other popular regression models.<p>Neural GAMs with learned basis functions have already been proposed, so I&#x27;m a bit surprised that the prior art is not mentioned in this new paper. Previous applications focused more on interpretability.
评论 #40264761 未加载
montebicyclelo大约 1 年前
The success we&#x27;re seeing with neural networks is tightly coupled with the ability to scale - the algorithm itself works at scale (more layers), but it also scales well with hardware, (neural nets mostly consist of matrix multiplications, and GPUs have specialised matrix multiplication acceleration) - one of the most impactful neural network papers, AlexNet, was impactful because it showed that NNs could be put on the GPU, scaled and accelerated, to great effect.<p>It&#x27;s not clear from the paper how well this algorithm will scale, both in terms of the algorithm itself (does it still train well with more layers?), and ability to make use of hardware acceleration, (e.g. it&#x27;s not clear to me that the structure, with its per-weight activation functions, can make use of fast matmul acceleration).<p>It&#x27;s an interesting idea, that seems to work well and have nice properties on a smaller scale; but whether it&#x27;s a good architecture for imagenet, LLMs, etc. is not clear at this stage.
评论 #40220823 未加载
cs702大约 1 年前
It&#x27;s so <i>refreshing</i> to come across new AI research different from the usual &quot;we modified a transformer in this and that way and got slightly better results on this and that benchmark.&quot; All those new papers proposing incremental improvements are important, but... everyone is getting a bit tired of them. Also, anecdotal evidence and recent work suggest we&#x27;re starting to run into fundamental limits inherent to transformers, so we may well need new alternatives.[a]<p>The best thing about this new work is that it&#x27;s not an either&#x2F;or proposition. The proposed &quot;learnable spline interpolations as activation functions&quot; can be used <i>in conventional DNNs</i>, to improve their expressivity. Now we just have to test the stuff to see if it really works better.<p>Very nice. Thank you for sharing this work here!<p>---<p>[a] <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=40179232">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=40179232</a>
评论 #40225436 未加载
评论 #40225676 未加载
评论 #40223792 未加载
mxwsn大约 1 年前
From the preprint - 100 input dimensions is considered &quot;high&quot;, and most problems considered have 5 or fewer input dimensions. This is typical of physics-inspired settings I&#x27;ve seen considered in ML. The next step would be demonstrating them on MNIST, which, at 784 dimensions is tiny by modern standards.
评论 #40221265 未加载
ubj大约 1 年前
Very interesting! Kolmogorov neutral networks can represent discontinuous functions [1], but I&#x27;ve wondered about how practically applicable they are. This repo seems to show that they have some use after all.<p>[1]: <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2311.00049" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2311.00049</a>
评论 #40231434 未加载
reynoldss大约 1 年前
Perhaps a hasty comment but linear combinations of B-splines are yet another (higher-degree) B-spline. Isn&#x27;t this simply fitting high degree B-splines to functions?
评论 #40225550 未加载
Lichtso大约 1 年前
1. Interestingly the foundations of this approach and MLP were invented &#x2F; discovered around the same time about 66 years ago:<p>1957: <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Kolmogorov%E2%80%93Arnold_representation_theorem" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Kolmogorov%E2%80%93Arnold_repr...</a><p>1958: <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Multilayer_perceptron" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Multilayer_perceptron</a><p>2. Another advantage of this approach is that it has only one class of parameters (the coefficients of the local activation functions) as opposed to MLP which has three classes of parameters (weights, biases, and the globally uniform activation function).<p>3. Everybody is talking transformers. I want to see diffusion models with this approach.
评论 #40228022 未加载
评论 #40229957 未加载
评论 #40228750 未加载
评论 #40228431 未加载
cbsmith大约 1 年前
Feels like someone stuffed splines into decision trees.
评论 #40221182 未加载
评论 #40222089 未加载
adityang5大约 1 年前
Very cool stuff! Exciting to see so many people sharing their works on KANs. Seeing as the authors claim that KANs are able to reduce the issues of catastrophic forgetting that we see in MLPs, I thought &quot;Wouldn&#x27;t it be nice if there was an LLM that substituted MLPs with KANs?&quot;. I looked around and didn&#x27;t find one, so I built one!<p>- PyTorch Module of the KAN GPT<p>- Deployed to PyPi<p>- MIT Licence<p>- Test Cases to ensure forward-backward passes work as expected<p>- Training script<p>I am currently working on training it on the WebText dataset to compare it to the original gpt2. Facing a few out-of-memory issues at the moment. Perhaps the vocab size (50257) is too large?<p>I&#x27;m open to contributions and would love to hear your thoughts!<p><a href="https:&#x2F;&#x2F;github.com&#x2F;AdityaNG&#x2F;kan-gpt">https:&#x2F;&#x2F;github.com&#x2F;AdityaNG&#x2F;kan-gpt</a><p><a href="https:&#x2F;&#x2F;pypi.org&#x2F;project&#x2F;kan-gpt&#x2F;" rel="nofollow">https:&#x2F;&#x2F;pypi.org&#x2F;project&#x2F;kan-gpt&#x2F;</a>
cloudhan大约 1 年前
This reminds me of Weight Agnostic Neural Networks <a href="https:&#x2F;&#x2F;weightagnostic.github.io&#x2F;" rel="nofollow">https:&#x2F;&#x2F;weightagnostic.github.io&#x2F;</a>
apolar大约 1 年前
Article 2021: <a href="https:&#x2F;&#x2F;www.sciencedirect.com&#x2F;science&#x2F;article&#x2F;abs&#x2F;pii&#x2F;S0952197620303742" rel="nofollow">https:&#x2F;&#x2F;www.sciencedirect.com&#x2F;science&#x2F;article&#x2F;abs&#x2F;pii&#x2F;S09521...</a><p>Seminar 2021: <a href="https:&#x2F;&#x2F;warwick.ac.uk&#x2F;fac&#x2F;sci&#x2F;maths&#x2F;research&#x2F;events&#x2F;seminars&#x2F;areas&#x2F;applmath&#x2F;2020-21&#x2F;#WEEK1" rel="nofollow">https:&#x2F;&#x2F;warwick.ac.uk&#x2F;fac&#x2F;sci&#x2F;maths&#x2F;research&#x2F;events&#x2F;seminars...</a><p>Article in archive 2023: <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2305.08194" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2305.08194</a><p>Video 2021: <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=eS_k6L638k0" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=eS_k6L638k0</a><p>Extension to stochastic models where KAN builds the distribution 2023: <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=0hhJIpzxPR0" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=0hhJIpzxPR0</a>
评论 #40266162 未加载
yobbo大约 1 年前
<a href="https:&#x2F;&#x2F;kindxiaoming.github.io&#x2F;pykan&#x2F;intro.html" rel="nofollow">https:&#x2F;&#x2F;kindxiaoming.github.io&#x2F;pykan&#x2F;intro.html</a><p>At the end of this example, they recover the symbolic formula that generated their training set: exp(x₂² + sin(3.14x₁)).<p>It&#x27;s like a computation graph with a library of &quot;activation functions&quot; that is optimised, and then pruned. You can recover good symbolic formulas from the pruned graph.<p>Maybe not meaningful for MNIST.
评论 #40225764 未加载
diwank大约 1 年前
It’d be really cool to see a transformer with the MLP layers swapped for KANs and then compare its scaling properties with vanilla transformers
评论 #40224506 未加载
评论 #40219995 未加载
评论 #40220987 未加载
SpaceManNabs大约 1 年前
How does back propagation work now? Do these suffer from vanishing or exploding gradients?
评论 #40226155 未加载
评论 #40227168 未加载
phpkar大约 1 年前
<a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2404.05903" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2404.05903</a>
ALittleLight大约 1 年前
I can&#x27;t assess this, but I do worry that overnight some algorithmic advance will enhance LLMs by orders of magnitude and the next big model to get trained is suddenly 10,000x better than GPT-4 and nobody&#x27;s ready for it.
评论 #40220191 未加载
评论 #40226834 未加载
评论 #40220169 未加载
erwincoumans大约 1 年前
If you like this, you may also like this 2019 research paper: &quot;Deep networks and the Kolmogorov–Arnold theorem&quot; <a href="https:&#x2F;&#x2F;hadrien-montanelli.github.io&#x2F;2019-06-25.html" rel="nofollow">https:&#x2F;&#x2F;hadrien-montanelli.github.io&#x2F;2019-06-25.html</a> <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;1906.11945" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;1906.11945</a>
mipt98大约 1 年前
A more elaborate implementation of this was published years ago, and it wasn&#x27;t the very first one <a href="https:&#x2F;&#x2F;www.science.org&#x2F;doi&#x2F;10.1126&#x2F;science.1165893" rel="nofollow">https:&#x2F;&#x2F;www.science.org&#x2F;doi&#x2F;10.1126&#x2F;science.1165893</a>
评论 #40258342 未加载
kevmo314大约 1 年前
This seems very similar in concept to the finite element method. Nice to see patterns across fields like that.
syassami大约 1 年前
Nice implementation I&#x27;ve been playing with, <a href="https:&#x2F;&#x2F;github.com&#x2F;Blealtan&#x2F;efficient-kan">https:&#x2F;&#x2F;github.com&#x2F;Blealtan&#x2F;efficient-kan</a> alongside @GistNoesis&#x27;s.
Maro大约 1 年前
Interesting!<p>Would this approach (with non-linear learning) still be able to utilize GPUs to speed up training?
评论 #40219635 未加载
coderenegade大约 1 年前
I was under the impression that graph neural nets already trained learnable functions on graph edges rather than nodes, albeit typically on a fully connected graph. Is there any comparison to just a basic GNN here?
评论 #40233625 未加载
renonce大约 1 年前
So a new type of neural network that has been proven to work well on regression tasks common in physics? And tested in practice to fit well on elementary algebra and compositions of complex functions. But no evidence at all if it works on the most basic machine learning tasks like MNIST, not to mention language models.<p>I mean it&#x27;s great but at the current state it seems better suited for tasks where an explicit formula exists (though not known) and the goal is to predict it on unknown points (and be able to interpret the formula as a side effect). Deep learning tasks are more of a statistical nature (think models with a cross entropy loss - it&#x27;s statistically predicting the frequency of different choices of the class&#x2F;next token), it requires a specialized training procedure and it is designed to fit 100% rather than somewhat close (think linear algebra - it won&#x27;t be good at it). It would very likely take a radically different idea to apply it to deep learning tasks. The recently updated &quot;Author&#x27;s note&quot; also mentions this: &quot;KANs are designed for applications where one cares about high accuracy and&#x2F;or interpretability.&quot;<p>It&#x27;s great but let&#x27;s be patient before we see this improve LLM accuracy or be used elsewhere.
nico大约 1 年前
Looks super interesting<p>I wonder how many more new architectures are going to be found in the next few years
ComplexSystems大约 1 年前
Very interesting! Could existing MLP-style neural networks be put into this form?
nu91大约 1 年前
I am curious to know if this type of network can help with causal inference.
评论 #40238308 未加载
评论 #40268721 未加载
brrrrrm大约 1 年前
doesn&#x27;t KA representation require continuous univariate functions? do B-splines actually cover the space of all continuous functions? wouldn&#x27;t... MLPs be better for the learnable activation functions?
评论 #40247087 未加载
arianvanp大约 1 年前
This really reminds me of petrinets but an analog version? But instead of places and discrete tokens we have activation functions and signals. You can only trigger a transition if an activation function (place) has the right signal (tokens).
keynesyoudigit大约 1 年前
Eli5: why aren&#x27;t these more popular and broadly used?
评论 #40225055 未加载
yza大约 1 年前
Bayesian KANs, KAN Transformers and KAN VAE&#x27;s in 3.2...
WithinReason大约 1 年前
Looks very interesting, but my guess would be that this would run into the problem of exploding&#x2F;vanishing gradients at larger depths, just like TanH or sigmoid networks do.