Ask HN: Which lately analysis paper blow your thoughts?

![]() |
|
That’s pretty interesting actually. Someone must have a copy somewhere. Seems like a real failure of scholarship if it’s truly lost, and a serious argument against walled gardens-style publishing.
|
![]() |
|
I don’t know why parent comment is stirring up drama but:
1. Not available online doesn’t mean the paper’s existence is made up. It’s a very bold claim to make for the authors that they cite work that is fabricated. From the available information, this looks like a technical report by a, probably now defunct, company back in the 80s. If this was its only form of publication, and not on some conference proceedings for example, it would be only found available on select university libraries as a physical copy. But most important, 2. This isn’t even as an impactful paper as the parent comment states. Or if its proposed concept is, the original idea is probably derived from some other paper that is indeed the one that is highly cited and most definitely available online. Accumulative citations number from Google Scholar and IEEEXplore doesn’t exceed fifteen for the particular paper though. https://scholar.google.com/scholar?cites=1491448744595502026… |
![]() |
|
Integral Neural Networks (CVPR 2023 Award Candidate), a nifty way of building resizable networks.
My understanding of this work: A forward pass for a (fully-connected) layer of a neural network is just a dot product of the layer input with the layer weights, followed by some activation function. Both the input and the weights are vectors of the same, fixed size. Let’s imagine that the discrete values that form these vectors happen to be samples of two different continuous univariate functions. Then we can view the dot product as an approximation to the value of integrating the multiplication of the two continuous functions. Now instead of storing the weights of our network, we store some values from which we can reconstruct a continuous function, and then sample it where we want (in this case some trainable interpolation nodes, which are convoluted with a cubic kernel). This gives us the option to sample different-sized networks, but they are all performing (an approximation to) the same operation. After training with samples at different resolutions, you can freely pick your network size at inference time. You can also take pretrained networks, reorder the weights to make the functions as smooth as possible, and then compress the network, by downsampling. In their experiments, the networks lose much less accuracy when being downsampled, compared to common pruning approaches. Paper: https://openaccess.thecvf.com/content/CVPR2023/papers/Solods… |
![]() |
|
I thought the AlphaZero paper was pretty cool: https://arxiv.org/abs/1712.01815
Not solely did we get an entire new kind of Chess engine, it was additionally fascinating to see how the engine considered completely different openings at varied phases in its coaching. For example, the Caro-Kann, which is my weapon of selection, was favored fairly closely by it for a number of hours after which seemingly rejected (maybe it even refuted it?!) close to the top. |
![]() |
|
> First time for ML that is not deep learning
What do you mean by this? Virtually all “classic” or “shallow” ML can be GPU-accelerated, from linear regression to SVM to GBM. |
![]() |
|
I recently read “Enabling tabular deep learning when d ≫ n with an auxiliary knowledge graph” (https://arxiv.org/pdf/2306.04766.pdf) for certainly one of my graduate courses. Basically, when there are considerably extra knowledge factors than options (n >> d), machine studying normally works superb (assuming knowledge high quality, an underlying relationship, and so forth.). However, for sparse datasets the place there are fewer knowledge factors than options (d >> n), most machine studying strategies fail. There’s simply not sufficient knowledge to study all of the relationships. This paper builds a information graph based mostly on relationships and different pre-existing information of knowledge options to enhance mannequin efficiency on this case. It is actually fascinating – I hadn’t realized there have been methods to get higher efficiency on this case.
|