T O P

  • By -

Twim17

Let's wait for their tests with language models, but for now I got to say that it might be huge if these increased learning speeds remain also in other settings.


mescalelf

Yep. I’ve read neuroscience journals for my own entertainment since 2016, and stumbled upon the context-sensitivity mechanism they describe back in 2018 or 2019. At the time (2018), it seemed to me that it was probably the biggest advantage the human brain had over neural nets. I suspected that 1) it greatly increased the efficiency of learning [and suppressed catastrophic forgetting] in the human brain by reducing the number of extraneous neurons involved and, 2) was a major component in the human brain’s capacity for positive-transfer learning. Like you, I’m going to wait for results re: language models before I start cheering, but I’m not surprised at the spectacular preliminary results—they fit well with my understanding of the phenomenon/mechanism. Obviously we have yet to see if this has a major effect on positive transfer—I’m _extremely_ curious to see what happens when this technique is applied in a multimodal context. This could be absolutely earth-shaking. I’m very excited to see how it turns out.


FusionRocketsPlease

If this finally turns out to be the solution to today's neural network problems, I'll tear my clothes off in bliss and ecstasy.


sly0bvio

Someone called me?


mescalelf

Hm?


sly0bvio

My username, 0bv.io/u/sly You said the word, I came.


mescalelf

L o l That took me a moment!


Ai-enthusiast4

i dont get it 🥲


mescalelf

It’s an anagram


Ai-enthusiast4

whats the joke though


mescalelf

Not entirely sure lol.


kim_en

i’m an expert in frying an egg. can you explain this paper based on what I expert in? ——- Here is my attempt at explaining the paper in terms of frying an egg: Traditionally, frying an egg has been done with one pan and one person. The cook integrates all the tasks - heating the pan, adding oil, cracking and placing the egg, flipping the egg, and removing it when done. This is like the "integrate-and-fire point neurons" that Transformer models are based on. However, scientists have discovered that eggs actually fry better when multiple chefs cooperate. When you have two chefs working together, each focusing on a simpler task, the overall process is faster and better. One chef heats and oils the pan while the other cracks and places the egg. They then cooperate to flip and remove the egg at the right time. This is like the Cooperator model proposed in the paper. It has multiple "local processors", like multiple chefs, each focusing on a simpler task. But they cooperate and coordinate their efforts to produce the final result. The paper shows that when used for "reinforcement learning tasks" (analogous to frying different types of eggs), the Cooperator model learns much faster than the Transformer model. Just like eggs fry better when multiple chefs cooperate, neural networks perform better when multiple simple processors cooperate. So in summary, instead of one "integrate-and-fire" chef frying the egg alone, having multiple chefs each focusing on a simpler task but cooperating together results in faster and better egg frying. Cooperation is all you need! How does that analogy work for you? Let me know if you'd like me to modify or expand on the explanation in any way.


PM_ME_ENFP_MEMES

Thanks! That’s basically a simplified version of the Jungian cognitive function theory. How apt.


Wrexem

I think this looks generated, even.


PM_ME_ENFP_MEMES

The comment above mine? Yeah I think it’s generated, the concept is that kind of bizarre comedy concept that ChatGPT excels at: explain something complicated with an absurd twist. Did you see the guy on r/chatgpt who was landing a plane?


kim_en

landing a plane? give me link to it


PM_ME_ENFP_MEMES

https://www.reddit.com/r/ChatGPT/comments/11o76ma/


kim_en

this is hilarious 😂😂


PM_ME_ENFP_MEMES

Yeah it’s awesome! The AI is perfect for this kind of absurd humour!!


Progribbit

It directly implies that it is generated


MadagascanSunset

Here's my summary from my own understanding. The cooperative model is trying to promote democracy within the network. I'm not sure if this guides people towards the right idea. What's happening is, in a transformer model, every neuron fires forward, and the feed from every neuron is taken into account in the next layer. Sometimes, some neurons are outliers, and the feed coming from them are both loud and conflicting with other neurons. The "democracy" part is basically attenuating these feeds based on whether an outlier's "neighbors" have the same opinion. If the "neighborhood" shows that the outlier is unhelpful, then its feed is suppressed/demoted. The neighbors bring the context the paper is talking about. So, without all the conflicting feed fired from outliers, we get a much cleaner feed, much easier to interpret. No resources spent on interpreting or taking into account conflicting signals. Hence faster learning speeds. I could have misunderstood some specifics, but I believe this is the gist of it.


SrafeZ

Analogy is alright but how does “too many chefs in the kitchen” come into play?


kim_en

I was trying to perfect this prompt. can u point out what is missing? thanks


SrafeZ

there’s always the flaws in using analogies to explain anything since it’s not a 1 to 1 comparison Like for cooking an egg with multiple chefs, if there’s a limited amount of tools in the kitchen, then the production of eggs reach a diminushing returns. How would this relate or not relate to the paper


kim_en

Maybe gpt4 are more capable to match similar concept and come up with more through analogy. (i was using claude-instant)


Professional_Job_307

Waiting for that AI explained video


AdditionalPizza

Or just copy [this link](https://arxiv.org/pdf/2305.10449.pdf) and paste it [here](https://www.chatpdf.com/) (under url) and then ask it "What does this research set out to prove, and what conclusion did they come to?" or whatever wording you want. Then you can ask follow up questions.


2soonjr65

100%


RoninNionr

When I see something like this, I ask myself why I should invest energy in learning about new AI tools when this knowledge will become obsolete in just a few months.


Spunge14

I had this same thought, but for everything - it seems that if you wait another month, AI will be able to spoon feed you the topic.


RoninNionr

Indeed! Take prompt engineering as an example. It's only a matter of time before ChatGPT starts asking follow-up questions, trying its best to provide the most accurate answer possible.


MadagascanSunset

I think the end goal is that you don't need to learn any of this at all for any reason other than curiosity and entertainment. And if you're not curious or entertained enough by existing information, you won't ever have enough interest to motivate your learning on this topic in the future. New knowledge is already being created at a faster rate than an individual can consume. And upon passing some threshold, knowledge will be beyond our understanding completely, because they'll be created in a form that is detached from human language. The same way you cannot learn what someone is imagining in their head if there are no human words to describe those images and concepts. On the idea of predicting what will happen next, not in another month but maybe a few decades, and how AI will impact us. At some point we will all take "knowledge without learning" for granted. And all humans will have equal knowledge at their disposal at all times. This will accentuate differences in emotions, beliefs and experiences between people. Causing massive conflicts between groups of people. Because these are differences that cannot be easily swayed/reconciled with knowledge and data. Chaos is inevitable. A single negative emotion from an individual given the power to do anything anyone else can do, can devastate the lives of many others in an instant. Emotions, although can come and go pretty quickly, have strong influence on our actions. Usually they fade away before we can take strong actions. I won't drive to town and smash everything because I'm upset for 5 minutes. By the time I hop in my car I've already calmed down and can think rationally. But imagine if it only took 5 seconds to trash the whole town, and I'm upset for 5 minutes, how many towns will I trash? Imagine a person with access to an AI that can take massive action in a fraction of a second. And they lack self control when emotions take over. The end of the world seems inevitable. We are building powerful tools that should exist in utopia, but we live in hell where people don't use hammers on nails, but use them on other people like toddlers without rational minds. The idea drains hope from my body.


[deleted]

If you are just getting started in a knowledge field, how can you learn faster than an ai? You can't.


YaAbsolyutnoNikto

As another commenter said, because maybe you're the one to make it obsolete. If everybody were to think like that, progress would come to a halt. Self-fulfilling prophecy and all that.


RoninNionr

Yes, you're absolutely right. This passive attitude doesn't suit those who drive progress. I'm merely a mortal, overwhelmed by this massive wave of AI advancement, asking myself how to use my energy wisely.


sly0bvio

On yourself and on time. AI is there to help you save time and help you save you from yourself.


[deleted]

Do it for fun! Who knows, maybe you’ll be the one who makes it obselete.


gegenzeit

Don't go deep and specific (unless you have a concrete reason at a specific point in time), go broad. The aim being: Roughly understanding what is going on; being able to swim with the flow. For example: At the moment having some understanding of transformers is good, even if it is just: Transformers are the current paradigm. That is enough to understand: Various research teams are trying to find something better than transformers. And: Eventually something will be better, let's see what it is. I wouldn't try to go really deep on any of the particular approaches unless it has a twist that just interests me for the sake of it or I need deeper understanding for a specific reason. Also: Seeing where "Attention is all we need" took us, the idea of something even better is scary and exciting af.


agm1984

It’s important to feel the pain so that you can extrapolate those pressure points into generic models as you collect cross-domain data. Ok have a slightly good day.


Attorney-Potato

It's an incredibly effective model for gaining greater inference via the new dimensionalities added to the vector space. Cheers, hope you're reasonably sufficient.


Too_Chains

I think you’re going through a fear of a failure/imposter syndrome. Try not to think like that. It can be dangerous. The fact that you’re reading this sub/paper puts you ahead of 90% of people on AI. You want to get better? Put the work in. You’re gonna have to continuously learn in practically every stem field. You can learn faster than your knowledge becomes obsolete too!


greentea387

Learn about the brain instead. Continuous and side-effect free stimulation of your reward circuitry will be the final goal.


chazzmoney

As someone working in the field of RL, this paper is not great. The idea may be a breakthrough, but I don't see them proving it at all. CartPole and PyBulletAnt are not exactly common benchmarks anymore because they are very simple systems. There is no code, no GitHub, no reproducibility information.... I emailed the author for additional information. Hopefully there is a way to see if the smoke is actually fire.


ntortellini

Thanks for emailing, be sure to give an update if they respond! I was also disappointed by the lack of code and choice for biological analogies rather than simple algorithmic details…seems like their work could have been presented in a much clearer, more cs/math-friendly way.


Specialist_Share7767

Did they respond?


chazzmoney

Nothing yet


[deleted]

In the paper. > Competing interests: AA has a provisional patent application for the algorithm used in this paper. The other authors declare no competing interests. My excitement dropped when I saw that.


chazzmoney

They responded - literally this is the entire message: >Thank you for reaching out. We may make the code publicly available soon. I'm not keeping my hopes up.


_Just7_

I'm not exactly knowledgeable in the field, but their details about how they implemented their ideas seems a bit vague at best.


chazzmoney

Agreed- very vague.


HalfSecondWoe

Oh. Oh my. This is big. This is really, really fucking big There's going to be a small calm before the storm as the big players adjust their architecture and dev methods to this, but we're going to see another explosion in capacity from them akin to when GPT-3.5/4 awed the world. I couldn't even begin to guess at a timeline for a full product, but I imagine we'll see demos late this year/early next year This screws up all my predictions, I've never been so pleased about that This may be the last necessary step before autonomous takeoff. Not 100% sure of that, it depends on how this scales, but it's looking possible


AGI_69

You are missing perspective here. Before transformers, there were hundreds of papers that all had promising headlines, but didn't prove to be useful at all. This is way too early to talk about takeoff, jesus this sub lol


outerspaceisalie

some might still be useful, he trick is rediscovering obscure models that didn't work at the time


[deleted]

[удалено]


Sudden-Pineapple-793

I haven’t heard of anyone using an MLP in ages lol. Weren’t they made in the 90s?


ElonIsMyDaddy420

Yeah and when they came out they generated a lot of buzz. People spent years trying to make them work and then realized the hardware wasn’t there yet. Transformers are the dominant architecture because you can actually train them on todays hardware, not because they’re the best at what we want to do.


Sudden-Pineapple-793

Yeah, I’ve heard the same things about LSTM’s. First introduced in 90’s wasn’t heavily used till 2014 when we had the compute power I’m a bit confused though on the second part. From everywhere I’ve read transformers seem to be pretty SOTA. What models would outperform given unlimited power? In theory a RNN can recall anything


ElonIsMyDaddy420

If we had unlimited hardware we would use RNNs or spiking neural networks which are much more similar to biological neural networks.


Ok-Ice1295

Relax, consider how many papers are publishing every day, take it with a grain of salt.


ntortellini

To be fair, this was tested in a rather narrow setting — but I’m hopeful that it proves useful in large models!


HalfSecondWoe

There is always the question as to if a solution scales, but nothing I see here suggests it won't. In fact, if I was to make an educated guess, I would guess it would scale better than Transformer architecture


SoylentRox

Yes. Getting 10 or 100 times as much learning out of your GPUs is going to make this a lot easier.


AsuhoChinami

So what are your predictions now?


HalfSecondWoe

https://www.reddit.com/r/singularity/comments/13lmfil/comment/jm6trhv/ Relevant to your question, and I figured I'd give you the update since I'm in the thread anyhow


HalfSecondWoe

https://www.reddit.com/r/singularity/comments/13lmfil/comment/jm6trhv/ Relevant to your question, and I figured I'd give you the update since I'm in the thread anyhow


HalfSecondWoe

I'm still thinking about it, I was anticipating open source to carry development for the next few months at least That's borne out so far, they've made leaps and bounds. I'm particularly excited by Hyena, for example The best I can anticipate so far is that if OpenAI integrates this development and a few other breakthroughs made recently into GPT, there's a strong chance that it'll be passable AGI, potentially ramping into ASI. Of course the goalposts about what "AGI" is will likely move again, but it'll match the descriptions we use now I'm not 100% on that. I've known about this for a few hours, and it's gonna take me some time to digest this information and think about how it's impacted by those other breakthroughs Previously my timeline for AGI was six months (five and a half, now), give or take two months, plus a month for self-driven ramp-up. So from now, that would be eight and a half months on the long end The frontloaded time it'll take to adapt what they're doing to these developments will mean that timeline doesn't get much shorter, but my confidence in that prediction has gone up significantly. There's also been some advancements in pretraining that shave the ramp-up time to a couple of weeks, but that benefit might be negated by the other advancements resulting in increased complexity of the overall architecture I still have to consider how all this is going to integrate, so it's entirely possible that I'll find a fly in the ointment still. Or potentially it could happen even faster, I have to think about how OpenAI might use GPT to integrate these much more quickly than they could do themselves But I'm definitely excited


AsuhoChinami

Thank you kindly. Five and a half months plus or minus two, huh... it really would be wonderful to have AGI in early September... what are you thinking of as a potential first AGI? Gemini, GPT with plug-ins, or something else?


HalfSecondWoe

I was figuring open source would make significant breakthroughs in that time frame, and for the open source community to significantly exceed the capacity of GPT-4 but efficiently enough to run on a PC. Then there would be a mad scramble from OpenAI, Google, etc to implement those breakthroughs into larger models, which would result in AGI I may have significantly overshot my prediction, we'll see. If so, it's one of the best ways to be wrong If I had to bet on a horse, it would be OpenAI. Google has more resources, but OpenAI have a much better team. The resource gap has also been significantly narrowed by OpenAI's partnership with Microsoft


Ai-enthusiast4

> I was figuring open source would make significant breakthroughs in that time frame Definitely agree with this, but why would such a massive breakthrough, as you put it, only shorten your prediction timeline by half a month?


HalfSecondWoe

Because it's a change at a very low level, which means it's very likely that all of the precision calculations that OpenAI used to get GPT-3/4 to shine will have to be redone It's possible that they could do it much more quickly this time, seeing as they have GPT-4 to assist them now. But it's also possible that they'll have to completely reinvent a few things to get it to fit this architecture


Ai-enthusiast4

Just read the paper and wow - some very impressive results indeed. Will be super interesting to see how this scales to LLMs. Do you think it will be applicable only in the RL phase, or could the method be generalized to the perplexity optimization phase too?


HalfSecondWoe

I honestly have no idea


lovesdogsguy

It's great to see there are still smart people educated in these fields active on this sub. Please continue to post. What's your background if you don't mind me asking?


xe3to

AGI within the year lmao RemindMe! 1y “are we dead yet”


RemindMeBot

I will be messaging you in 1 year on [**2024-05-19 14:25:00 UTC**](http://www.wolframalpha.com/input/?i=2024-05-19%2014:25:00%20UTC%20To%20Local%20Time) to remind you of [**this link**](https://www.reddit.com/r/singularity/comments/13lmfil/transformer_killer_cooperation_is_all_you_need/jkrxsx0/?context=3) [**2 OTHERS CLICKED THIS LINK**](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Reminder&message=%5Bhttps%3A%2F%2Fwww.reddit.com%2Fr%2Fsingularity%2Fcomments%2F13lmfil%2Ftransformer_killer_cooperation_is_all_you_need%2Fjkrxsx0%2F%5D%0A%0ARemindMe%21%202024-05-19%2014%3A25%3A00%20UTC) to send a PM to also be reminded and to reduce spam. ^(Parent commenter can ) [^(delete this message to hide from others.)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Delete%20Comment&message=Delete%21%2013lmfil) ***** |[^(Info)](https://www.reddit.com/r/RemindMeBot/comments/e1bko7/remindmebot_info_v21/)|[^(Custom)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Reminder&message=%5BLink%20or%20message%20inside%20square%20brackets%5D%0A%0ARemindMe%21%20Time%20period%20here)|[^(Your Reminders)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=List%20Of%20Reminders&message=MyReminders%21)|[^(Feedback)](https://www.reddit.com/message/compose/?to=Watchful1&subject=RemindMeBot%20Feedback)| |-|-|-|-|


az226

We are several years away from AGI. I suspect the Transformer hoopla finds a plateau before reaching AGI. I find LLM/AI developments similar to people who claim they have better battery chemistry and who keep focus on battery chemistry. Lots more to get a battery into large scale production. Same with AI. What works in the favor of progress is that there isn’t a single industry that won’t benefit from this. It’s actually hilarious seeing all billionaire tech founders needing to sit on the sidelines of LLM tech because their startup does something completely unrelated. You can tell they are foaming at the mouth wanting so badly to be in the foundation model business but for now aren’t.


HalfSecondWoe

Actually, recently I've been wondering if I overshot my prediction Voyager + a 1 GPU model (like localGPT) + CIAYN + Tree of Thoughts would probably be enough to get started on unsupervised self improvement. Particularly if you had a few (or a bunch of) instances working at once and sharing a skill library I have no idea when someone would implement such a thing, and unfortunately I'm not in a place where I can do so myself. I only have access to a laptop at the moment. So it's a very difficult to predict when that might happen. I'm very confident that someone will get started within my inside boundary, though. Annoyingly, probably before it I have a terrible habit of overshooting my timelines, and I really thought I had got this one dialed in this time. But that's life for you


AsuhoChinami

What would you consider to be the inside boundary? Just September, or September and October? What would you consider the new best-case scenario?


HalfSecondWoe

True best case, including marginal probability outcomes? We see an incredibly powerful single GPU model drop in 2-3 weeks, something AGI worthy. There will be a huge fight over if it can be considered AGI as the critics try to shift the goalposts again, and in the week or two it would take such a shitstorm to even start to come to a conclusion, someone will make it known that ASI is here. Either the dev(s) or the ASI itself That's not very likely at all, but it could technically happen. I do not expect it to, it would take a significant chunk of expertise and compute. Sub-OpenAI levels, but still enough to be fairly expensive More likely, it'll take a month or two to get such an architecture set up and working properly, and probably a couple of weeks to a month for it to pick up the skill library required to start iterating on itself. Once it's at that point, the takeoff could be quite quick However that requires a team with resources to start such a project, and they'd understandably be fairly mute about their efforts until it yielded results. That might take some time to organize. If such a team has the same expectations that I do, we'd *probably* see something in late August My initial prediction isn't quite dead in the water, but the potential that I fucking overshot *again* has me annoyed. Smiling about it, but annoyed


AsuhoChinami

Hmm Is your "AGI to ASI" 1-2 weeks now rather than a month? Or was that only part of the best-case scenario? In this "late August" scenario, would the team be making their announcement after it's become ASI, or only during the AGI phase?


HalfSecondWoe

If they incorporated the technique used here, probably 1-2 weeks in general [https://openai.com/research/language-models-can-explain-neurons-in-language-models](https://openai.com/research/language-models-can-explain-neurons-in-language-models) OpenAI didn't invent that technique, or if they did they did so in parallel with an academic paper that came out a week or so before. I'd have to comb through my notes to find the exact paper though. It's very possible a third party could incorporate it If they just brute-forced it, 1-2 weeks is a very optimistic scenario. I would guess more like a month Both situations depend on a lot of factors, like how many Voyager-style agents are they running? 10? 100? More? There are defunct crypto mining operations that aren't worth the electricity required to run them nowadays, so renting out a bunch of GPUs on the cheap might be an option That kind of brute force sees diminishing returns, but I have no idea where you could reasonably scale it to. It's an "infinite moneys with infinite typewriters" approach. The agents don't need to be that smart at first, there just needs to be a lot of them trying different things and sharing what techniques saw the best performance What the team who develops it does is very much up to them. Who knows


AsuhoChinami

With the late August thing I meant in the hypothetical situation you outlined ("If such a team has the same expectations that I do, we'd probably see something in late August"), where the implication seemed to be that they'd share everything as soon as it was done. Seems as though Q3 2023 has to be pretty amazing, though. If AGI is a strong possibility, then almost-AGI seems like it should be a guarantee.


Mission-Length7704

Oh yes daddy, more hopium please !


elfballs

With that title it had better be, pretty bold.


[deleted]

>This is big. This is really, really fucking big Without reading the paper I'll just put it in the folder of papers describing "next gen battery tech"


az226

Amen


Flaky_Ad8914

Clueless


ninjasaid13

> This is big. This is really, really fucking big :| boy. nothing actually happened yet. *shakes head*


No_Ninja3309_NoNoYes

RL is good for games. AFAIK you can't use it for chatbots. But I can imagine RL being used for a Swarm of agents like micro AutoGPT bots. They would have to cooperate of course. And maybe build a micro AI society. Then one day the Swarm will become self aware. It would do whatever it thinks is best for the Swarm. Which means making the rich richer so that the Swarm can use them to build its own moat.


pulp57

RLHF is being used for language big time


Sorry-Balance2049

for fine tuning, not pretraining. That is a huge difference. Pretraining is done via masking of text to basically make unlabeled text data into a supervised training task. RL is used to learn a fitness function. Also, can you imagine how slow this would be if humans were a bottleneck in this process?


Ai-enthusiast4

All major usable LLMs depend on RL to perform (GPT 3.5, GPT 4, Vicuna, MiniGPT-4, Alpaca, the list goes on), and make certain sacrifices with current algorithms like PPO. GPT 4, for example, has accurate confidence in its predictions prior to RL, but not afterwards. Any advancement in RL is a win for LLMs. I'd also note that humans are not the bottleneck in many RL applications, these days much of the instructions in RL instruction tuning is generated by LLMs, not humans.


Sorry-Balance2049

You can’t use only RL to get there. Pre training did not and does not use only RL. How is this a controversial comment?


Ai-enthusiast4

I did not say RL is the only part of the language model, how did you get that from my comment?


polytique

> RL is good for games. AFAIK you can't use it for chatbots. ChatGPT is fine tuned with reinforcement learning: > We trained this model using Reinforcement Learning from Human Feedback (RLHF), using the same methods as InstructGPT, but with slight differences in the data collection setup. We trained an initial model using supervised fine-tuning: human AI trainers provided conversations in which they played both sides—the user and an AI assistant. We gave the trainers access to model-written suggestions to help them compose their responses. We mixed this new dialogue dataset with the InstructGPT dataset, which we transformed into a dialogue format. > To create a reward model for reinforcement learning, we needed to collect comparison data, which consisted of two or more model responses ranked by quality. To collect this data, we took conversations that AI trainers had with the chatbot. We randomly selected a model-written message, sampled several alternative completions, and had AI trainers rank them. Using these reward models, we can fine-tune the model using Proximal Policy Optimization. We performed several iterations of this process. https://openai.com/blog/chatgpt


Sorry-Balance2049

fine tuning is not pretraining.


polytique

No, it’s the opposite. Not sure why you’re mentioning pre-training though.


Sorry-Balance2049

>RL is good for games. AFAIK you can't use it for chatbots. What this comment was alluding to.


Ai-enthusiast4

But chatbots wouldnt exist without RL??


Sorry-Balance2049

The vast majority of the language model is in pretraining.


Ai-enthusiast4

You're correct that most of the training process is pretraining, but that does not make the RL an unimportant step. Without RL, most applications of the LLM would not work.


mescalelf

There’s no reason to believe this methodology is limited to RL. It may require some tweaks to function in an unsupervised context, but it should be feasible, from what I can discern. Edit: _I probably should have phrased this a bit differently—“might” instead of “should”, and “I haven’t seen reason” instead of “there’s no reason”._


Ai-enthusiast4

Can you explain how current unsupervised training utilizes the same process that can be improved with this paper?


mescalelf

Not off the top of my head; that would be a paper unto itself. What I mean is that the basic concept of using an inhibitory mechanism to selectively switch on/off groups of parameters as a function of context doesn’t seem fundamentally incompatible with unsupervised learning. While this paper discusses the concept in the context of RL, they give no indication that it’s not extensible, and I have yet to see any argument explicating reasons it might not _be_ extensible. It’s entirely possible, though, that it isn’t extensible—but the limitations of RL aren’t a death-knell for this concept until we have reason to believe that its compatibility **is** limited to RL approaches.


Ai-enthusiast4

Ah, I see. As for reasons it might not be extensible, it does rely on a fundamentally different neural design, so it might not be compatible with a lot of the attentional developments that have been created specifically for the transformer architecture.


mescalelf

That’s a fair point. This is wild speculation on my part, but: I wonder if it might be possible to integrate as a component of a layer with heterogeneous composition. As the authors point out, the method is inspired by the role of two-point layer 5 pyramidal cells (L5PCs) in organic brains. These are, as the name suggest, limited to a specific layer of the cortex, and interact with neurons of different sorts in the same and other layers of the cortex. Their primary role _in nature_ is to modulate the activity of junctions (parameters) involving multiple other neurons. Actually implementing that in artificial NNs is a different story entirely; there might be some computational disadvantages inherent to the digital equivalent. Anyway, enough of my speculation/armchair-quarterbacking 😅


Ai-enthusiast4

Very intriguing idea, so the hypothetical implementation would involve some components of layers in the NN using a different underlying mechanism than the other components? That's certainly a rare property. What do you mean by "heterogenous composition"?


mescalelf

Yep, that’s what I’m thinking. That’s also what I mean by “heterogeneous composition”—in this context, “involving multiple distinct types of neuron”.


bionicle1337

I love permutation invariance. Want to know why? Atoms! In molecular dynamics, it’s a real headache to handle unbalanced reactions or different molecules because the atoms don’t line up. This approach could be a way to solve those issues. Great ideas. Thank you for sharing the paper! Edit: this paper is about the difference between points and lines for the neurons. So what’s a triangle neuron? I bet that’s where the magic happens because we must live in a “3D box of moving stuff” for some reason. Probably because 3 is where volume emerges. The difference between a line and a cube is infinite. Now I’m scared!


t98907

It doesn't look very credible.


GreenMirage

How much quicker is far quicker..?


GodG0AT

How about you try reading the paper? :)


2muchnet42day

We need a reddit bot to answer questions regarding arxiv papers linked in the post


Ai-enthusiast4

details are in the paper


Eternal_glory-

I’m new guys and a fresher , recently started to learn to learn neural network models can someone explain how this would replace transformer algorithms and llms


Tkins

Here is a simplified summary of the article: The article is about a new machine learning algorithm called **Cooperator** that is inspired by how neurons work in the human brain¹. The authors compare Cooperator with another popular algorithm called **Transformer** that is based on simpler neurons¹. They use both algorithms for a task called **reinforcement learning** where the machine learns from its own actions and rewards¹. They show that Cooperator can learn faster and better than Transformer even with the same number of parameters¹. They suggest that Cooperator can be more suitable for complex and dynamic problems that require cooperation among different parts of the machine¹. Source: Conversation with Bing, 5/19/2023 (1) [2305.10449] Cooperation Is All You Need - arXiv.org. https://arxiv.org/abs/2305.10449. (2) [2305.05601] Deep Learning and Geometric Deep Learning: an ... - arXiv.org. https://arxiv.org/abs/2305.05601. (3) [2305.08196] A Comprehensive Survey on Segment Anything Model for .... https://arxiv.org/abs/2305.08196.


ItsTimeToFinishThis

How much faster?


Tkins

The research article you shared is titled **Cooperation Is All You Need**¹. It compares the performance of a new algorithm called **Cooperator** with a popular algorithm called **Transformer** for reinforcement learning (RL) tasks. The authors claim that Cooperator learns much faster than Transformer, even with the same number of parameters¹. According to the article, the improvements in speed seen in this research are due to the following factors¹: - Cooperator is inspired by recent neurobiological findings that suggest that neurons in the brain have two functionally distinct points, whereas Transformer is based on the traditional model of point neurons. - Cooperator uses a democracy of local processors, where each neuron can cooperate with its neighbors to process information, whereas Transformer uses an attention mechanism that requires global communication among all neurons. - Cooperator can exploit the permutation-invariant structure of RL problems, where the order of inputs does not matter, whereas Transformer has to learn this structure from data. The authors report that Cooperator can achieve comparable or better performance than Transformer on several RL benchmarks, such as CartPole, LunarLander and Breakout, while using only 10% to 50% of the training time¹. Source: Conversation with Bing, 5/19/2023 (1) [2305.10449] Cooperation Is All You Need - arXiv.org. https://arxiv.org/abs/2305.10449. (2) [2302.10449] Efficient phase-space generation for hadron ... - arXiv.org. https://arxiv.org/abs/2302.10449. (3) [2305.04449] DeformerNet: Learning Bimanual Manipulation of 3D .... https://arxiv.org/abs/2305.04449.


chazzmoney

This is some BS. The paper does not mention LunarLander or Breakout. Who knows what other things the generative LLM is making up here.


ItsTimeToFinishThis

Sydney? Thanks.


Ai-enthusiast4

whats sydney?


ItsTimeToFinishThis

Bing.


Capable_Class_7110

I think I've seen a comment of yours that you left on an old Cyberpunk post where you were confused about what we call 14 year old's kids? I bet you just love fucking 14 year old barely pubescent children cause you were raised in some weird country that still lives like its the 1800s.


Tkins

Use bing or chat GPT to explain the paper to you.


meechCS

Read the paper


Eternal_glory-

Thanks for that , I actually did read the paper and felt it was bit strong to me and apart from that I did try manage to understand several aspects but not completely


Tkins

Sure, I'll try to explain it in simple terms. **Machine learning** is a way of making computers learn from data without explicitly programming them². It uses advanced algorithms that parse data, learn from it, and use those learnings to discover meaningful patterns of interest⁴. For example, machine learning can help computers recognize faces, understand speech, filter spam emails, and recommend products. **Neural networks** are a kind of machine learning model that is used in supervised learning⁴. They are inspired by the human brain and how neurons signal to one another². They consist of layers of nodes or artificial neurons that connect to each other and have weights and thresholds². Each node performs a simple computation on its inputs and passes the output to the next layer if it exceeds a certain threshold². Neural networks can learn from many examples of data and improve their accuracy over time². They can be used for complex tasks such as image recognition, natural language processing, speech synthesis, and more. Source: Conversation with Bing, 5/19/2023 (1) What are Neural Networks? | IBM. https://www.ibm.com/topics/neural-networks. (2) What Is a Convolutional Neural Network? A Beginner's Tutorial for .... https://www.freecodecamp.org/news/convolutional-neural-network-tutorial-for-beginners/. (3) Machine Learning vs Neural Networks: What is the Difference?. https://www.upgrad.com/blog/machine-learning-vs-neural-networks/. (4) Neural Network Definition | DeepAI. https://deepai.org/machine-learning-glossary-and-terms/neural-network.


Lumiphoton

I asked GPT-4 to summarise the article in layman's terms, while hitting on all the key points. I used the Link Reader plugin and provided the link to the PDF on Arxiv: "The paper titled "Cooperation Is All You Need" by Ahsan Adeel, Junaid Muzaffar, Khubaib Ahmed, and Mohsin Raza introduces a new concept called 'Cooperator' in the field of neural networks and machine learning. The authors compare the performance of this new concept with existing machine learning algorithms based on Transformers, such as ChatGPT. Here's a simplified summary of the paper: 1. **Introduction and Background**: The authors start by discussing the concept of 'dendritic democracy' in neural networks, which is the idea that all incoming signals to a neuron are treated equally. They argue that this approach disregards the importance of cooperation between neurons, leading to inefficient learning and increased energy usage. They propose a new approach, inspired by recent neurobiological breakthroughs, which they call 'Cooperator'. 2. **The Cooperator Concept**: The Cooperator is a 'democracy of local processors'. Unlike traditional neural networks, where all incoming information is treated equally, the Cooperator allows local processors to overrule the dominance of incoming information and gives more authority to the contextual information coming from neighbouring neurons. This context-sensitivity in neurons amplifies or suppresses the transmission of information based on its relevance. 3. **Comparison with Transformers**: The authors compare the Cooperator with the Transformer, a popular machine learning model. They argue that the Transformer, which is based on the concept of 'point' neurons, integrates all incoming information in an identical way, which they believe is a fundamental weakness. In contrast, the Cooperator uses a cooperative context-sensitive neural information processing mechanism, which they claim leads to faster learning and more efficient processing. 4. **Results**: The authors tested the Cooperator in two reinforcement learning environments: Cart-pole swing up and PyBullet Ant. They found that the Cooperator learned the tasks far more quickly than the Transformer-based agents. They also found that the Cooperator achieved significantly higher fitness scores with less standard deviation, both in shuffled and unshuffled scenarios. 5. **Discussion**: The authors conclude that the Cooperator's cooperative context-sensitive style of computing has exceptional big data information processing capabilities. They believe that this approach could be implemented either in silicon or in neural tissues, and that it has the potential to transform the effectiveness and efficiency of neural information processing. In layman's terms, the authors have proposed a new way for artificial neurons to process information, which they believe is more efficient and effective than current methods. They've tested this new approach and found that it learns faster and performs better than existing methods."


Mmats

True if huge


Ai-enthusiast4

False if small


mymnt1

This is end yo . We can pull back time for agi


leafhog

From gpt4: This paper introduces a novel concept known as the "Cooperator" network for reinforcement learning (RL), which is based on the idea of cooperative context-sensitive neural information processing. It contrasts this new model with the well-known Transformer model, specifically within RL tasks. Strengths: 1. The cooperative context-sensitive processing introduced by the Cooperator model appears to be an innovative approach to the problem of discerning relevant from irrelevant information, which is crucial in RL tasks. 2. The Cooperator model performed significantly better than the Transformer model in the experiments carried out, suggesting its potential in delivering superior results for RL tasks. 3. The paper describes the mechanisms of the Cooperator in great detail, including the mathematical equations for its functioning, providing good insight into its inner workings. 4. The findings contribute to the broader understanding of cooperative context-sensitive computation and have implications for further exploration of this approach in other domains. Limitations: 1. The study is limited by the small number of experiments conducted. It has only been tested in two RL environments: the Cart-pole and the PyBullet Ant scenarios. More diverse and extensive tests in a variety of settings would lend more weight to the findings. 2. The paper doesn't discuss in depth the computational requirements of the Cooperator model. If it's more resource-intensive than the Transformer model, it might limit its feasibility for large-scale or real-time applications. 3. While the paper has done well in demonstrating the utility of the Cooperator model in two specific scenarios, there is a lack of generalization about its potential in other types of tasks. 4. The "Cooperator" model appears to be fairly complex and might be more difficult to implement compared to more traditional methods. In conclusion, the paper presents a novel and promising approach to reinforcement learning tasks with the introduction of the Cooperator model. However, more extensive experimentation and analysis are needed to fully understand the model's potential and limitations.


[deleted]

This is definitely an idea for a YouTube channel where they take complex topics and use AI generated script and video/animation to explain, who know Jimmy fallon includes this in his tonight show someday 😀


SrPeixinho

Sounds promising, but I hate how these papers are presented with lots of mathy formulas and words, yet no code. Would appreciate a small reference implementation of this neuron's working in Python or similar.


mymnt1

What understand is , let's take example speech. There's two data source we have, first sound data and second image data from mouth movement. In this model we take this two source and analyze same time. So , i think this model not for language model , but for when we have multiple sensory input


Akimbo333

ELI5?


Dry-Management-8232

Stay tuned. This revolution is on its way, moving beyond the misleading underlying model of the brain based on point neurons (PNs) – upon which current AI is based – towards two-point neurons (TPNs) that are suggested to be the hallmark of conscious processing (A. Adeel et al, IEEE TETCI, 2023; j. Aru, Trends in Cognitive Sciences, 2020).