Sam Altman Steps Down as OpenAI CEO

Joined
Apr 3, 2014
Messages
71,910
Reputation
17,058
Daps
305,987
whats with this logic that if i don't condemn something that it must mean i'm open to doing the same thing myself?

I don't care if someone chooses to use drugs, I do not use drugs nor do I have a desire to do so.

She can do whatever she's comfortable with doing as long as she's not harming others.




Well thats the difference between us. I'm against drug use. I'm not for people destroying their bodies and shortening their lives with drug use.

Your drug use won't affect me personally true, but the loss of my fellow man diminishes me.

I want all my brothas to be well and live well. :wow:
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,057
Reputation
8,239
Daps
157,733
Well thats the difference between us. I'm against drug use. I'm not for people destroying their bodies and shortening their lives with drug use.

Your drug use won't affect me personally true, but the loss of my fellow man diminishes me.

I want all my brothas to be well and live well. :wow:

I'm not FOR that at all but I accept that there will always be a contingent amount of people who will always want to experiment and take risks no matter the amount of warnings. I think the policies created in pursuit of stopping said people are far more damaging than the behavior that it's meant to stop.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,057
Reputation
8,239
Daps
157,733
That conversation tells me that it’s hard to recreate nature

How can you “teach” a system to fully think, discover new things, do people even know how humans think?

Trial and error. Unless you believe that God simply gave humans intelligence, then human brains must have learned to think the same way at some point.


from another tweet

The MLC process using examples:

Imagine a language model named Charles.

Let's say Charles is presented with some novel words and a few examples that demonstrate their meaning:

Study examples:
glerk → RED CIRCLE
blicket → BLUE CIRCLE

Now Charles is given a new word and asked to interpret it systematically based on the examples:

Query:
fep

Charles leverages its compositional skills nurtured by MLC to infer that fep likely maps to a new color, since the examples mapped individual words to colored circles. It responds:

GREEN CIRCLE

In another episode, Charles is given some examples showing words that combine other words:

Study examples:
glerk kiki blicket → BLUE CIRCLE RED CIRCLE
blicket kiki glerk → RED CIRCLE BLUE CIRCLE

Query:
fep kiki glip

Charles recognizes kiki is combining words. It systematically composes the likely meanings of fep and glip from the examples, responding:

PURPLE CIRCLE ORANGE CIRCLE

By training on many such episodes requiring rapid generalization, MLC enables Claude to learn how to learn - to systematically compose meanings from limited examples.

This illustrates how the curriculum of compositional reasoning tasks teaches the model to exhibit human-like systematicity in novel situations, as quantified by its strong performance matching people



snippet:

Human-like systematic generalization through a meta-learning neural network​

Nature volume 623, pages115–121 (2023)Cite this article

Abstract​

The power of human language and thought arises from systematic compositionality—the algebraic ability to understand and produce novel combinations from known components. Fodor and Pylyshyn1 famously argued that artificial neural networks lack this capacity and are therefore not viable models of the mind. Neural networks have advanced considerably in the years since, yet the systematicity challenge persists. Here we successfully address Fodor and Pylyshyn’s challenge by providing evidence that neural networks can achieve human-like systematicity when optimized for their compositional skills. To do so, we introduce the meta-learning for compositionality (MLC) approach for guiding training through a dynamic stream of compositional tasks. To compare humans and machines, we conducted human behavioural experiments using an instruction learning paradigm. After considering seven different models, we found that, in contrast to perfectly systematic but rigid probabilistic symbolic models, and perfectly flexible but unsystematic neural networks, only MLC achieves both the systematicity and flexibility needed for human-like generalization. MLC also advances the compositional skills of machine learning systems in several systematic generalization benchmarks. Our results show how a standard neural network architecture, optimized for its compositional skills, can mimic human systematic generalization in a head-to-head comparison.

Main​

People are adept at learning new concepts and systematically combining them with existing concepts. For example, once a child learns how to ‘skip’, they can understand how to ‘skip backwards’ or ‘skip around a cone twice’ due to their compositional skills. Fodor and Pylyshyn1 argued that neural networks lack this type of systematicity and are therefore not plausible cognitive models, leading to a vigorous debate that spans 35 years2,3,4,5. Counterarguments to Fodor and Pylyshyn1 have focused on two main points. The first is that human compositional skills, although important, may not be as systematic and rule-like as Fodor and Pylyshyn indicated3,6,7. The second is that neural networks, although limited in their most basic forms, can be more systematic when using sophisticated architectures8,9,10. In recent years, neural networks have advanced considerably and led to a number of breakthroughs, including in natural language processing. In light of these advances, we and other researchers have reformulated classic tests of systematicity and reevaluated Fodor and Pylyshyn’s arguments1. Notably, modern neural networks still struggle on tests of systematicity11,12,13,14,15,16,17,18—tests that even a minimally algebraic mind should pass2. As the technology marches on19,20, the systematicity debate continues.
In this Article, we provide evidence that neural networks can achieve human-like systematic generalization through MLC—an optimization procedure that we introduce for encouraging systematicity through a series of few-shot compositional tasks (Fig. 1). Our implementation of MLC uses only common neural networks without added symbolic machinery, and without hand-designed internal representations or inductive biases. Instead, MLC provides a means of specifying the desired behaviour through high-level guidance and/or direct human examples; a neural network is then asked to develop the right learning skills through meta-learning21.


EDIT: asked AI to dumb it(the tweet) down.

Sure, let's break it down:

This is about a big breakthrough in artificial intelligence (AI). For a long time, people thought that AI couldn't think like humans because it's based on statistics and not on understanding things the way we do. But this new research shows that AI can actually learn to think more like us.

The researchers used a new method called MLC to train an AI model. They gave the model lots of small tasks that required it to think in a human-like way. The model did really well, and it even made some of the same mistakes that humans do, which shows that it's not just blindly following rules but actually understanding things.

This is a big deal because it means we can use AI to understand more about how our own brains work. It also means that AI can be more useful because it can understand things in a more human-like way.

The researchers also shared their work so that other people can learn from it and build on it. They published a paper about it and even shared the code they used.

Here's an example of how it works:

Let's say we have an AI model named Charles. We teach Charles that the word "glerk" means "red circle" and "blicket" means "blue circle". Then we give Charles a new word, "fep", and ask it to guess what it means. Charles guesses that "fep" might mean "green circle", because it's learned that these kinds of words usually refer to colored circles.

Then we give Charles some more complex examples. We teach it that "glerk kiki blicket" means "blue circle red circle" and "blicket kiki glerk" means "red circle blue circle". Then we give Charles a new phrase, "fep kiki glip", and ask it to guess what it means. Charles guesses that it might mean "purple circle orange circle", because it's learned that "kiki" is used to combine words.

So, by training Charles on lots of these kinds of tasks, it learns to understand and use words in a human-like way. This is a big step forward in AI research.
 
Last edited:

Save $ not Hoes

Superstar
Supporter
Joined
Aug 3, 2012
Messages
4,158
Reputation
986
Daps
18,796
All I know is watching this new mission impossible and trying that picture version of chat gpt my mind thinking it’s the AI 😂

That picture shyt is the most impressive shyt I’ve seen in years
 

Gritsngravy

Superstar
Joined
Mar 11, 2022
Messages
8,177
Reputation
577
Daps
16,513
This is interesting, makes me want to get into machine learning, need to find some courses
 
Top