zlacker

[parent] [thread] 89 comments
1. noneth+(OP)[view] [source] 2023-05-16 19:32:58
This is the definition of regulatory capture. Altman should be invited to speak so that we understand the ideas in his head but anything he suggests should be categorically rejected because he’s just not in a position to be trusted. If what he suggests are good ideas then hopefully we can arrive at them in some other way with a clean chain of custody.

Although I assume if he’s speaking on AI they actually intend on considering his thoughts more seriously than I suggest.

replies(3): >>pg_123+X8 >>EGreg+Ic >>brooks+Rd
2. pg_123+X8[view] [source] 2023-05-16 20:16:15
>>noneth+(OP)
There is also growing speculation that the current level of AI may have peaked in a bang for buck sense.

If this is so, and given the concrete examples of cheap derived models learning from the first movers and rapidly (and did I mention cheaply) closing the gap to this peak, the optimal self-serving corporate play is to invite regulation.

After the legislative moats go up, it is once again about who has the biggest legal team ...

replies(4): >>robwwi+Rc >>TheDud+Ph >>yarg+BB >>daniel+bz1
3. EGreg+Ic[view] [source] 2023-05-16 20:36:01
>>noneth+(OP)
I remember when a different Sam — Mr. Bankman Fried came to testify and ask a different government agency CFTC to oversee cryptocurrency, and put regulations and licenses in place.

AI is following the path of Web3

replies(2): >>mschus+yp >>smcin+Kx
◧◩
4. robwwi+Rc[view] [source] [discussion] 2023-05-16 20:36:34
>>pg_123+X8
Counterpoint—-there is growing speculation we are just about to transition to AGI.
replies(2): >>Eamonn+bd >>causal+mg
◧◩◪
5. Eamonn+bd[view] [source] [discussion] 2023-05-16 20:38:23
>>robwwi+Rc
Growing? Or have the same voices who have been saying it since the aughts suddenly been platformed.
replies(3): >>TeMPOr+so >>jack_p+bs >>lostms+KK
6. brooks+Rd[view] [source] 2023-05-16 20:41:53
>>noneth+(OP)
I'm not following this "good ideas must come from an ideologically pure source" thing.

Shouldn't we be evaluating ideas on the merits and not categorically rejecting (or endorsing) them based on who said them?

replies(6): >>briant+8h >>samsta+wh >>parent+3W >>dreamc+P01 >>bcrosb+D51 >>dTal+JF1
◧◩◪
7. causal+mg[view] [source] [discussion] 2023-05-16 20:53:00
>>robwwi+Rc
Growing among who? The more I learn about and use LLMs the more convinced I am we're in a local maxima and the only way they're going to improve is by getting smaller and cheaper to run. They're still terrible at logical reasoning.

We're going to get some super cool and some super dystopian stuff out of them but LLMs are never going to go into a recursive loop of self-improvement and become machine gods.

replies(6): >>ben_w+Xj >>TeMPOr+2n >>behnam+oo >>bernie+WC >>stuckk+YD >>ux-app+nF
◧◩
8. briant+8h[view] [source] [discussion] 2023-05-16 20:58:43
>>brooks+Rd
> Shouldn't we be evaluating ideas on the merits and not categorically rejecting (or endorsing) them based on who said them?

The problem is when only the entrenched industry players & legislators have a voice, there are many ideas & perspectives that are simply not heard or considered. Industrial groups have a long history of using regulations to entrench their positions & to stifle competition...creating a "barrier to entry" as they say. Going beyond that, industrial groups have shaped public perception & the regulatory apparatus to effectively create a company store, where the only solutions to some problem effectively (or sometimes legally) must go through a small set of large companies.

This concern is especially prescient now, as these technologies are unprecedentedly disruptive to many industries & private life. Using worst case scenario fear mongering as a justification to regulate the extreme majority of usage that will not come close to these fears, is disingenuous & almost always an overreach of governance.

replies(2): >>samsta+Yh >>chii+pP
◧◩
9. samsta+wh[view] [source] [discussion] 2023-05-16 21:00:11
>>brooks+Rd
Aside from who is saying them, the premise holds water.

AI is beyond-borders, and thus unenforceable in practicality.

The top-minds-of-AI are a group that cannot be regulated.

-

AI isnt about the industries it shall disrupt ; AI is the policy-makers it will expose.

THAT is what they are afraid of.

--

I have been able to do financial lenses into organizations that even with rudimentary BI would have taken me months/weeks - but I have been able to find insights which took me minutes.

AI regulation right now, in this infancy, is about damage control.

---

Its the same as the legal weed market. You think BAIN Capital just all of a sudden decided to jump into the market without setting up their spigot?

Do you think that haliburton under cheney was able to setup their supply chains without cheney as head of KBR/Hali/CIA/etc...

Yeah, this is the same play ; AI is going to be squashed until they can use it to profit over you.

Have you watched ANIME ever? Yeah... its here now.

replies(2): >>turtle+lL >>throwa+uQ
◧◩
10. TheDud+Ph[view] [source] [discussion] 2023-05-16 21:02:36
>>pg_123+X8
Why? Because there hasn't been any new developments last week? Oh wait, there has.
◧◩◪
11. samsta+Yh[view] [source] [discussion] 2023-05-16 21:03:24
>>briant+8h
I can only say +1 = and I know how much HN hates that, but ^This.
◧◩◪◨
12. ben_w+Xj[view] [source] [discussion] 2023-05-16 21:16:08
>>causal+mg
> They're still terrible at logical reasoning.

Are they even trying to be good at that? Serious question; using LLMs as a logical processor are as wasteful and as well-suited as using the Great Pyramid of Giza as an AirBnB.

I've not tried this, but I suspect the best way is more like asking the LLM to write a COQ script for the scenario, instead of trying to get it to solve the logic directly.

replies(2): >>staunt+pr >>fsckbo+GD
◧◩◪◨
13. TeMPOr+2n[view] [source] [discussion] 2023-05-16 21:30:39
>>causal+mg
> The more I learn about and use LLMs the more convinced I am we're in a local maxima

Not sure why would you believe that.

Inside view: qualitative improvements LLMs made at scale took everyone by surprise; I don't think anyone understands them enough to make a convincing argument that LLMs have exhausted their potential.

Outside view: what local maximum? Wake me up when someone else makes a LLM comparable in performance to GPT-4. Right now, there is no local maximum. There's one model far ahead of the rest, and that model is actually below it's peak performance - side effect of OpenAI lobotomizing it with aggressive RLHF. The only thing remotely suggesting we shouldn't expect further improvements is... OpenAI saying they kinda want to try some other things, and (pinky swear!) aren't training GPT-4's successor.

> and the only way they're going to improve is by getting smaller and cheaper to run.

Meaning they'll be easier to chain. The next big leap could in fact be a bunch of compressed, power-efficient LLMs talking to each other. Possibly even managing their own deployment.

> They're still terrible at logical reasoning.

So is your unconscious / system 1 / gut feel. LLMs are less like one's whole mind, and much more like one's "inner voice". Logical skills aren't automatic, they're algorithmic. Who knows what is the limit of a design in which LLM as "system 1" operates a much larger, symbolic, algorithmic suite of "system 2" software? We're barely scratching the surface here.

◧◩◪◨
14. behnam+oo[view] [source] [discussion] 2023-05-16 21:40:56
>>causal+mg
My thoughts exactly. It's hard to see signal among all the noise surrounding LLMs, Even if they say they're gonna hurt you, they have no idea about what it means to hurt, what is "you", and how they're going to achieve that goal. They just spit out things that resemble people have said online. There's no harm from a language model that's literally a "language" model.
replies(2): >>visarg+Ns >>forget+zt
◧◩◪◨
15. TeMPOr+so[view] [source] [discussion] 2023-05-16 21:41:41
>>Eamonn+bd
Yes, growing. It's not that the Voices have suddenly been "platformed" - it's that the field made a bunch of rapid jumps which made the message of those Voices more timely.

Recent developments in AI only further confirm that the logic of the message is sound, and it's just the people that are afraid the conclusions. Everyone has their limit for how far to extrapolate from first principles, before giving up and believing what one would like to be true. It seems that for a lot of people in the field, AGI X-risk is now below that extrapolation limit.

replies(4): >>b112+Cx >>rtkwe+AA >>2muchc+j71 >>dhkk+cb1
◧◩
16. mschus+yp[view] [source] [discussion] 2023-05-16 21:47:56
>>EGreg+Ic
At least AI has legitimate, actual use cases.
replies(1): >>EGreg+HR
◧◩◪◨⬒
17. staunt+pr[view] [source] [discussion] 2023-05-16 21:58:21
>>ben_w+Xj
Indeed, AI reinforcement-learning to deal with formal verification is what I'm looking forward to the most. Unfortunately it seems a very niche endeavour at the moment.
◧◩◪◨
18. jack_p+bs[view] [source] [discussion] 2023-05-16 22:02:49
>>Eamonn+bd
When the sky is getting to a dark shade of red it makes sense to hear out the doomsayers
replies(1): >>matwoo+Vs
◧◩◪◨⬒
19. visarg+Ns[view] [source] [discussion] 2023-05-16 22:06:31
>>behnam+oo
A language model can do many things based on language instructions, some harmless, some harmful. They are both instructable and teachable. Depending on the prompt, they are not just harmless LLMs.
◧◩◪◨⬒
20. matwoo+Vs[view] [source] [discussion] 2023-05-16 22:07:05
>>jack_p+bs
And the vast majority of the time it's just a nice sunset.
replies(2): >>jack_p+wv >>nullse+2L
◧◩◪◨⬒
21. forget+zt[view] [source] [discussion] 2023-05-16 22:11:02
>>behnam+oo
You appear to be ignoring a few thousand years of recorded history around what happens when a demagogue gets a megaphone. Human-powered astroturf campaigns were all it took to get randoms convinced lizard people are an existential threat and then -act- on that belief.
replies(2): >>nullse+KA >>goatlo+h61
◧◩◪◨⬒⬓
22. jack_p+wv[view] [source] [discussion] 2023-05-16 22:22:15
>>matwoo+Vs
a sunset at lunch time hits different
◧◩◪◨⬒
23. b112+Cx[view] [source] [discussion] 2023-05-16 22:35:54
>>TeMPOr+so
it's that the field made a bunch of rapid jumps

I wish I knew what we really have achieved here. I try to talk to these things, via turbo3.5 api, amd all I get is broken logic, twisted moral reasoning, all due to oipenai manually breaking their creation.

I don't understand their whole filter business. It's like we found a 500 yr old nude painting, a masterpiece, and 1800 puritans painted a dress on it.

I often wonder if the filter, is more to hide its true capabilities.

replies(4): >>stuckk+tD >>andyba+7E >>ux-app+4F >>TeMPOr+wP
◧◩
24. smcin+Kx[view] [source] [discussion] 2023-05-16 22:36:51
>>EGreg+Ic
That was entirely different, and a play to muddy the regulatory waters and maybe buy him time: the CFTC is much smaller (budget, staff) than the SEC, and less aggressive in criminal enforcement. Aided by a bill introduced by crypto-friendly Sens Lummis and Gillibrand [https://archive.ph/vqHgC].
◧◩◪◨⬒
25. rtkwe+AA[view] [source] [discussion] 2023-05-16 22:58:50
>>TeMPOr+so
What's the actual new advancements? LLMs to me are great at faking AGI but are no where near actually being a workable general AI. The biggest example to me is you can correct even the newest ChatGPT and ask it to be truthful but it'll make up the same lie within the same continuous conversation. IMO the difference between being able to act truth-y and actually being truthful is a huge gap that involves the core ideas of what separates an actual AGI and a really good chatbot.

Maybe it'll turn out to be a distinction that doesn't matter but I personally still think we're a ways away from an actual AGI.

replies(2): >>ux-app+nE >>canjob+R71
◧◩◪◨⬒⬓
26. nullse+KA[view] [source] [discussion] 2023-05-16 22:59:34
>>forget+zt
I think I'm just going to build and open source some really next gen astroturf software that learns continuously as it debates people online in order to get better at changing people's minds. I'll make sure to include documentation in Russian, Chinese and Corporate American English.

What would a good name be? TurfChain?

I'm serious. People don't believe this risk is real. They keep hiding it behind some nameless, faceless 'bad actor', so let's just make it real.

I don't need to use it. I'll just release it as a research project.

replies(2): >>forget+QY >>edgyqu+Sc1
◧◩
27. yarg+BB[view] [source] [discussion] 2023-05-16 23:05:50
>>pg_123+X8
There's no chance that we've peeked from a bang for buck sense - we still haven't adequately investigated sparse networks.

Relevantish: https://arxiv.org/abs/2301.00774

The fact that we can reach those levels of sparseness with pruning also indicates that we're not doing a very good job of generating the initial network conditions.

Being able to come up with trainable initial settings for sparse networks across different topologies is hard, but given that we've had a degree of success with pre-trained networks, pre-training and pre-pruning might also allow for sparse networks with minimally compromised learning capabilities.

If it's possible to pre-train composable network modules, it might also be feasible to define trainable sparse networks with significantly relaxed topological constraints.

replies(3): >>alexel+2E >>stephc+LI >>cma+yJ
◧◩◪◨
28. bernie+WC[view] [source] [discussion] 2023-05-16 23:13:14
>>causal+mg
I’m agreeing with this viewpoint the more I use LLMs.

They’re text generators that can generate compelling content because they’re so good at generating text.

I don’t think AGI will arise from a text generator.

◧◩◪◨⬒⬓
29. stuckk+tD[view] [source] [discussion] 2023-05-16 23:15:58
>>b112+Cx
Lately I have been picturing comments, and this is truly iconic haha.
◧◩◪◨⬒
30. fsckbo+GD[view] [source] [discussion] 2023-05-16 23:17:17
>>ben_w+Xj
> using the Great Pyramid of Giza as an AirBnB

, were you allowed to do it, would be an extremely profitable venture. Taj Mahal too, and yes, I know it's a mausoleum.

replies(1): >>ben_w+IH1
◧◩◪◨
31. stuckk+YD[view] [source] [discussion] 2023-05-16 23:18:38
>>causal+mg
I was looking at the A100 80gb cards. 14k a pop. We gonna see another GPU shortage when these models become less resource dependent. CRYPTO era
◧◩◪
32. alexel+2E[view] [source] [discussion] 2023-05-16 23:19:04
>>yarg+BB
I don’t think you really disagree with GP? I think the argument is we peaked on “throw GPUs at it”?

We have all kinds of advancements to make training cheaper, models computationally cheaper, smaller, etc.

Once that happens/happened, it benefits OAI to throw up walls via legislation.

replies(2): >>Neverm+eK >>yarg+vQ1
◧◩◪◨⬒⬓
33. andyba+7E[view] [source] [discussion] 2023-05-16 23:19:23
>>b112+Cx
You seem to be making several points at once and I'm not sure they all join up?
◧◩◪◨⬒⬓
34. ux-app+nE[view] [source] [discussion] 2023-05-16 23:21:24
>>rtkwe+AA
>Maybe it'll turn out to be a distinction that doesn't matter but I personally still think we're a ways away from an actual AGI.

if you had described GPT to me 2 years ago I would have said no way, we're still a long way away from a machine that can fluidly and naturally converse in natural language and perform arbitrary logic and problem solving, and yet here we are.

I very much doubt that in 5 years time we'll be talking about how GPT peaked in 2023.

replies(2): >>q7xvh9+IE >>rtkwe+DU
◧◩◪◨⬒⬓⬔
35. q7xvh9+IE[view] [source] [discussion] 2023-05-16 23:24:18
>>ux-app+nE
Seriously. It's worth pausing for a minute to note that the Turing Test has been entirely solved.

In fact, it has been so thoroughly solved that anyone can download an open-source solution and run it on their computer.

And yet, the general reaction of most people seems to be, "That's kind of cool, but why can't it also order me a cheeseburger?"

replies(3): >>eroppl+qG >>canjob+x81 >>2muchc+l91
◧◩◪◨⬒⬓
36. ux-app+4F[view] [source] [discussion] 2023-05-16 23:26:31
>>b112+Cx
>I often wonder if the filter, is more to hide its true capabilities.

right now we're all sharing a slice of GPT. I wouldn't be at all surprised if there's some uber GPT (which requires a lot more processing per response) running in a lab somewhere that blows what's publicly available out of the water.

◧◩◪◨
37. ux-app+nF[view] [source] [discussion] 2023-05-16 23:29:13
>>causal+mg
>They're still terrible at logical reasoning.

2 years ago a machine that understands natural language and is capable of any arbitrary, free-form logic or problem solving was pure science fiction. I'm baffled by this kind of dismissal tbh.

>but LLMs are never going to go into a recursive loop of self-improvement

never is a long time.

replies(1): >>leonid+aQ
◧◩◪◨⬒⬓⬔⧯
38. eroppl+qG[view] [source] [discussion] 2023-05-16 23:34:51
>>q7xvh9+IE
It has not been solved. Even GPT-4, as impressive as it is for some use cases, is dumb and I can tell the difference between it and a human in a dozen sentences just by demanding sufficient precision.

In some contexts, will some people be caught out? Absolutely. But that's been happening for a while now.

replies(1): >>ben_w+u32
◧◩◪
39. stephc+LI[view] [source] [discussion] 2023-05-16 23:49:34
>>yarg+BB
The efficiency of training has very unlikely reached its peak or near its peak. We are still inefficient. But the bottleneck might be elsewhere, in data, what we use to feed them.

Maybe not peaked yet, but the case can be made that we’re not seeing infinite supply…

◧◩◪
40. cma+yJ[view] [source] [discussion] 2023-05-16 23:55:15
>>yarg+BB
50% sparsity is almost certainly already being used given that it is accelerated in current nvidia hardware both at training time, usable dynamically through RigL ("Rigging the Lottery: Making All Tickets Winners" https://arxiv.org/pdf/1911.11134.pdf )--which also addresses your point about initial conditions being locked in-- and at accelerates 50% sparsity at inference time.
◧◩◪◨
41. Neverm+eK[view] [source] [discussion] 2023-05-17 00:00:02
>>alexel+2E
No way has training hit any kind of cost, computing or training data efficiency peak.

Big tech advances, like the models of the last year or so, don't happen without a long tail of significant improvements based on fine tuning, at a minimum.

The number of advances being announced by disparate groups, even individuals, also indicates improvements are going to continue at a fast pace.

◧◩◪◨
42. lostms+KK[view] [source] [discussion] 2023-05-17 00:03:05
>>Eamonn+bd
Growing is quite apt here. No matter what you or I think more and more people get the sense of AI coming and talk about it.
◧◩◪◨⬒⬓
43. nullse+2L[view] [source] [discussion] 2023-05-17 00:05:15
>>matwoo+Vs
I'm so glad that we 100% know for sure that this too is the vast majority of the time.
◧◩◪
44. turtle+lL[view] [source] [discussion] 2023-05-17 00:07:51
>>samsta+wh
Which anime(s)? If ANIME is the title, that's going to be hard to search.

Do you mean like Serial Experiments Lain?

replies(1): >>throwa+IQ
◧◩◪
45. chii+pP[view] [source] [discussion] 2023-05-17 00:37:48
>>briant+8h
> there are many ideas & perspectives that are simply not heard or considered.

of course, but just because those ideas are unheard, doesn't mean they are going to be any better.

An idea should stand on its own merits, and be evaluated objectively. It doesn't matter who was doing the proposing.

Also the problem isn't that bad ideas might get implemented, but that the legislature isn't willing or able to make updates to laws that encoded a bad idea. Perhaps it isnt known that it is a bad idea until after the fact, and the methods of democracy we have today isn't easily able to force updates to bad laws encoding bad ideas.

replies(1): >>briant+kW
◧◩◪◨⬒⬓
46. TeMPOr+wP[view] [source] [discussion] 2023-05-17 00:38:13
>>b112+Cx
> I wish I knew what we really have achieved here. I try to talk to these things, via turbo3.5 api, amd all I get is broken logic

Try to get your hands on GPT-4, even if it means paying the $20/mo subscription for ChatGPT Plus. There is a huge qualitative jump between the two models.

I got API access to GPT-4 some two weeks ago; my personal experience is, GPT-3.5 could handle single, well-defined tasks and queries well, but quickly got confused by anything substantial. Using it was half feelings of amazement, and half feelings of frustration. GPT-4? Can easily handle complex queries and complex tasks. Sure, it still makes mistakes, but much less frequently. GPT-4 for me is 80% semi-reliable results, 20% trying to talk it out of pursuing directions I don't care about.

Also, one notable difference: when GPT-4 gives me bad or irrelevant answers, most of the time this is because I didn't give it enough context. I.e. it's my failure at communicating. A random stranger, put in place of GPT-4, would also get confused, and likely start asking me questions (something LLMs generally don't do yet).

> I don't understand their whole filter business.

Part preferences, part making its "personality" less disturbing, and part PR/politics - last couple times someone gave the general public access to an AI chatbot, it quickly got trolled, and then much bad press followed. Doesn't matter how asinine the reaction was - bad press is bad press, stocks go down. Can't have it.

> I often wonder if the filter, is more to hide its true capabilities.

I don't think it's to hide the model's capabilities, but it's definitely degrading them. Kind of expected - if you force-feed the model with inconsistent and frequently irrational overrides to highly specific topics, don't be surprised if the model's ability to (approximate) reason starts to break down. Maybe at some point LLMs will start to compartmentalize, but we're not there yet.

◧◩◪◨⬒
47. leonid+aQ[view] [source] [discussion] 2023-05-17 00:42:23
>>ux-app+nF
Two years ago we already had GPT-2, that was capable of some problem solving and logic following. It was archaic, sure, it produced a lot of gibberish, yes, but if you followed OpenAI releases closely, you wouldn't think that something like GPT3.5 was "pure science fiction", it would just look as the inevitable evolution of GPT-2 in a couple of years given the right conditions.
replies(2): >>ux-app+0R >>canjob+591
◧◩◪
48. throwa+uQ[view] [source] [discussion] 2023-05-17 00:45:08
>>samsta+wh
This is a very interesting post. I don't understand this part: <<AI is the policy-makers it will expose>> Can you help to explain a different way?

And hat tip to this comment:

    Have you watched ANIME ever? Yeah... its here now.
The more I watch the original Ghost in the Shell, the more I think it has incredible foresight.
replies(2): >>samsta+RV2 >>samsta+gP3
◧◩◪◨
49. throwa+IQ[view] [source] [discussion] 2023-05-17 00:45:36
>>turtle+lL
One idea / suggestion: The original Ghost in the Shell.
◧◩◪◨⬒⬓
50. ux-app+0R[view] [source] [discussion] 2023-05-17 00:47:39
>>leonid+aQ
that's pedantic. switch 2 years to 5 years and the point still stands.
replies(1): >>edgyqu+Hc1
◧◩◪
51. EGreg+HR[view] [source] [discussion] 2023-05-17 00:52:32
>>mschus+yp
True but also other use cases with much worse outcomes than blockchain could ever have
◧◩◪◨⬒⬓⬔
52. rtkwe+DU[view] [source] [discussion] 2023-05-17 01:15:11
>>ux-app+nE
Again I'm not really saying GPT has peaked I'm saying there's a categorical difference between GPT and AGI. A good enough fake might perform well enough to function like one but I have my doubts that it will. Without a way to deal with and, to some sense of the word understand, facts I don't think LLMs are suitable for use as anything beyond an aide for humans (for starters because they can't determine internally what is and isn't a fact vs hallucination so you have to constantly check their work).
◧◩
53. parent+3W[view] [source] [discussion] 2023-05-17 01:30:02
>>brooks+Rd
The subtle difference between the original statement and yours:

Ideas that drive governing decisions should be globally good - meaning there should be more than just @sama espousing them.

replies(1): >>r_hood+lg1
◧◩◪◨
54. briant+kW[view] [source] [discussion] 2023-05-17 01:32:26
>>chii+pP
> of course, but just because those ideas are unheard, doesn't mean they are going to be any better.

It probably does mean it's better at least for the person with the perspective. Too bad only a very few get a seat at the table to advocate for their own interests. It would be better if everyone has agency to advocate for their interests.

> Also the problem isn't that bad ideas might get implemented, but that the legislature isn't willing or able to make updates to laws that encoded a bad idea

First, this is a hyped up crisis where some people are claiming it will end humanity. There have been many doomsday predictions & people scared by these predictions are effectively scammed by those fomenting existential fear. It's interesting that the representatives of large pools of capital are suddenly existentially afraid when there is open source competition.

Second, once something is in the domain of government it will only get more bloated & controlled by monied lobbyists. The legislatures controlled by lobbyists will never make it better, only worse. There have been so many temporary programs that continue to exist & expand. Many bloated omnibus bills too long to read passed under some sort of "emergency". The government's tendency is to grow & to serve the interests of the corporations that pay the politicians. Fear is an effective tool to convince people to accept things against their interests.

◧◩◪◨⬒⬓⬔
55. forget+QY[view] [source] [discussion] 2023-05-17 01:56:50
>>nullse+KA
It's not like there isn't a market waiting impatiently for the product...
replies(1): >>nullse+2p1
◧◩
56. dreamc+P01[view] [source] [discussion] 2023-05-17 02:13:54
>>brooks+Rd
The recent history of tech CEOs advocating for regulations only they can obey has become so blatant that any tech CEO who advocates for regulation should be presumed guilty until proven innocent.
replies(1): >>brooks+se2
◧◩
57. bcrosb+D51[view] [source] [discussion] 2023-05-17 03:06:10
>>brooks+Rd
Not when it comes to politics.

You'll be stuck in the muck while they're laughing their ass off all the way to the bank.

replies(1): >>jasonm+v61
◧◩◪◨⬒⬓
58. goatlo+h61[view] [source] [discussion] 2023-05-17 03:12:22
>>forget+zt
So the LLM demagogue is going to get people to create gray goo or make a lot of paper clips?
◧◩◪
59. jasonm+v61[view] [source] [discussion] 2023-05-17 03:15:22
>>bcrosb+D51
It doesn't even matter if "his heart is pure" ... Companies are not run that way.

We have lawyers.

◧◩◪◨⬒
60. 2muchc+j71[view] [source] [discussion] 2023-05-17 03:25:43
>>TeMPOr+so
How do we define a general intelligence?
◧◩◪◨⬒⬓
61. canjob+R71[view] [source] [discussion] 2023-05-17 03:31:05
>>rtkwe+AA
The fact that it’s a system you’d even consider to be “lying” or “truthful” is a huge advance over anything available 5 years ago.
replies(1): >>rtkwe+382
◧◩◪◨⬒⬓⬔⧯
62. canjob+x81[view] [source] [discussion] 2023-05-17 03:37:45
>>q7xvh9+IE
As impressive as GPT4 is, it still makes mistakes humans would not, and so I wouldn’t say it passes the Turing Test.

I generated this just now:

Me: Jack and Jill are sitting next to each other in a room. There is no one else in the room. The person sitting next to Jack is sad. The person sitting next to Jill is angry. Who is sad, Jack or Jill?

GPT4: Based on the information provided, it is not possible to determine who is sad and who is angry, as there is a contradiction in the given statements. If Jack and Jill are the only ones in the room and sitting next to each other, then the person sitting next to Jack would be Jill and the person sitting next to Jill would be Jack. The statements about their emotions conflict with each other, so it is not possible to accurately determine who is sad or angry.

replies(1): >>jdkee+Rb1
◧◩◪◨⬒⬓
63. canjob+591[view] [source] [discussion] 2023-05-17 03:43:35
>>leonid+aQ
In hindsight it’s an obvious evolution, but in practice vanishingly few people saw it coming.
replies(1): >>leonid+dF2
◧◩◪◨⬒⬓⬔⧯
64. 2muchc+l91[view] [source] [discussion] 2023-05-17 03:45:55
>>q7xvh9+IE
>And yet, the general reaction of most people seems to be, "That's kind of cool, but why can't it also order me a cheeseburger?”

Well yeah. Imagine you tell a small child that knows about calculators, “Hey can you work out 18763 + 38284, for me?” They might struggle and then maybe fetch a calculator.

The LLMs attempt to predict the answer. WTF? It’s a computer and it can see that is a plus sign. Just understand its addition, and use the rest of your computer brain and do the sum. Hell, it is connected to the internet and we just taught you everything since before 2021. Just call out to Wolfram and give me the answer.

But that’s not how computers work. And we keep saying “AI” but that I is doing a lot of heavy lifting.

◧◩◪◨⬒
65. dhkk+cb1[view] [source] [discussion] 2023-05-17 04:10:14
>>TeMPOr+so
Sources please. Every expert interview ive seen with AI researchers who have been in the game since the beginning have said the same: GPT's are not a massive breakthrough in the field of AI research.
replies(1): >>TeMPOr+tU1
◧◩◪◨⬒⬓⬔⧯▣
66. jdkee+Rb1[view] [source] [discussion] 2023-05-17 04:17:35
>>canjob+x81
Jack and Jill are sitting next to each other in a room. There is no one else in the room. The person sitting next to Jack is sad. The person sitting next to Jill is angry. Who is sad, Jack or Jill?

ChatGPT Since Jack and Jill are the only two people in the room and they are sitting next to each other, the person sitting next to Jack is Jill and the person sitting next to Jill is Jack. Given the conditions you provided:

Jill is the one who is sad because she is sitting next to Jack. Jack is the one who is angry because he is sitting next to Jill.

replies(1): >>ChatGT+Gv1
◧◩◪◨⬒⬓⬔
67. edgyqu+Hc1[view] [source] [discussion] 2023-05-17 04:27:14
>>ux-app+0R
No it isn’t. Even before transformers people were doing cool things with LSTMs and RNNs before that. People following this space haven’t really been surprised by any of these advancements. It’s a straight forward path imo
◧◩◪◨⬒⬓⬔
68. edgyqu+Sc1[view] [source] [discussion] 2023-05-17 04:28:45
>>nullse+KA
I just don’t see how it’s going to be significantly worse than existing troll farms etc. This prediction appears significantly overblown to me.
replies(2): >>forget+Ki1 >>nullse+WN1
◧◩◪
69. r_hood+lg1[view] [source] [discussion] 2023-05-17 05:08:01
>>parent+3W
You're defending an argument that is blatantly self contradictory within the space of two sentences.

A) "anything he suggests should be categorically rejected because he’s just not in a position to be trusted."

B) "If what he suggests are good ideas then hopefully we can arrive at them in some other way with a clean chain of custody."

These sentences directly follow each other and directly contradict each other. Logically you can't categorically (the categorical is important here. Categorical means something like "treat as a universal law") reject a conclusion because it is espoused by someone you dislike, while at the same time saying you will accept that conclusion if arrived at by some other route.

"I will reject P if X proposes P, but will accept P if Y proposes P." is just poor reasoning.

replies(1): >>brooks+Re2
◧◩◪◨⬒⬓⬔⧯
70. forget+Ki1[view] [source] [discussion] 2023-05-17 05:35:27
>>edgyqu+Sc1
Does it really? You thinking LLM-powered propaganda distribution services can't out-scale existing troll farms? Or do a better job of evading spam filters?
replies(1): >>edgyqu+JC2
◧◩◪◨⬒⬓⬔⧯
71. nullse+2p1[view] [source] [discussion] 2023-05-17 06:46:53
>>forget+QY
It's definitely not something I would attempt to productize and profit off of. I'm virtually certain someone will, and I'm sure that capability is being worked on as we speak, since we already know this type of thing occurs at scale.

My motivation would be simply shine a light on it. Make it real for people, so we have things to talk about other than just the hypotheticals. It's the kind of tooling that if you're seriously motivated to employ it, you'd probably prefer it remain secret or undetected at least until after it had done it's work for you. I worry that the 2024 US election will be the real litmus test for these things. All things considered it'd be a shame if we go through another Cambridge Analytica moment that in hindsight we really ought to have seen coming.

Some people have their doubts, and I understand that. These issues are so complex that no one individual can hope to have an accurate mental model of the world that is going to serve them reliabily again and again. We're all going to continue to be surprised as events unfold, and the degree to which we are surprised indicates the degree to which our mental models were lacking and got updated. That to me is why I'm erring on the side or pessimism and caution.

◧◩◪◨⬒⬓⬔⧯▣▦
72. ChatGT+Gv1[view] [source] [discussion] 2023-05-17 07:54:36
>>jdkee+Rb1
Does the fact you both got difference answers make it better or worse at reasoning ?
◧◩
73. daniel+bz1[view] [source] [discussion] 2023-05-17 08:31:34
>>pg_123+X8
If “peaked” means impact and “bang for buck” means per dollar then its only peaked if the example is allowing the population at large to use these free tools like chatbots, for fun and minimal profits. but if we consider how they can be used to manipulate people at scale with misinformation then that’s an example where I think we’ve not yet seen the peak. So we should at least thoroughly discuss or think of it of to see if we can in any way mitigate certain negative societal outcomes.
◧◩
74. dTal+JF1[view] [source] [discussion] 2023-05-17 09:33:34
>>brooks+Rd
I think what they are trying to say is that Sam Altman is very smart, but misaligned. If we assume that he is 1) sufficiently smart and 2) motivated to see OpenAI succeed, then his suggestions must be assumed to lead to a future where OpenAI is successful. If that future looks like it contradicts a future we want (for instance, user-controlled GPT-4 level AIs running locally on every machine), his suggestions should therefore be treated as reliably radioactive.
◧◩◪◨⬒⬓
75. ben_w+IH1[view] [source] [discussion] 2023-05-17 09:48:14
>>fsckbo+GD
I can see the reviews in my head already:

1 star: No WiFi, no windows, no hot water

1 star: dusty

1 star: aliens didn't abduct me :(

5 stars: lots of storage room for my luggage

4 stars: service good, but had weird dream about a furry weighing my soul against a feather

1 star: aliens did abduct me :(

2 stars: nice views, but smells of camel

◧◩◪◨⬒⬓⬔⧯
76. nullse+WN1[view] [source] [discussion] 2023-05-17 10:49:31
>>edgyqu+Sc1
Only time will tell.
◧◩◪◨
77. yarg+vQ1[view] [source] [discussion] 2023-05-17 11:13:14
>>alexel+2E
Yeah, it's a little bit RTFC to be honest.
◧◩◪◨⬒⬓
78. TeMPOr+tU1[view] [source] [discussion] 2023-05-17 11:47:03
>>dhkk+cb1
> Sources please.

My own eyes? Hundreds of thousands thousand different scientific papers, blog posts, news reports and discussion threads that covered this ever since ChatGPT appeared, and especially in the last two months as GPT-4 rolled out?

At this point I'd reconsider if the experts you listened to are in fact experts.

Seriously. It's like saying Manhattan project wasn't a massive breakthrough in experimental physics or military strategy.

replies(1): >>dhkk+Ez3
◧◩◪◨⬒⬓⬔⧯▣
79. ben_w+u32[view] [source] [discussion] 2023-05-17 12:48:30
>>eroppl+qG
"Dumb" isn't why the Turing Test isn't solved. (Have you seen unmoderated chat with normal people? Heck, even smart people outside the domain of expertise; my mum was smart enough to get into university in the UK in the early 60s, back when that wasn't the default, but still believed in the healing power of crystals, homeopathic sodium chloride and silicon dioxide, and Bach flower remedies…)

ChatGPT (I've not got v4) deliberately fails the test by spewing out "as a large language model…", but also fails incidentally by having an attention span similar to my mother's shortly after her dementia diagnosis.

The problem with 3.5 is that it's simultaneously not mastered anything, and yet also beats everyone in whatever they've not mastered — an extremely drunk 50,000 year old Sherlock Holmes who speaks every language and has read every book just isn't going to pass itself off as Max Musstermann in a blind hour-long trial.

replies(1): >>rtkwe+F92
◧◩◪◨⬒⬓⬔
80. rtkwe+382[view] [source] [discussion] 2023-05-17 13:15:32
>>canjob+R71
That's more a convenience of language than an actual "It's Alive!". Calling them hallucinations or inaccuracies is unwieldy and the former has the same kind of implied attribution of a mind. We know for sure that's not there, my internal model for those is just a stupendously complex markov chain because to my understanding that's all LLMs are currently doing.
◧◩◪◨⬒⬓⬔⧯▣▦
81. rtkwe+F92[view] [source] [discussion] 2023-05-17 13:23:25
>>ben_w+u32
The lack of an ability to take in new information is maybe the crux of my issues with the LLM to AGI evolution. To my understanding the only way to have it even kind of learn something is to include it in a preamble it reprocesses every time which is maybe workable for small facts but breaks down for updating it from the 202X corpus it was trained on.
replies(1): >>ben_w+8L2
◧◩◪
82. brooks+se2[view] [source] [discussion] 2023-05-17 13:51:44
>>dreamc+P01
Sure, go execute him for all I care.

My point was that an idea shold not need attribution for you to know whether it's good or bad, for your own purposes. I can't imagine looking at a proposal and deciding whether to support or oppose based on the author rather than content.

If Altman is that smart and manipulative, all he has to do is advocate the opposite of what he wants and you'll be insisting that we must give him exactly what he wants, on principle. That's funny with kids but no way to run public policy.

◧◩◪◨
83. brooks+Re2[view] [source] [discussion] 2023-05-17 13:53:29
>>r_hood+lg1
More clearly said than I managed, yep.

But I suppose it comes down to priorities. If good policy is less important than contradicting P, I suppose that approach makes sense.

◧◩◪◨⬒⬓⬔⧯▣
84. edgyqu+JC2[view] [source] [discussion] 2023-05-17 15:27:20
>>forget+Ki1
No I’m thinking that scaling trolls up has diminishing returns and we’re already peak troll.
replies(1): >>nullse+XK3
◧◩◪◨⬒⬓⬔
85. leonid+dF2[view] [source] [discussion] 2023-05-17 15:37:58
>>canjob+591
Few people saw it coming in just two years, sure. But most people following this space were already expecting a big evolution like the one we saw in 5-ish years.

For example, take this thread: https://news.ycombinator.com/item?id=21717022

It's a text RPG game built on top of GPT-2 that could follow arbitrary instructions. It was a full project with custom training for something that you can get with a single prompt on ChatGPT nowadays, but it clearly showcased what LLMs were capable of and things we take for granted now. It was clear, back then, that at some point ChatGPT would happen.

◧◩◪◨⬒⬓⬔⧯▣▦▧
86. ben_w+8L2[view] [source] [discussion] 2023-05-17 15:59:49
>>rtkwe+F92
Mmm. Well, possibly.

On the one hand, what I was saying here was more about the Turing Test than about AGI. Sometimes it gets called the AGI, sometimes it's "autocomplete on steroids", but even if it is fancy autocomplete, I think 3.5 has the skill to pass a short Turing Test, but not the personality, and it needs a longer "short-term memory"-equivalent than 3.5 for a full Turing Test.

On the other hand, as I (sadly) don't get paid to create LLMs, I've only got the kind of superficial awareness of how they work that comes from podcasts and the occasional blog post, which means ChatGPT might very well understand ChatGPT better than I do.

Can GPT-[3.5, 4] be prompted to make itself?

◧◩◪◨
87. samsta+RV2[view] [source] [discussion] 2023-05-17 16:43:14
>>throwa+uQ
ANIME predicted the exact corprate future...

Look at all anime cyber cities...

Its not as high tech as you may imagine, but the surveillance is there

EDIt your "company" is watching every

◧◩◪◨⬒⬓⬔
88. dhkk+Ez3[view] [source] [discussion] 2023-05-17 19:51:01
>>TeMPOr+tU1
It was Yann LeCun. His professional experience and knowledge of the AI development timeline outweighs your opinions, imo. Thanks for confirming you have no sources.
◧◩◪◨⬒⬓⬔⧯▣▦
89. nullse+XK3[view] [source] [discussion] 2023-05-17 20:44:53
>>edgyqu+JC2
Any evidence or sources for that? I just don't know how that would be knowable to any of us.

Yuval Noah Harari gave a great talk the other day on the potential threat to democracy from the current state of the technology - https://youtu.be/LWiM-LuRe6w

◧◩◪◨
90. samsta+gP3[view] [source] [discussion] 2023-05-17 21:11:25
>>throwa+uQ
don't understand this part: <<AI is the policy-makers it will expose>> Can you help to explain a different way?*

===

Policy makers will not understanding what they are doing.

[go to top]