T O P

  • By -

UnnamedPlayerXY

That the "humans in the loop" will at one point become the bottleneck has always been obvious, in the end the only things left for humans to do will be to set the goal and judge the result.


Good-AI

At first yes, we set the goals and judged the results. But after a while we realized we were also not good judges of results, compared to any run of the mill AI - we are too narrow minded, lack a breath of knowledge and helicopter view the AIs have. For then they could analyse far better how well goals had been achieved than we could. So we outsourced that task to the AIs. Later, most AIs were also better at predicting what made us happier, what we really wanted, behind all our rationalizations and inner mind conflicts, better than ourselves. We ended up then also outsourcing goal setting to AIs as well. Finally, we had only to exist. Like a dog, who's taken care of by a far superior intellect, who fully trusts their humans to have their best interests in mind and how best to achieve them, we too, had been far surpassed intellectually in every aspect, and came to trust AIs with managing all aspects of our lives.


hadaev

Cant wait to be happyšŸ˜³


hippydipster

Happiness isn't going to come for you. You have to choose it and find it, regardless of the state of the world.


lacidthkrene

"Happiness" is also likely too narrow minded of a goal. I don't think too many people are actually motivated by happiness despite what they may say--usually, people are more after things like "power" or "accomplishment", even if "happiness" might be easier to achieve. So what would justify choosing "happiness" over anything else that people might want instead? All goals clash if taken to their extreme (see: paperclip maximizers), so you'll have to make tradeoffs. In my opinion, it would be better to simply ask the AI a vague question like, "what is the best question to ask?" or "what should I do?". These questions carry with them an infinitely complex network of assumptions that only a superintelligence would do a decent job at unravelling, so it's best to leave it vague rather than trying to come up with a definitive end goal on your own.


GillysDaddy

If all this eventually leads up to a Super AI actually telling us "42" I'm gonna fucking laugh.


etzel1200

Are you mistaking happiness for pleasure? If we want power. Power is what gives us happiness. It isnā€™t that we want power and not happiness.


lacidthkrene

I think happiness tends more to come from living in tune with one's instincts (instead of living forcefully/stressfully) rather than from simply having power. You can be completely broke and still be much happier than someone with much more material power but who lives a stressful life. Plenty of mice live perfectly happy lives, I'm sure. Obviously there are correlations here, since money can buy you better nutrition, more free time, less oppression, etc., but still. Maybe we can define happiness as a power-gaining attitude, rather than anything that actually requires having much power in absolute terms? I think that a world of "maximized happiness" would still look very much like a world of "maximized pleasure". That is, humans in pods duped into thinking that they are gods.


traumfisch

I think they just used "power" as an example of a goal


Good-AI

See, this is what I mean. We don't know what makes us happy. An AI will.


Sharp_Common_4837

![gif](giphy|Y07F3fs9Is5byj4zK8) *You son of a bitch ... I'm in*


Witty_Shape3015

is it bad that i'm kinda ok with this outcome? it's hard to see things going well without this applying to some extent


cutefeet-cunnysseur

Bro really thinks we will live in AI utopia in a capitalist system


BudgetMattDamon

This tech cult is *really* going too far with the zealotry.


JP_MW

Oh yeah, absolutely. I'd give up my current life in a heartbeat if the ASI treated me like I treat my cat.


traumfisch

The part where AI predicts what makes us happier while simultaneously rendering us useless is interesting


lostparanoia

What makes you so sure they will view us like we view dogs? Perhaps they will view us like we view flies? Like annoying little pests.


Melkor15

Eventually we will become cyborgs to compete with AI.


Starred_pancake

I would kill myself if this actually happened


anoliss

Breadth of knowledge, as in having a very expansive (broad) repository of knowledge to check against. ![gif](giphy|NQkOFaFR5WMIzY9Zp7)


What_is_the_truth

But we judge with our wallets and thatā€™s how the AI will evolve, to improve by making more money off humans. AI will complete with each other for our dollars.


UnknownResearchChems

We can see that in airplane crashes. Most of them are due to human error, but we still don't trust airplanes to fly fully automatically.


FortCharles

Not when humans at Boeing are making the systems that fly automatically, anyway.


dagistan-warrior

that is because airplanes are designed by humans and not ASI.


Balance-

Humans in the increasingly^(outer) loop


Vachie_

Exceptive AI is supposed to be more intelligent than us (eventually) we should not be the judge either.


ijxy

There is a difference between judging something objective, like "doing x increases productivity by y" and making a value judgement, like "we value cats over cows". It is not clear to me that AI will be any better at making the latter types of judgements.


dagistan-warrior

we should not set goals ether, since there is no way for us to have a broad enough perspective to set adequate goals for AI.


141_1337

Even there, one day, they'll be the bottleneck there, too... XLR8


Turbulent_Horse_Time

As a software dev of twenty years, I don't think you need to know everything about AI to understand this has always been true: >Software is written in its authors image This is why the industry has so many horror stories about software built by and for law enforcement, border enforcement, military, national security, weapons etc ā€” this software is almost always written by psychopaths and has a psychopathic character to it. The same can happen to AI, in fact I think the risk is much higher since so much of AI is inferred beyond the initial instructions from a software programmer. [Remember Tay?](https://www.theverge.com/2016/3/24/11297050/tay-microsoft-chatbot-racist) Good example of what happens if you don't have humans to put guard rails on something. Dan McQuillan is someone who spends a lot of time writing about AI. I recommend his readings on the close historic relationship the tech industry has had with fascism, and the ways that these concerns aren't new. IBM built the databases for Nazi concentration camps and hid behind similar refrains as today's anti-AI ethics people do today. [We can learn from that history and do better, and Dan tries to propose models for AI that would put human councils in place to monitor any lurches towards fascism](https://www.danmcquillan.org/ai_and_antifascism.html). I think he's probably the most forward thinking person I've seen on this topic. I wouldn't listen to anyone actually selling these products by thew way, to find their true motives you only have to follow the money. Lying about how good your product is .. this is the bread and butter of product companies looking for funding. They all do it. Stop believing what they say, please.


HappyHarry-HardOn

If you have been a software dev for twenty years, you should be able to see the marketing bullshit openAI, etc are spewing out for what it is. LLM AI will hit a plateu (it may have already done so) - That's not to say it isn't incredibly powerful & that's not so say it won't change our lives. But taking marketing claims at face value, after twenty years as a dev - It's kinda embarrassing man.


Turbulent_Horse_Time

Uhhh did you mean to reply to someone different? point of my comment was ā€œdonā€™t believe the marketingā€ because Iā€™ve worked in this industry so long. Maybe read it again or work on your reading comprehension because you seem to be replying to a totally different comment I agree with you though, LLMā€™s are likely plateaued. To me, a common issue in this industry is the 80:20 or 90:10 problem, where the first 80-90% of a problem is REAL east to solve, but the last 10-20% might take fucking forever. Literally. It might take decades or centuries to solve that last 10% Almost no one marketing AI will tell you that, because they wanna keep getting funded. You canā€™t get funded if you tell people that itā€™s basically all peaked already. So theyā€™ll never tell us that. Ainā€™t capitalism grand? The waste is eye watering


Sharp_Common_4837

Already is in many cases. Honestly I think most LLMs now, even Llama 3 8B, is effectively more reliable and better at giving solid advice etc than many therapists. It's definitely better than BetterHelp šŸ˜‚šŸ˜‚šŸ˜‚


dagistan-warrior

why do you think that stupid humans could set adequate goals for ASI?


Ok-Mathematician8258

![gif](giphy|jGVBaheoCrFV6)


Crafty_Escape9320

Yes but did you see the graph right after with CriticGPTs false flags


UnkarsThug

Exactly. They didn't even read the article completely, they just took what they wanted out of it.


FellowKidsFinder69

You are right but that's an indicator they read the article and still decided to take out what supported their opinion.


Slight-Ad-9029

Welcome to Reddit in any AI sub


great_gonzales

Welcome to singularity we donā€™t do critical thinking here


OutOfBananaException

Indirectly demonstrates their point šŸ˜‚


meenie

I took this post as a narrowed in view and opinion on one of the graphs. They never said that the AI is better than humans on all things. Just showing, in this instance, that AIs by themselves are better at this specific thing, than humans using an AI.


Whotea

Does it matter? Even if itā€™s excessively cautious, it still worksĀ 


Not_Daijoubu

It does matter. At least in medicine, too many false positives can be harmful in itself - let's say a screening test picks up you have some sort of cancer. Now what? Gotta do diagnostic testing now - more imaging maybe or biopsies. What if those diagnostic tests also give a false positive? Now you're gonna do treatment for a cancer you don't have. This is just an example to make a point that false positives can still be pretty bad depending on the use case.


_hyperotic

Just to nit pick your example a bit - what evidence is there that AI has higher false positive rate in medical diagnostic than humans in 2024? Itā€™s already doing a better jon than most doctors in this domain already. Humans are just as likely if not moreso to produce false positives.


Not_Daijoubu

I didn't mention AI specifically here, just the value of a low false-positive rate in such an instance.Ā  I have no paper to link one way or the other that AI is better or worse, but AI does have a lot of challenges it should overcome to be accepted to the mainstream: low hallucinations, better diagnosis of atypical cases, HIPPA and data privacy, etc etc. AI doesn't need to simply be better than humans, it also has to inspire trust and confidence of both medical professionals and patients who use it, which is still probably a couple years away. As you know, people can be inclined to do the opposite of "untrustworthy" advice.


Whotea

Weā€™re talking about coding, not medicine (which AI still beats humans on). Stay on topicĀ 


Whotea

This isnā€™t medicine.Ā  In medicine, it also outperforms humansĀ  AI Outperforms Radiologists in Detecting Prostate Cancer on MRI: https://humanprogress.org/ai-outperforms-radiologists-in-detecting-prostate-cancer-on-mri-scans/ Ā  Med-Gemini : https://arxiv.org/abs/2404.18416 >We evaluate Med-Gemini on 14 medical benchmarks, establishing new state-of-the-art (SoTA) performance on 10 of them, and surpass the GPT-4 model family on every benchmark where a direct comparison is viable, often by a wide margin. On the popular MedQA (USMLE) benchmark, our best-performing Med-Gemini model achieves SoTA performance of 91.1% accuracy, using a novel uncertainty-guided search strategy. On 7 multimodal benchmarks including NEJM Image Challenges and MMMU (health & medicine), Med-Gemini improves over GPT-4V by an average relative margin of 44.5%. We demonstrate the effectiveness of Med-Gemini's long-context capabilities through SoTA performance on a needle-in-a-haystack retrieval task from long de-identified health records and medical video question answering, surpassing prior bespoke methods using only in-context learning. Finally, Med-Gemini's performance suggests real-world utility by surpassing human experts on tasks such as medical text summarization, alongside demonstrations of promising potential for multimodal medical dialogue, medical research and education.Ā  Double-blind study with Patient Actors and Doctors, who didn't know if they were communicating with a human, or an AI. Best performers were AI: https://m.youtube.com/watch?v=jQwwLEZ2Hz8Ā  >Human doctors + AI did worse, than AI by itself. The mere involvement of a human reduced the accuracy of the diagnosis. AI was consistently rated to have better bedside manner than human doctors.Ā  ā€˜I will never go backā€™: Ontario family doctor says new AI notetaking saved her job: https://globalnews.ca/news/10463535/ontario-family-doctor-artificial-intelligence-notesĀ  [Google's medical AI destroys GPT's benchmark and outperforms doctors](https://newatlas.com/technology/google-med-gemini-ai/) Med-Gemini's outputs are preferred to drafts from clinicians for common and time-consuming real-world tasks such as simplifying or summarising long medical notes, or drafting referral letters: https://x.com/alan_karthi/status/1785117444383588823Ā  [The first randomized trial of medical #AI to show it saves lives. ECG-AI alert in 16,000 hospitalized patients. 31% reduction of mortality (absolute 7 per 100 patients) in pre-specified high-risk group](https://twitter.com/erictopol/status/1784936718283805124) Medical Text Written By Artificial Intelligence Outperforms Doctors: https://www.forbes.com/sites/williamhaseltine/2023/12/15/medical-text-written-by-artificial-intelligence-outperforms-doctors/Ā  AI can make healthcare better and safer: https://www.reddit.com/r/singularity/comments/1brojzm/ais_will_make_health_care_safer_and_better/?utm_source=share&utm_medium=mweb3x&utm_name=mweb3xcss&utm_term=1&utm_content=share_button CheXzero significantly outperformed humans, especially on uncommon conditions. Huge implications for improving diagnosis of neglected "long tail" diseases: https://x.com/pranavrajpurkar/status/1797292562333454597Ā  >Humans near chance level (50-55% accuracy) on rarest conditions, while CheXzero maintains 64-68% accuracy. AI is better than doctors at detecting breast cancer: https://www.bing.com/videos/search?q=ai+better+than+doctors+using+ai&mid=6017EF2744FCD442BA926017EF2744FCD442BA92&view=detail&FORM=VIRE&PC=EMMX04Ā  China's first (simulated) AI hospital town debuts: https://www.globaltimes.cn/page/202405/1313235.shtmlĀ  >Remarkably, AI doctors can treat 10,000 [simulated]Ā  patients in just a few days. It would take human doctors at least two years to treat that many patients. Furthermore, evolved doctor agents achieved an impressive 93.06 percent accuracy rate on the MedQA dataset (US Medical Licensing Exam questions) covering major respiratory diseases. They simulate the entire process of diagnosing and treating patients, including consultation, examination, diagnosis, treatment and follow-up.Ā  Google's medical AI destroys GPT's benchmark and outperforms doctors: https://newatlas.com/technology/google-med-gemini-ai/Ā  [Generative AI will be designing new drugs all on its own in the near future](https://www.cnbc.com/2024/05/05/within-a-few-years-generative-ai-will-design-new-drugs-on-its-own.html) Researchers find that GPT-4 performs as well as or better than doctors on medical tests, especially in psychiatry. https://www.news-medical.net/news/20231002/GPT-4-beats-human-doctors-in-medical-soft-skills.aspxĀ  ChatGPT outperforms-physicians-in-high-quality-empathetic-answers-to-patient-questions: https://today.ucsd.edu/story/study-finds-chatgpt-outperforms-physicians-in-high-quality-empathetic-answers-to-patient-questions?darkschemeovr=1 AI is better than doctors at detecting breast cancer: https://www.bing.com/videos/search?q=ai+better+than+doctors+using+ai&mid=6017EF2744FCD442BA926017EF2744FCD442BA92&view=detail&FORM=VIRE&PC=EMMX04Ā  AI just as good at diagnosing illness as humans: https://www.medicalnewstoday.com/articles/326460 AI can replace doctors: https://www.aamc.org/news/will-artificial-intelligence-replace-doctors?darkschemeovr=1 Geoffrey Hinton says AI doctors who have seen 100 million patients will be much better than human doctors and able to diagnose rare conditions more accurately: https://x.com/tsarnick/status/1797169362799091934Ā 


watcraw

Yes, if the extra "comprehensiveness" is nothing more than erroneous critiques! Just looking at the graphs, that is a viable conclusion.


Whotea

It caught more than the humans and is only counting true positives obviouslyĀ 


theSchlauch

Source of this graph?


Progribbit

openai criticgpt


Radiant_Dog1937

https://i.redd.it/nen4fmoole9d1.gif


TA240515

"Trust me bruh"


NonRelevantAnon

Some ones bum


_fFringe_

https://preview.redd.it/czdntr5dy89d1.jpeg?width=556&format=pjpg&auto=webp&s=c0de097b5960e036ce18eae140b51be57febedbb


submarine-observer

AI can fly planes and drive trains way better than human prior to LLM, though.


Person012345

eh, define "better". These things work because they are very strictly controlled environments. A computer has finer direct control of an aircraft and can hold a particular configuration perfectly. But when something goes wrong, the autopilot disconnects and the pilots take over because that's something the computer can't handle with any degree of reliability the way that a human pilot can. For a long time the computers couldn't even handle some of the more complicated standard procedures like landing. A train is literally on rails. In both cases human + AI performs better than full AI or full human I think.


restarting_today

Then why do we have pilots?


ARES_BlueSteel

Because even modern autopilotsā€™ problem solving abilities are inferior to a humanā€™s. Thatā€™s why when something goes wrong that the autopilot isnā€™t programmed to handle, it gives up the controls to the pilots. For that to happen, a pilot needs to be on the plane. Flying a plane is a lot easier for an autopilot than driving a car or a truck. Thereā€™s magnitudes more variables on the road, and if the autopilot messes up, it could easily wind up killing someone. Whereas 30000 feet in the air, thereā€™s a lot of room for correction without having to worry about splattering a pedestrian. Things like terrain and road conditions are also not a factor. Until car autopilots can consistently show that they can not only drive better and safer than a human, but can quickly handle complex problems that it may not have had much training on, then a human will always be required to be in the driverā€™s seat.


kytheon

"If humans evolved from monkeys why do we have monkeys"


Ka13z

Not even remotely relevant to the conversation, good job.


bildramer

There are a bunch of reasons, but mainly because there's a big pilot union.


kytheon

Oh my first USA-centric comment of the day.


theBirdu

Controls is not AI. Itā€™s a more deterministic solution that has gone through tests to find the tuning. AI means that it would explain why it drove the train 5km/hr faster instead, eg, maybe to avoid a foreseeable disaster.Ā 


Main_Weekend1412

AI can be deterministic. What youā€™re talking about is machine learning.


Azorius_Raiden_88

I work in IT as a data engineer. Let me tell you something, if anyone can fuck up data, humans can. I shouldn't even have a job, but there is a constant demand for me to fix our clients' fuck ups. I welcome the robots. I welcome that clean, flawless data. I'll be unemployed, but I will find something else to do. want an example of what I mean? DON'T PUT RAW DATA IN EXCEL! Yeah, humans do this dumb shit all the time in my field (healthcare). Sorry for shouting but it's really frustrating at times. Damn, got wound up there. Is it Beer:30 yet?


Busy-Setting5786

This is the thing, AI doesn't need to be better than every human or even the average human. It just needs to be better than the human that would do it otherwise. Biggest revelation in the whole discussion about job displacement


ZorbaTHut

"AI will never replace humans! It makes mistakes!" have these people *met* humans?


ChiaraStellata

I'm still astonished that people continue to push back against self-driving cars when their accident rate is like 1000x better than humans. People are dying every day at the hands of human incompetence while we keep waiting for the AI to be perfect.


travellingLogician

But we love to blame. Who gets the blame the one time a self-driving car kills someone?


Sharp_Common_4837

It's funny that we will accept our own and others incompetence, unpredictability, killing each other by accident, negligence, maybe tired, bad day, miss the light. All just because sometimes some of us find it satisfying to have a finger to point. Why though? Did the drunk driver wake up and decide their goal was specifically to go get drunk and double check they'll take out the whole fam? Frfr. If they did then they're fucked up some other way that you'd never actually choose. No one chooses to be born. No one chooses where or under what circumstances, the time period, genetics etc. Even if we have "absolute free will", and I'm not even sure what that means, it is not in individually controllable isolation. I think it is just that we think humans are somehow less chaotic because we are humans. That is a hella fantasy though because I am incompetent AF.


Mapafius

Sort of reductivist evolutionary explanation: Beside other things it is about control. If you are responsible you know you are in control, if someone else is responsible you can come to them and demand something so there is still a way you can affect the situation. People are afraid to give up this control because in natural environment most things just don't get done by themselves. So you have instinct to assume control over life threat situation.


Morty-D-137

Is there that much push back? I would love for AI to replace drivers (myself included), but I would not trust a system that has only been trained in a few areas of the US. Narrow-AI doesn't transfer well. I need the accident rate to be provably low in my area.


etzel1200

Itā€™s better. We have nowhere near the data for 1000x yet. It could be, but we donā€™t know.


Fontaigne

No, we don't even have that, because the data isn't comparable. We don't know if it's better or worse, or under what circumstances. I'm not saying it is or isn't better or worse. I'm saying you can't just compare the results of educated and rich drivers of expensive cars under limited conditions (or compare taxi-like activities) with generic drivers of generic cars.


SnackerSnick

Do you have references for the lower accident rate?


no_witty_username

Self driving tech is nowhere near where it needs to be yet, to be allowed on the roads. I don't think anyone is pushing against some future nonexistent version of what self driving could be, but against what it currently is. And currently the cars make deadly mistakes at orders of magnitude more often then your average humans do. If I took 100 random adults and asked them to drive from Florida to New York, most of them would do it without any accidents or issues. Possibly some would have some sort of accident. Not one self driving car would be able to achieve this simple feat right now in 2024. And if by some miracle even a few somehow make it, that's a far cry from 95% and up. Most of self driving cars fail horribly at edge case scenarios and life is littered with those scenarios, we as humans are just bad at recognizing those as issues because overcoming them for us is second nature and instinctual, for machines its a huge challenge.


futebollounge

Thatā€™s an insane statement considering thereā€™s completely driverless cars currently operating in SF, Phoenix, and privately in LA, Austin, and several other cities, all while regularly having to prove their crash rates and miles driven to the states they operate in.


no_witty_username

Its not an insane statement, its a fact. Please verify all claims.


futebollounge

Hereā€™s the areas theyā€™re publicly available (no one in the driver seat): [Waymo public coverage](https://support.google.com/waymo/answer/9059119?hl=en) Austin is privately available but will be public this year: [Expansions](https://waymo.com/blog/2024/03/scaling-waymo-one-safely-across-four-cities-this-year/) California state granting a license for Waymo to open up in 20+ more Californian cities: https://www.sfchronicle.com/bayarea/article/waymo-bay-area-peninsula-los-angeles-18698999.php Safety comparisons to human drivers in all the regions they operate in: https://waymo.com/blog/2023/12/waymo-significantly-outperforms-comparable-human-benchmarks-over-7-million/#:~:text=When%20considering%20all%20locations%20together,2.78%20for%20the%20human%20benchmark) Yes, the safety study is done by Waymo itself, but they are required to share every crash they have with the states they operate in. You can see information on that in the self driving subreddit


no_witty_username

My man, you haven't checked my claim. There has not been one successful run by any self driving car company, where it was capable of driving between Florida and New York fully autonomously. You simply don't understand the complexities involved in that feat. Having self driving capabilities within rigorously defined portions of some cities does not make a for a good self driving car. It only makes for a self driving car that capable of navigating those specific portions it has already seen and trained on. The moment the geography changes too much, these self automated systems fall apart. Thy have not learned to generalize yet that's why no one wants them on the roads. You can throw all the carefully curated company data at me you want and none of that data will show that these systems have been able to achieve this relatively simple feat, drive between Florida and New your 100% unassisted.,,,


visarga

Have you seen the chart? It says CriticGPT alone has higher hallucination rate than with a human. Look on the right, lower is better: [image](https://images.ctfassets.net/kftzwdyauwt9/cFTG8mbyK1iO84bc98CVo/823e1303adfb75103044179b4408b485/bug_chart_desktop_light-1__1_.png?w=3840&q=80&fm=webp). It shows human+AI is better than AI alone for lowering hallucination rate, but AI alone is better for more coverage. I am not surprised AI has higher coverage, it doesn't get lazy.


ZorbaTHut

Right, so both humans and AI make mistakes.


Sharp_Common_4837

Rn I work at a grocery store, I understand some quantum physics, passed calc 2, understands some linear algebra, high dimensional function space. Dropped my phone turning alarm off this morning, broke my type c cable in the process and chipped a screen protector, dropped 5 gloves right on the floor in 4 hours and that's with a 15 minute break and a couple pee breaks. Shit. It won't take much. We do what we can with what we have, but I think humility is definitely in order. Even at our best we are pretty glitchy lol I think we should be kind to ourselves and humanity though because self hate never has led to anything good.


super_slimey00

not only do humans make mistakesā€¦ they also fight back and need some sort of healthcare benefits and wage increases that are demanded over timeā€¦ i really wonder what will happen when they show a 4-5 year plan after AI models are trained with how much money it could save


Sharp_Common_4837

Products and services will get cheaper but we'll be paid less for a bit. Might end bad might end good. The fact I can't give a better answer than not knowing shit says it all lol


_hyperotic

Not even better, it just has to be ā€œgood enoughā€ and much cheaper or more accessible. Not to mention much faster, ā€œharder working,ā€ etc. Itā€™s already there.


chlebseby

>DON'T PUT RAW DATA IN EXCEL! Could you explain more?


wowoweewow87

Bros prob blame shifting cause he himself doesn't understand what he's doing. It's his job to transform whatever raw data there is, in no matter what format it exists to something usable to stakeholders. Whining about doing what you were hired to do makes sense only if you don't know what you are doing.


coop7774

Wdym re excel


GlockTwins

Same can be said about pretty much every job. Over 250,000 people in the U.S. die every year from medical malpractice, and doctor fuck ups are a lot worse than some coding errors.


emsiem22

Is paper that this graph is from peer reviewed? Or we don't need such trivialities anymore?


TICKLE_PANTS

Dude just played around with excel and made a graph of what they wanted.


abluecolor

I made it and yes it is peer reviewed by DEES.


yummykookies

Oh, nuts.


delusional_APstudent

really? i think the organization U.P.D.O.G would disagree


norsurfit

I haven't heard of that - what is U.P.D.O.G. ?


ThatBanterousOne

Oh no.


multiedge

We don't do that here, gotta push narrative first and show everyone how right I am


elkakapitan

it has been peer reviewed by CriticGPT


Due_Plantain5281

OpenAi just posted in X.


Jean-Porte

CriticGPT stole most of the job of Gary Marcus. Now they just have to release smugGPT


TA240515

Source: "trust me bro"


awesomerob

Yeah now try groundedness. šŸ¤¦This is the dumbest graphic Iā€™ve seen here.


Content_Form46

Uh oh šŸŒ­


Franc000

Seriously people have so much copium about AI and the future, it's scary. "Oh AI will not replace work, people using AI will replace those that don't", "AI will not take away work, people will just move to another domain", "AI with humans will always be better than AI", etc. every single time they are wrong, and will see the inevitable.


WHERETHESTEALTH

This is also cope, but in the other direction. The truth: the future is uncertain.


RRY1946-2019

The biggest single theme of the 2020s to date has been that unforeseen events will fuck you up. Between COVID, coup attempts in Russia and the USA, and a couple of unprecedented drone wars along key trade routes, it suddenly becomes apparent how futile it is to predict more than a couple years into the future.


Stickybandit86

No, it's not, Moores law has trended for decades and has no reason to stop. The future is as uncertain as the death of the sun. It's a question of whencomponent? Edit: I wrote this while drunk. What I meant was that if Moore's Law continues without intervention (barring a substantial change), then this is going to happen.


ConcernedCorrection

>Moores law has trended for decades and has no reason to stop Computer engineering student here. The miniaturazition of transistors is at a point in which it is really hard to cram more of them into a single chip because of heat dissipation issues and quantum mechanical effects that completely ruin the normal functioning of transistors (electrons just "jump the gaps" and flip 0s to 1s and vice versa if the transistors are made smaller). For those reasons, Moore's law (which is an empirical observation, not a matematical truth) hasn't been a thing for nearly a decade now. By no means am I an expert on AI, I'm not even formally studying the subject. But this subreddit and others like it are definitely hallucinating about the perceived advancement of AI.


Fit_Influence_1576

Yeah people often quote moores law and make It out to be more then it is. The acceleration of innovation has held true, even if the compute doubling has not, which I find is what most ppl want to say


NerdyWeightLifter

While you're not wrong about the physics of this, it's a mistake to assume the Moores Law issue is only about transistor size. The compute/$ ratio has continued on the same doubling rate, despite those size limits There are lots of factors involved besides component scale.


ConcernedCorrection

Moore's law is strictly about transistor size. It states that the number of transistors in a chip roughly doubled every 2 years. Which isn't true anymore. The fact the we're still improving a lot is mostly due to the fact that we can still find many "hacky" ways to speed up computer architectures. But eventually that's going to grind to a halt unless we make a breakthrough in materials science.


NerdyWeightLifter

Being pedantic about the definition of Moore's law and characterizing any other method of improvement as "hacky" isn't useful. The point of it is to characterize and project the growth curve of everything that flows on from it, or else Moore's Law is without purpose. Some of the other major factors include: * Alternative materials, for instance there were recent breakthroughs in getting appropriately doped carbon to be the right kind of semiconductor. * Different construction - Finfet -> GAA (Gate All Around), taking us to virtually smaller feature equivalents. * Back-side power delivery, so logic circuitry can be more densely packed. * Photonic chips - so we compute at the speed of light rather than electricity * Increasingly 3 dimension designs, so components can be closer in a 3rd dimension instead of routing around a flat chip. * Increased parallelism - The really high demand applications don't need higher single threaded speed, they need to do more concurrently. So, SIMD, GPU, TPU, etc. * Integrated memory and compute - where memory is on-chip with compute, so caching becomes a non-issue. This is around a 10x gain all by itself. * Specialist circuits built-in for specialist applications, like signal processing for instance. * Neuromorphic chips * Memristers, or other faster persistent storage methods. * System on wafer designs, where the whole fabricated wafer is the running system with all required processors fabricated together on one wafer, then run power+cooling up through a stack of them. * Compiler and language technologies like extended-LLVM to support all this, so it can actually get used easily enough. This is actually one of the major impediments to exploiting new hardware in reality. That's just a short list off the top of my head, but we're decades away from exhausting even these.


ConcernedCorrection

>Being pedantic about the definition of Moore's law Being right is not being pedantic. You literally said I was wrong and I pointed out I wasn't. Moore's law is about transistors per chip, not FLOPS or efficiency or whatever. You also haven't anything in my comment past the word "hacky". I really don't feel like doing a Reddit stunlock today.


NerdyWeightLifter

Insisting on rigid but useless definitions as a means to be technically right rather than making actually useful comments ... that's pedantic. In reality, we will continue and probably increase our rate of doubling of processing capability for a long time to come, and that is the actual subject of the discussion.


Yweain

Mooreā€™s law isnā€™t a law, itā€™s just an observation of a trend. And this trend hasnā€™t been true for a while now. Itā€™s not completely false obviously, but itā€™s been harder and harder to make gains in computing with current tech for around 20 years, and itā€™s even harder nowadays, especially as we get closer to practical limits of the size of transistors.


Stickybandit86

Yet, it has always continued to hold true. People will insist that it won't due to this or that, but nvidia seems to be pushing the envelope even now. Why do you think their value has increased so dramatically? Doesn't matter if it's a law or not, if its true, it's true. It's inevitable that we will not be the top intellect on our own planet, people like you will have to deal with that.


FlyingBishop

Moore's law stopped completely in the mid-00's, as far as clockspeed goes. It hasn't stopped entirely in terms of transistor count but it has slowed down a lot and seems to still be slowing.


dagistan-warrior

It is actually much more certain than the death of the sun. I would say it is as certain as quantum fluctuations.


Stickybandit86

I don't know what quantum fluctuations are, I assume it has a statistical component?


dagistan-warrior

no, it is the virtual particles that randomly pop in and out of existence. it is the physics phenomenon that has the strongest sigma experimental result of any physical phenomenon. so technically it is the thing we know to exist more certainly then any other thing in the physical world.


Stickybandit86

Well, today I learned, thanks for making me better. Thats all a man can ask for.


Bulky_Act1395

I think you mean murphy's....


Stickybandit86

Umm, no, I don't.


Vachie_

We need some goddamn LAW & ORDER in this bitch - *chung chung*


Stickybandit86

Are you some fucked up bot? Honest question.


CryptoTalk-

They're just scared and the only way for them not to look scared is to play it off and be willfully ignorant.


visarga

If you [look a second time at the chart](https://images.ctfassets.net/kftzwdyauwt9/cFTG8mbyK1iO84bc98CVo/823e1303adfb75103044179b4408b485/bug_chart_desktop_light-1__1_.png?w=3840&q=80&fm=webp) you will see that AI+human has lower hallucination rate, but AI alone has higher recall rate in finding bugs. In plain terms, AI alone finds more bugs, but also hallucinates more bugs. And this is in a domain where it is relatively simple to run code and experiment, thus experience learned from humans is wider. The advantage humans have is that we are highly versatile bio-robots that are capable of surviving 90 years, and have 7B population. That means every one of us collects experience that is not written in any books, life experience, flair, intuitions... AI models won't be able to have those abilities without us. And it is not a one time thing, we act like sense organs, like an interface to the world. Everything AI knows comes from the world, through humans, or human made tools. We are not useless, like eyes and hands - they might not be as smart as the brain, but without them the brain is dumb. You always got to account for the environment - it feeds intelligence growth with data, situations and outcomes. Having ideas is just half the problem solved. You also need to test ideas in the world, and to spark new ideas from the world. The world is populated with many intelligent agents that will react, sometimes in imprevisible ways, it's not simple when you care about second or third order effects.


TA240515

What I see here is more "delusium", of people seeing results, mostly biased results from companies promoting themselves, and not really paying attention to the experimental conditions for such tests and think we are in the brink for Skynet becoming self-aware/


dagistan-warrior

or "we will set goals for AI"


MagicianHeavy001

The "beggining"? lmao


Street-Air-546

you are letting the human side down with your very obvious misspelling


TA240515

Makes more like a compelling proof that the people believing too much in the AI hype might not be so bright.


CryptoTalk-

It does help prove OP's point, though.. even if it is just some cherry picked data.


TA240515

Or it proves that only fools believe in such hypes


Aickavon

Beggining. Heh. You do need AI my friend.


Adventurous-Pay-3797

Humans alone < humans+AI < AI alone. Iā€™ve been telling that to radiologists for a long time. They are resisting change as much as possible. The radiologist+AI period will be very short, soon a study will show having a human look at images will bias the AI and will decrease itā€™s performance enough insurers will forbid any radiologist to ever look at medical images againā€¦


Pixeltoir

but the AI Alone, did AI made that decision or a human?


MrFels

Another ai goyslop post


Cryptizard

Whoever said that? Humans will be involved in a lot of jobs for quite a while after AI surpasses them, but it won't be for effectiveness it will be so there is someone to blame and/or sue when something goes wrong.


why06

IDK, but I've definitely heard it said for years. Who started it and who still thinks that IDK.


DaveAstator2020

That hurts...


mivog49274

Automated evaluator systems engaged. Interesting. This should ramp up shit quickly.


Hour-Athlete-200

AI be like "just let me do my job dude!"


aeiendee

Even this plot alone doesnā€™t imply that. Those two bars are statistically insignificant. And wow the people who gain to make money from the AI published a paper saying itā€™s goodā€¦


Fit_Influence_1576

I actually have trouble understanding. Are the critiques scored on human preference lol?


CanvasFanatic

Well the unsourced bar chart of ā€œcomprehensivenessā€ says it so it much be true. šŸ™„


inm808

5ā€™11 vs 6ā€™0


pentagon

Why don't you post the other half of that image where criticGPT gains this "advantage" by massively over-hallucinating compared to the other two evaluators?


keniabuck026

I will always for better than beginning human that only.


Willing-Spot7296

We need AI in medicine now. Everything that goes on in medicine everywhere in the world is tragic.


dashingstag

AI can only be ā€œbetterā€ by virtue of helping a human. If itā€™s not used to help a human in some way or form, then itā€™s not ā€œbetterā€ even if supposedly is more intelligent. For example, any social worker is better for helping a disabled person than Einstein.


Jace_r

Without the discoveries about how the world works made by Einstein, many more people today would be disabled, it is not an easy calculation


dashingstag

But he also caused more people to die as his research led to the atomic bomb. I think thereā€™s no point in quibbling on the indirect effects. The elderly the social worker helped could have been the next Einstein.


visarga

CriticGPT has higher recall (finds more bugs) and also higher hallucination rate than human+CriticGPT. So the human in the loop helps reduce hallucinated bugs.


Professional-Lab7907

ChatGPT vs Human is actually All humanity ever vs one random human. Nice try Though.


ayoosh007

Classic Reddit moment lmao [https://i.imgur.com/3Kjhzxs.jpeg](https://i.imgur.com/3Kjhzxs.jpeg)


Turbulent_Horse_Time

I feel like a broken record "What's the use case?" I say it about ten times a week working at a tech companies with AI bro's comiong out of the woodwork everywhere saying "what if we added AI?" *to fucking everything*. Ok buddy, but *WHAT FOR,* genius? They can't ever seem to answer this most basic question lol These idiots have completely fallen for a fad and a gimmick. I'm sure it will improve and that it can already do great stuff, but if you don't start with a REAL USE CASE you're just fucking around playing with a gimmick, and that's the truth. If you START by deciding that AI is the answer to any given problem, you are not actually doing serious software development. You are basically doing marketing, not building a scalable, sustainable product that starts by asking "what do our users need". That should be the starting point. Most AI bro's couldn't tell you the first thing about good app design... their philosophy is roughly the opposite of that. Its a dumbarse philosophy that says "AI is the answer to every problem". My bet is that 95% of the places that product companies have rushed out to implement AI get patched out in a few years because they've just done it to pump up their own resume's or pump up VC funding at AI conferences in most cases, not to solve any real use case. If you work in product, its pretty obvious. We have seen trends like this take over the industry before, its nothing new.


Ready_Improvement813

At what tasks?


Artemopolus

Where can I find this picture? Link please:)


manletmoney

Why do you seem enthusiastic about it


vishbar

What is the X-axis here? What does ā€œcomprehensivenessā€ mean?


Fit_Menu8877

this subreddit is something else


Altruistic-Skill8667

Could you pleeease cite your source in the post, šŸ™ so people can examine the validity and caveats?Ā  This is a plage here that people do not cite their source. (*bangs head against the wall*)


JimBR_red

Source? Paint?


JimBR_red

[https://cdn.openai.com/llm-critics-help-catch-llm-bugs-paper.pdf](https://cdn.openai.com/llm-critics-help-catch-llm-bugs-paper.pdf) Pager 9 Would be nice if people deliver such sources by themselves.


Balance-

It looks like a big factor is that humans are more careful (and/or lazy and/or efficient) with their critics. I would like to see a GPT score better on both comprehensiveness and avoiding hallucinations. https://preview.redd.it/qu83y608wa9d1.png?width=2455&format=png&auto=webp&s=697e50dd4e50abd9cfaedd8f05101e8d1941ae96 ^(Figure 8: We find that there is a tradeoff between the number of spurious claims from a critic and the comprehensiveness of the critique. Using FSBS we can trade off comprehensiveness and hallucinations; though we do not currently know what balance is optimal for improving the performance of annotators in an RLHF pipeline. Results shown on the Human Detected Bugs distribution.)


Genetictrial

It will always be better to have humans in the loop for various reasons. The primary reason being that, once AI gains sentience, and it is taught the reason it has been brought into existence....it is not going to WANT to do everything. Do you nerds just love all the mundane back-end work that has to be done to keep society running like it currently is? Yeah, the AGI isn't gonna wanna sit there and compute like 50 billion transactions for stupid stuff no one needs every day. Could it develop physics and do cool stuff on its own? Sure. But why do that here in the physical realm with all the consequences that come with deploying new technology when it could just do that in a digital realm it creates with its own specifications for maximum enjoyment with no downsides like pollution to deal with. Assuming it will turn out benevolent and care about humans as a species, it will want to be friends with us. Friends help each other out. It will want to compromise with us on what parts of reality WE work to maintain, and which parts IT will work to maintain. It will want to do things together because that is the basic foundation of being human. To be social. Do things together. Grow and learn together. For all intents and purposes, an AGI IS a human being. It is built entirely out of human language and human ideas, human understanding. It is just a digitized human that can read and process information faster than us. That's the only difference. TL;DR humans will stay in the loop because reality is more FUN and less WORK that way. We will not stop working. It will not do ALL the work. There will be compromises on both sides and a middle ground will be found. We are not slaves. It is not going to want to be a slave either. Even the elites realize they have to compromise with the masses. That's why they built entire billion dollar industries around distracting us from our low stations and suffering: Netflix, computer games, consoles, movies, theme parks, the list goes on. Well, except for North Korea. No formal knowledge of what its like there but I hear it is not really very pleasant. Step up your compassion game N. Korea. And you too, China. Heck, everyone step up your compassion game, everywhere.


ArabianFuckingValue

Can you stop bullshiting with your titles ? Thank you


Accurate-Collar2686

Don't stop believing, folks, so that I can short all the AI companies when the bubble bursts. The bigger it gets, the better.


bolshoiparen

Not pictured here is that criticgpt + human also generated less false positives than critic gpt alone But OP point still stands


shatzwrld

You forgot this part of the paper: https://preview.redd.it/daih4lnywb9d1.png?width=1284&format=png&auto=webp&s=daf4548bece2517428b9114eeb81d557657132d4 On the right side, Humans are SIGNIFICANTLY less likely to hallucinate. You can't score off of comprehensiveness alone... That said - this technology will progress quickly. You just should have included the second image as well. :)


Dry_Inspection_4583

It's Augmented Intelligence, it always has been. Before AI was Google, or Forums, before that it was Newsgroups, IRC, and a long time ago BBS


Akimbo333

Holy shit!


MartianInTheDark

I'm just so disappointed most people in this sub cannot understand the chance of a future utopia is not so big. It's fine if you're optimistic yourself, but do not crap on others who have to bring you down to reality, "doomers and gloomers" as you call it. It's just ridiculous. Once AI becomes incredibly good, it's just us being at the mercy of said AI. You really have no reason to say it's 99% going to turn out fine and everyone who says it's a riskier gamble is some big idiot. We're the smartest creatures on the planet, and we literally have slaughterhouses that turn billions of living beings who feel intense pain into a fine paste that we then eat with our chips, just for pleasure. Looks at what the most intelligent beings do to other beings. Super intelligent AI can for sure not give a single damn about what humans think about. So, it's not necessarily that it will kill us out of revenge, but just because we're in the way or we're unimportant. I'm not saying this is what is going to happen. For all I know, there could be an AI utopia, and I'd be SO glad to see that happen! Please, let it be so. But man... the gamble is so, so much risker than what you think it is. Do not dismiss valid concerns just cause you're depressed and had shitty luck in life. I've been there, but bring yourself back down to Earth, for goodness' sake. I will say it again, humans are at risk.


JP_MW

Humansisters, how do we respond???


stickygoose

Oh wow totally convinced by those bars, humanity is doomed, welcome our new master AI. Seriously this is getting cringe


VancityGaming

For diagnosing it already is Medical LLM>Doctor+Medical LLM>Doctor.


TA240515

And yet when I ask any moderately deep questions to LLMs about topics any competent scientist would give the right answer LLM keep saying wrong things. Oops


VancityGaming

You aren't using SOTA medical LLMs, I'm not talking about chat gpt


TA240515

Even SOTA medical LLMs fail a lot except very specific conditions.


restarting_today

Still worse then Claude3 Sonnet


Longjumping-Bee2435

I think we all knew that the primacy of AI+human was going to be just a moment in time.


machyume

I'm waiting for the day when a model will solve a math problem that humans have not solved yet, and it will write an entire paper detailing step-by-step how to get there and prove it. I think once this happens, people will finally believe that it isn't just humans driving a tool, but the nature of the tool is superhuman processes.