Avatar

self

self@awful.systems
Joined
38 posts • 838 comments
Direct message

oh cool, the logo’s just a barely modified sparkle emoji so you know it’s horseshit, and it’s directly funded by Scale AI and a Rationalist thinktank so the chances the models weren’t directly trained on the problem set are vanishingly thin. this is just the FrontierMath grift with new, more dramatic, paint.

e: also, slightly different targeting — FrontierMath was looking to grift institutional dollars, I feel. this one’s designed to look good in a breathless thinkpiece about how, I dunno…

When A.I. Passes This Test, Look Out

yeah, whatever the fuck they think this means. this one’s designed to be talked about, to be brought up behind closed doors as a reason why your pay’s being cut. this is vile shit.

permalink
report
parent
reply

gonna start referring to awful.systems like how a twitch streamer refers to chat

permalink
report
parent
reply

oh boy: https://social.wake.st/@liaizon/113868769104056845 iOS devices send the contents of Signal chats to Apple Intelligence by default

e: this fortunately doesn’t seem to be accurate; excuse my haste. here’s the word from the signal forums

permalink
report
parent
reply

Chuds keep posting pictures of Democratic Party politicians (particularly Kamala Harris) with their arm raised

of course they are. there’s no convincing these fuckers because they’re collaborators looking to strengthen the conviction of other collaborators by any inane means necessary.

permalink
report
parent
reply

do you figure it’s $1000/query because the algorithms they wrote with their insider knowledge to cheat the benchmark are very expensive to run, or is it $1000/query because they’re grifters and all high mode does is use the model trained on frontiermath and allocate more resources to the query? and like any good grifter, they’re targeting whales and institutional marks who are so invested that throwing away $1000 on horseshit feels like a bargain

permalink
report
parent
reply

holy shit, that’s the excuse they’re going for? they cheated on a benchmark so hard the results are totally meaningless, sold their most expensive new models yet on the back of that cheated benchmark, further eroded the scientific process both with their cheating and by selling those models as better for scientific research… and these weird fucks want that to be fine and normal? fuck them

permalink
report
parent
reply

absolutely; there’s no reason to hide the funding source and OpenAI’s access unless you’re grifting. I feel bad for the mathematicians working on FrontierMath who didn’t know though. imagine wasting valuable time on something like this then finding out it was all just a marketing stunt devised by grifters.

permalink
report
parent
reply

Besiroglu says OpenAI did have access to many of the FrontierMath problems and solutions — but he added “we have a verbal agreement that these materials will not be used in model training.”

ooh, a verbal agreement! incredible! altman & co didn’t even have to do the typical slimy corporate move and pay an intern to barely modify the original materials into the input for the training corpus, since that verbal agreement wasn’t legally binding and behind the scenes OpenAI can just go “oopsy woopsy we swear it won’t happen again” and who’s gonna stop them?

permalink
report
reply

it should be fixed… again. for some reason our image cache keeps getting into a state where it either stops accepting uploads or stops accepting requests at all. I plan to upgrade us to the latest version soon, but it’ll unfortunately involve a little bit of downtime: to upgrade pict-rs to a new point release, you have to run the migrate command, but it only works for the previous release. we’re two releases behind, so I have to custom package the in-between release just to get us there.

permalink
report
parent
reply