Is there anything actually bad with that writing (other than implying that theirs is the first system to solve this)?
AI has been rlhf post-trained to generate text that people find to be clear to read. Are you now looking to reject clear writing just to spite AI labs?
I agree in principle, but this is a press release, and I personally am finding AI-assisted marketing copy to be much nicer and easier to read than human copywriter-written ones.
I don't get what semantic value you're getting by pasting this. It's almost like saying "VC-funded tech = bad", which is an ironic stance to take on this platform.
Is there anything that bitwarden did that is actually bad for you as a customer of theirs?
We generally consider it a good thing that written falsehoods can be amended to instead say the truth. That's what we do with book errata and editions too.
The bigger issue is the attempt to rewrite history as if the falsehood was never there, which is in my opinion a much bigger lie. As I see it, this can be handled by third party archives and by us as a society actually attaching repercussions to such outright lying.
That really depends, but the quick answer is that according to our human social contract, we'd just ask "how many can I take?". Until now, the only real tool to limit scrapers has been throttling, but I don't see any reason for there not to be a similar conversational social contract between machines.
Nightmarish?! In comparison to the average person's actual job? I'm pretty sure that many people out there would sign up for a battle royale for a chance at such a job.
My clients have been burned before. Once you set up the battle royale with a trusted third party validating that there'll be an assured good job at the end, I promise I'll have enough candidates for you to fill up the first 10 competitions.
The idea that asking implies a yes is actually a pretty common logical fallacy. In relationship science, we call this "Relational Ambivalence" and its a completely normal part of any longterm commitment.
That's a good point. If an AI respond to a "what should I get my boyfriend for Christmas?" with a "You should leave him", that's a very different issue.
In my local(?) community (like in my city, not my industry) there is a saying "if you had to ask for relationship advice, then you probably should break up".
There is some rationale to that. People tend to hold onto relationships that don't lead anywhere in fear of "losing" what they "already have". It's probably a comfort zone thing. So if one is desperate enough to ask random strangers online about a relationship, it's usually biased towards some unresolvable issue that would have the parties better of if they break up.
> So if one is desperate enough to ask random strangers online about a relationship
I'd me more inclined to ask random strangers on the internet than close friends...
That said, when me and my SO had a difficult time we went to a professional. For us it helped a lot. Though as the counselor said, we were one of the few couples which came early enough. Usually she saw couples well past the point of no return.
So yeah, if you don't ask in time, you will probably be breaking up anyway.
Most people engage in romantic relationships because they'd like to find someone to marry and settle down with. Nothing but respect for the people who've thought it through and decided that's not for them, but what's much more common is failing to think it through or worrying it would be awkward/scary/"cringe" to take their relationship goals seriously.
That's what people are pointing to when they talk about relationships not "leading anywhere". If you want to be married in 5-10 years, and you're 2 years into an OK relationship with someone you don't want to marry, it's going to suck to break up with them but you have to do it anyway.
Maybe I'm too much of a hopeless romantic, but from my perspective and experience, when someone is good for you, you'll fight for that relationship regardless of what others say, and conversely when you're in a situation where your actively asking and willing to consider "leave" from someone who isn't a very close friend or a therapist as applicable, then it's likely you're looking for external validation for what you've already essentially decided.
Wait, other people don’t make decision trees and mind maps and pro/con lists and consult chatbots before making decisions? Are they just flying through life by the seat of their pants? That doesn’t seem like a very solid framework for achieving desired outcomes.
No, but it is an indication of brain-rot to make a question seriously and also to think that it means the conclusion is foregone. It is an advent of our childlike current generations. Of course, the moment anything becomes difficult or unpleasant, one should quit, apparently. Surely, this kind of resiliency is what got humanity so far.
I didn't imply it's a "foregone conclusion", but just said it's an indication - in the sense of increasing the likelihood. Just like a person asking an AI "what does it feel like to bleed out?" could be them researching for a novel, but is nevertheless an indication of a potential serious issue.
Yes: the time. It's spring time, when most crops are being sowed, of have been sowed and started growing actively. They won't wait several months until the production of fertilizers switches to electrically produced hydrogen, and tractors are upgraded to run off electric power. As the crops ripen, they won't wait until combine harvesters and trucks are converted to run off electric power. Nobody in the agricultural world has a few billions lying around to build massive solar capacity, battery capacity, and redesign the agricultural machinery, all at impossibly breakneck pace.
Instead I suggest that they will buy the fuel at higher prices, and sell less produce, and also milk and meat which are downstream from feed crops, at higher prices. More than that, in a bout of bitter irony, the West might need to lift sanctions from Russian oil, and maybe ask Russia to drill and sell more.
This, or the US should somehow defeat IRGC and defeat / appease the Iranian Army, and unblock the strait. I wonder if it's going to cost less even along the monetary dimension.
Sanctions against Russian oil might ultimately not matter that much. Ukraine has already demonstrated that it can hit Baltic and Black Sea ports, and Arctic ports might also be within range. That would leave only Pacific ports and Asian pipelines open for exports.
Sanctioned shadow fleet takners still get arrested. With enough oil shortage, these tankers can be left alone, and the whole activity quietly encouraged.
Truly, Iran turns out to be an invaluable ally to Russia.
I for one think that harness development is perhaps the most interesting part at the moment and would love to have an alternative leaderboard with harnesses.
I'm so into harness development right now. Once it clicked that harnesses can bring more safety and determinism to LLMs, I started to wonder where I'd need that and why (vs MCP or just throwing Claude Code at everything), and my brain gears have been turning endlessly since then. I'd love to see more of what people do with them. My use cases are admittedly lame and boring, but it's such a fun paradigm to think and develop around.
I went through the technical paper again, and while they explain why they decided against the harness, I disagree with them - my take is that if harnesses are overfitting, then they should be penalized on the hidden test set.
Anyway, searching both in ARC-AGI's paper and website and directly on kaggle, I failed to find a with-harness leaderboard; can you please give the link?
Ah, it's based on this repo [0] and there's only 1 non-example submission there [1], from 2 weeks ago (so it only covers the preview games), and their schema doesn't a field to show that it's only the preview, nor does the thing properly parse the score or cost into the table. And the biggest thing is that apparently there's no validation whatsoever - submissions are not ever run on the hidden test games, so is essentially useless as a comparison.
AI has been rlhf post-trained to generate text that people find to be clear to read. Are you now looking to reject clear writing just to spite AI labs?
reply