🚨 ANTHROPIC JUST LAUNCHED AN AI THAT REVIEWS ITS OWN CODE FOR SECURITY... BECAUSE YOLO The features - a terminal command and GitHub scanning - are being hailed as the "big moment" for "AI-native development." In other words, the robots are finally doing their own homework. This is either the next giant leap in tech security or the single greatest plot hole in human history. The same machine that writes the vulnerabilities is now the one we trust to find them. What could possibly go wrong? Experts are "split," which in tech terms means some are busy writing glowing press releases while others are quietly moving their servers into a Faraday cage. Source: StartupNews
Mario Nawfal
Mario Nawfal7.8. klo 17.40
ANTHROPIC DISCOVERS HOW TO IMMUNIZE AI AGAINST REALLY BAD DECISIONS Anthropic claims it can now vaccinate AI against evil. Using “persona vectors” to track traits like deception or sycophancy, researchers trained language models by deliberately injecting bad behavior, then correcting it mid-training. The result: less hallucination, more control, and no loss in intelligence. Instead of scrubbing traits after the fact, this method steers models before the rot sets in. It’s early, but the tech may finally let developers curb AI drift with surgical precision before bots start sweet-talking dictators or inventing lies to win arguments. Source: techxplore
44,57K