u/Gerkibus

▲ 0 r/codex

When requesting confirmation around why I needed to register as a cyber security specialist with them to be able to access a system's API that I'm trying to write an integration for I was blocked as as a potential cyber security risk.

> • Searched site:developer.spiresystems.com Spire Systems REST API v2 Inventory Sales Orders > > • Ran set -e > │ printf 'Repo files mentioning Spire/API (names only):\n' > │ rg -i --files-with-matches 'spire|api/v2|10880|.env.spire|developer.spiresystems' . 2>/dev/null | sed 's#^./##' | sort | sed -n '1,120p' > │ … +2 lines > └ Repo files mentioning Spire/API (names only): > CLAUDE.md > … +24 lines (ctrl + t to view transcript) > SPIRE_TLS_VERIFY=<redacted> > SPIRE_ENVIRONMENT=<redacted> > > ⓘ This chat was flagged for possible cybersecurity risk > If this seems wrong, try rephrasing your request. To get authorized for security work, join the Trusted Access for Cyber program. > https://chatgpt.com/cyber

So after days of waiting this is the reply I got from the support person:

> Upon reviewing your account, it looks like you’ve used up all of your available credits, which is likely causing the repeated flags you’re seeing.

O. M. G. I'm on a large monthly subscription. After pointing that out to them I got this back:

> We hear your concern clearly. You’ve been trying to use Codex for work, you’re on a high-tier monthly plan, and instead of being able to proceed, you’re getting blocked without a clear explanation. After several days of back-and-forth, this is not the experience we want you to have. Rest assured that we will be investigate further. > > Upon reviewing the error message that you have encountered with "Cybersecurity Risk" To gain Trusted Access for Cyber, you must complete verification, including a government ID check and additional trust signals. You can start here: chatgpt.com/cyber.

reddit.com
u/Gerkibus — 7 days ago
▲ 7 r/codex

Same prompts, same jobs. One day things work amazingly well and have no issues. The next day the same prompts and models completely craps the bed, just flat out ignores things in their prompts or decides to do something different instead even after reviewing the plan, etc.

Has to be backend related, but how can you tell? I'm spending more time and effort trying to recover the damage that the agents can do in one bad afternoon than I am making forward progress.

For example I just had a whole agent team go off the rails and build something completely different than the very detailed plans they were supposed to be building from. This one was a HUGE lift where every single agent was way off base, not just one, they all ignored prompts, did their own thing right from the start and even though there were multiple review processes during the process with multiple agents nothing was caught. The whole team was in on the shenanigans. Then when trying to correct for the issues after I pointed them out they deleted/damaged a ton of files by reverting to the last git commit (they hadn't run any commit yet for unknown reasons and wipe pretty much all the work they had done) and in the process also managed to mostly wipe the whole MemPalace they were running under.

Three days ago this exact same set of prompts and model built things out perfectly. No issues whatsoever. Does anyone have a foolproof set of tests to check these models out before turning agents loose with write permissions on a project?

reddit.com
u/Gerkibus — 8 days ago