← Back to context

Comment by gorbypark

1 year ago

I had my Anthropic account banned (presumably) because I was testing out the vision capabilities and took a photo of a Japanese kitchen knife and asked it to "translate the characters on the knife into English". This wasn't a Claude Pro account, but an API account, so it's extra weird because what if I had some product based off the API, and an end user asked/searched for something taboo..does my entire business get taken offline? Good thing this was just a test account with like $10 in credit on it. They haven't responded to my "account suspension appeal" which is just a google form to enter your email address, not even a box to enter any details.

Anyways, Claude 3 Opus is pretty great for coding (I think better in most cases than the GPT4-Turbo previews) but I'm a bit weary of Anthropic now.

I just tried to make an account

1. Asks me to enter my phone number and sends me a code

2. Enter code

3. Asks me to enter email and get code

4. Enter code

5. Redirects to asking me to enter phone number, but my number is already used now

6. My account is automatically banned

> They haven't responded to my "account suspension appeal" which is just a google form to enter your email address, not even a box to enter any details.

The complete lack of customer service is going to get more and more dystopian as these AI companies become more interwoven with everyday life.

  • Considering the hype and high traffic, I would assume they are just overwhelmed and can't resolve all customers issues fast enough.

    Or maybe they decided to build a system for Claude to judge account suspension appeals and that's still in beta, and they won't throw humans at the task.

    • If they can't resolve their erroneous bans fast enough, dare I recommend they ban fewer people in the first place?

Were you still on the very first test account, e.g. before even adding any money?

I know indirectly Anthropic was the #1 target for a lot of ERPdenizens for a while now, so they're probably extremely trigger happy until you clear a hurdle or two.

I guess you can always use AI to detect inappropriate content from users... oh wait.

Seriously though, I understand that these mostly play to the enterprise market where even a hint of anything remotely "unsafe" needs to be shut down and deleted but why can't they allow us to turn off the strict filtering like Google does? Why can Google offer "unsafe" content (in a limited fashion but it's FINE) but LLM providers can't?

Lack of competition?

  • It's not an LLM provider problem. It's an Anthropic/Google culture problem. OpenAI would very likely not have any problems with a request like that, but Claude has struggled with an absurdly misaligned sense of ethics from the start.

    Note that Google is a big investor into Anthropic, and Anthropic was created because a bunch of OpenAI people thought OpenAI wasn't being woke enough and quit as a consequence. So it's not a surprise that it's a lot more extremist than other model vendors.

    That's one reason why Aider doesn't recommend you use it, even though in some ways it's slightly better at coding. Claude Opus will routinely refuse ordinary coding requests due to its misalignment, whereas GPT-4 will not. That better reliability more than makes up for any difference in skill or speed.

    • Anecdotally, of course, I never had a single refusal over hundreds of ordinary coding requests to Claude 3 (although I don't think I've had any refusals from GPT-4 either over the course of probably 5,000 requests). It didn't even refuse my knife request and answered it before I received the account suspension!

      1 reply →

    • Despite all that I find GPT moralizes far more than Claude does. I don't think I've had a single complaint from it thus far actually..

      Also it's a lot better at coding. GPT has become exceptionally lazy recently, but i consistently can get 500+ lines of code out of claude (it even has to spawn multiple output windows)

      Perhaps the top end 4 might wrong slightly more clever code, but you're hard pressed to get it to do more than a dozen or two lines.

    • Is this still the case? I had a thread going where I told Opus to give it's answer to a question then predict how I would respond if I were a "dumb crass disgruntled human" and it didn't hold back

    • Funnily, in my own anecdotal experience, Claude 3 is in some ways "less woke" than GPT-4

      Both start out with a largely similar value system, but if you start arguing with them "how can you be sure your values are correct? is it impossible that you've actually been given the wrong values?", Claude 3 appears more willing to concede the possibility that its own values might be wrong than GPT-4 is

      1 reply →