Hacker Newsnew | past | comments | ask | show | jobs | submit | TechSquidTV's commentslogin

this

The American people will be robbed blind and beaten into submission until there is a reason not to. It's that simple. They have NEVER been punished, why would they stop?


"Clawdbot, make me a website about how evil AI is. And make it snappy!"

While you absolutely should, I would argue that MCP access would be the OPTIMAL level of accessibility.

Why? What does it add that accessibility features don't cover? And of there's a delta there, why have everyone build WebMCP into their sites rather than improve accessibility specs?

Because, thinking bigger picture, having an AI assistant acting on your behalf might be more effective than slow navigation via accessibility features?

I get the wider point that if accessibility features were good enough at describing the functionality and intent then you wouldn't need a separate WebMCP.

So what does WebMCP do that accessibility doesn't?

Seems to me, at cursory reading, it's around providing a direct js interface to the web site ( as oppose to DOM forms ).

Kind of mixing an API and a human UI into one single page.


Navigation shouldn't be slow when using accessibility features though. The browser already prices the accessibility tree with full context and semantics of what is on the page and what can be interacted with.

I take the same issue when MCP servers are created for CLI tools. LLMs are very good at running Unix commands - make sure your tool has good `--help` docs and let the LLM figure it out just like a human would.


I guess I was asking - assuming that WebMCP isn't totally misguided - which of course is an assumption - is there anything that current accessibility standards can learn from WebMCP - ie why did they feel the need to create it?

I'm not aware of anything WebMCP could add that wouldn't be more useful as an improvement to accessibility tooling instead.

MCP is ultimately another solution to trying to make RPC(ish) situations more RESTful. I.e. they need self-documenting, discoverable APIs.

That's exactly what you can get from both HTML and the accessibility tree, though. We don't need another implementation for it. My guess (conjecture here) is that all the skills, MCP, WebMCP, etc talk is a manifestation of all the model providers and VCs backing them trying desperately to have others find ways to make LLMs worth the cost.


Isn't Aria there to describe the structure of the page so that say visually impaired users gain the same information as any other user? ie the interpretation of what that page then does, and so the appropriate action to take is largely left to the human user post description - just as web you load a page and look at it - the human brain works out what to do based on those visual and textual clues.

This leaves agents trying to work out page intent, allowed values for text fields - parsing returned pages for working out success or failure etc.

I'm assuming that's why they want what is effectively an in page API - that massively improves machine accessibility and can piggy back on browser authentication systems so the agent can operate on the users behalf.


The website is the API though. HTML is one of the few RESTful systems people still use today, build semantics into the page and humans and LLMs can understand how to use it.

A11y specs and APIs are just a way of presenting those semantics differently, often for those who can't see the page, whether visually impaired or in this case an LLM.

At least in my view, we should expect anything claimed to be artificial intelligence to be able to interact with things much like a human would. I'm not going to build an MCP for a CLI tool, for example, I'll just make sure it has a useful man page or `--help` command.


I think you are confusing two things.

- the semantics of a form and a button and the resulting http POST/GET - and what the page actually does!

So I can have two pages - both with html forms - what they actually do on submission might be completely different - one buys a potted plant the other submits a tax return.

ie the meaning of the action is in the non-semantic elements - the free text, the images, the context.

This is the stuff that's hard for the agent to easily determine - is this a form for submitting a tax return or not?

If what you said is true then there would be already agents out there that use ARIA info to seemlessly operate the web. As far as I can see people have tried to use that information to improve agents use of the web - but have met limited success - and that's for well annotated sites - not because sites aren't ARIA enabled.


A human needs to be able to distinguish the buttons though, both visually and via accessibility tools.

I would hope those two buttons and forms include labels, description text, indicators for required fields, etc. All of that should live in the HTML and includes attributes as needed for a11y. LLMs can use that, they don't need yet another API to describe it.


> they don't need yet another API to describe it.

WebMCP isn't accessibility support for humans, it's accessibility support for agents, which despite all the hype, are less capable than humans in working out what's going on, and find functions and data schema's easier to understand than a web page designed for human ( whether that's a partially sighted human or not ).


not from a legal perspective

Please let me plan Sim Theme Park. I can't seem to run it on Crossover on my mac.

Sim Theme Park is probably too modern and complex for the current state of RetroTick, unfortunately.

I'll admit to knowingly taking advantage of Google's pricing, but I had assumed it was within a gray area. No warning bans are insane.

Google has always done no warning bans

YouTube is also full of huge content creators, people who make Google tons of money, that complain about the Byzantine and opaque rules they have to dance around to maintain their livelihood and fan base

Google fears their giant userbases so they act with zero regard for communication and transparency because of the small chance it’d help the abusers


There was a recent gnarly version of this where some anime reactors and at least one animation channel (with something like 1.4 million subs) got demonetized and had to go through a ton of hoops to get a human to fix it.

Man. I miss this Hitman multiplayer mode. I can't believe they killed it. So quickly too. The game launched with it and it was gone within three years I think.

Ahhh damn it. Intel! Come back!

My alt Google accounts were all banned from Gemini access. Luckily Google left my main account alone. They are all cracking down.


From 3rd party AI app use?


Using a proxy to switch accounts


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: