It depends on what they mean by AI. I can think of oodles of great uses:
An AI-powered adblock that removes all trackers, cookie confirmation popups, those annoying “please subscribe” popups, etc. would be badass. It would be virtually invisible but it would make the internet usable again.
A content filter that magically extracts the recipe you’re looking for out of the stupid blog post they write for SEO
Or to expand on that, an AI that goes through the page of search engine results and removes the ones that are SEO spam instead of actually useful content
An AI that can review at a page or email and determine if it’s a scam would save a TON of people by pointing out suspicious features.
Basically anything that requires you to copy data from one context to another is a good use of AI. You could probably have a nice resume-filling feature, for example.
But yeah, Mozilla will probably just go for a “chat with your browser” feature. Total waste of space.
The requirements to run good local LLMs have really been shrinking this past year… I have a lot of faith that there is a generally useful yet tiny AI tool within the grasp of Mozilla.
I can understand your thinking, but it could be as simple as giving the user the option to outsource the computation to a secure something or other, if their machine can’t handle it.
And yeah, the requirements are still quite high, but they are being reduced somewhat steadily, so I wouldn’t be surprised if average hardware could manage it in the long term.
Edit: For the record, Mozilla is one of the only companies I would trust if they said “the secure something or other is actually secure.” And they’d likely show actual proof and provide and explanation as to how.
It depends on what they mean by AI. I can think of oodles of great uses:
But yeah, Mozilla will probably just go for a “chat with your browser” feature. Total waste of space.
Removed by mod
The only reason this would be an issue is if it’s sending that data off to a third party. If its fully local, who cares what data it sees?
Removed by mod
The requirements to run good local LLMs have really been shrinking this past year… I have a lot of faith that there is a generally useful yet tiny AI tool within the grasp of Mozilla.
I can understand your thinking, but it could be as simple as giving the user the option to outsource the computation to a secure something or other, if their machine can’t handle it.
And yeah, the requirements are still quite high, but they are being reduced somewhat steadily, so I wouldn’t be surprised if average hardware could manage it in the long term.
Edit: For the record, Mozilla is one of the only companies I would trust if they said “the secure something or other is actually secure.” And they’d likely show actual proof and provide and explanation as to how.