The settings icon ‘sprouting’ cogs is really nice!
The editor also looks really nice.
Could this not be used online as well? Persistence on the server instead of browser cache?
(Curious what your use case is for an offline browser based editor?)
The use case is privacy. Data getting harvested by free and even paid for services isn't pleasant (targeted ads, data breach etc)
If I get to add some "server" capability it will rather be webrtc, basically P2P to sync between devices, or a config to plug our own store. E.g GitHub, Google drive, dropbox or a self hosted service to SCP the files.
It isn't just browser cached, one can export individual documents or the entire store as a zipped folder. And back that up.
You do understand that one of those “needs”affects others around you, and one of them leaves them in peace, right?
Also I’m sure parent wasn’t referring to emergency calls
regardless of if this text was written by an LLM or a human, it is still slop,with a human behind it just trying to wind people up . If there is a valid point to be made , it should be made, briefly.
If the point was triggering a reply, the length and sarcasm certainly worked.
I agree brevity is always preferred. Making a good point while keeping it brief is much harder than rambling on.
But length is just a measure, quality determines if I keep reading. If a comment is too long, I won’t finish reading it. If I kept reading, it wasn’t too long.
It’s how the LLM works. Anything accessed by the agent in the folder becomes input to the model. That’s what it means for the agent to access something. Those inputs are already “Input” in the ToS sense.
That an LLM needs input tokens to produce output was understood.
That is not what the privacy policy is about. To me the policy reads Anthropic also subsequently persists (“collects”) your data. That is the point I was hoping to get clarified.
The only thing Anthropic receives is the chat session. Files only ever get sent when they are included in the session - they are never sent to Anthropic otherwise.
Note that I am talking about this product where the Claude session is running locally (remote LLM of course, but local Claude Code). They also have a "Claude Code on the Web" thing where the Claude instance is running on their server. In principle, they could be collecting and training on that data even if it never enters a session. But this product is running on your computer, and Anthropic only sees files pulled in by tool calls.
So when using Cowork on a local folder and asking it to "create a new spreadsheet with a list of expenses from a pile of screenshots", those screenshots may[*] become part of the "collected Inputs" kept by Anthropic.
[*]"may" because depending on the execution, instead of directly uploading the screenshots, a (python) script may be created that does local processing and only upload derived output
Yes, in general. I think in your specific example it is more likely to ingest the screenshots (upload to Anthropic) and use its built-in vision model to extract the relevant information. But if you had like a million screenshots, it might choose to run some Python OCR software locally instead.
In either case though, all the tool calls and output are part of the session and therefore Input. Even if it called a local OCR application to extract the info, it would probably then ingest that info to act on it (e.g. rename files). So the content is still being uploaded to Anthropic.
Note that you can opt-out of training in your profile settings. Now whether they continue to respect that into the future...
When local compute is more efficient data may remain local (e.g. when asking it to "find duplicate images" in millions of images it will likely (hopefully) just compute hashes and compare those), but complete folder contents are just as likely to be ingested (uploaded) and considered "Inputs", for which even the current Privacy Policy already explicitly says these will be "collected" (even when opting-out of allowing subsequent use for training).
To be clear: I like what Anthropic is doing, they appear more trustworthy/serious than OpenAI, but Cowork will result in millions of unsuspecting users having complete folders full of data uploaded and persisted on servers, currently, owned by Anthropic.
That is the only implementation I can think of that might make me trust a third party with confidential data.
Of course these massive transformers are already insanely computer intensive and adding FHE would make that orders of magnitude worse.
I agree but that's why it needs to be researched more. Some changes in architecture may be able to address some performance problems. It could lead to algorithmic optimizations or even specialized hardware for this.
100% on board.
This would be a paradigm shift for cloud services.
(And take away a, for many significant, source of income - data they can sell, train on, etc - So I’m afraid the incentive to research an implement it will be lacking)
The editor also looks really nice. Could this not be used online as well? Persistence on the server instead of browser cache? (Curious what your use case is for an offline browser based editor?)
reply