it's funny how adding AI to notion actually made it a lot more usable. Most products force it on you, but here I feel like it's actually a massive benefit.
It was hard finding content and using the filters felt clunky. (And the whole UI either in a browser or their app feels buggy + slow). But with their notion AI / MCP it's gotten super easy to get information in and out.
Notion, as any other thin-AI product out there, is now in Anthropic/OpenAI/Google's crosshairs. Unless one has a moat the size of SharePoint or Google Docs or OneDrive, it's just a feature away.
I don't think they have added a Obsidian Bases / Notion Database like feature yet, right? Saw some discussion of adding a NocoDB integration, but also didn't see that happen yet.
I know this is probably out of scope, but I'd love it as well if Notion could slowly accrete the features of Airtable... at least expose some form of programmatic access to tables!
It works, but models seem to have these insane long traces to do the most basic things. I had to create a couple of skills so they know how to properly use the thing without breaking, so they don't always try to pass the wrong parameters to it.
It also doesn't let us change a couple of things (like icons). Or, if it does, not even Opus 4.6 can figure out how to do it.
From the announcement, I'm getting only a vague idea about what this is. It sounds like this Agent might run in some kind of sandbox with access to files? It would be nice if there were some documentation for the environment it runs in and what tools it has.
"If only I had an agent which does everything at work for me".
The logical continuation would be 'then I wouldn't have to work', right ?
No. Just as with coding agents, it doesn't mean less work - it means a lot more work of a different kind with the main challenge - don't loose your mind when managing the outputs from these agents.
Look at it another way - if these agents work perfectly and really increase productivity and profits and companies agentify all of their processes/development - then won't these companies essentially become extensions of OpenAI and not the other way around ?
Disclaimer: I work at OpenAI. Lots of people! I get a lot of value out of these already.
If it helps, here's an example: Our team shares wonderful customer demos in Slack. In the past, these would get lost in Slack unless someone took the time to manually create the documentation and log it in Notion.
That meant no one really logged them and the example of good work was lost.
Now, I tag workspace agent in the slack thread, it reads the thread, puts it into the right shape, and logs it (with considerably high fidelity). Saves us time, does the job no one wanted to do, and helps new hires+tenured folks like me learn from our colleagues.
(I used to use Codex to do the same thing by sharing the slack link, but now can skip that step).
Tried it to automate something that was on my to do list for the day. I had blocked off a few hours for this and managed to get the agent working reasonably well (85%) of the way there in < 15 mins.
The main remaining part is the poor docx / pdf / final output but will create a skill/workflow to get around that.
Looks like ChatGPTs answer to claude managed agents, but using existing ChatGPT Business subscription and not API Keys. With one Caveat , it needs to be invoked from ChatGPT or Slack does not support invoking from APIs, so cannot embed it. Also google launched agent cli today to build own one and integrate with Gemini enterprise https://developers.googleblog.com/agents-cli-in-agent-platfo...
This is the LLM integration approach I was pitching last year to some companies. Though in my case it was strictly tied to self-hosted inference.
Agents at the edge of business where they can work independently, asynchronously, is an approach that I don't feel was explored enough in business environments.
Sending your entire communication and documents to OpenAI would be a very bold choice.
Not only are businesses already doing that - they're not even cleaning up their source material so LLMs are generating garbage outputs from the old inconsistent trash that haunts Confluence, Google Drive, and all of the other dumping grounds for enterprise ephemera. Oftentimes "AI transformation" is just a slightly better search engine that regurgitates your old strategy (that didn't work the first time) and wraps it up in new sycophantic language that C-levels use to bulldoze the budgets and timelines of actual skilled front line employees.
I do believe that LLMs and AI provide actual value, but the "workspace" is usually the passive aggressive CYA battleground for employees to appear productive in-spite of leadership's blind-spots, ossified business practices, and "aligned" decision-making that doesn't actually fix a broken org. Maybe this release will be the one that finally challenges nepo-hires, not-invented here, and all of the other corpo crap that defines "enterprise" business.
Cleaning up source material is not easy work in companies that have massive piles of it and don't exactly know which parts of it are wrong. Quite often these documents are poorly versioned and do work for something but not exactly what you're looking for.
With this said, you can use your incorrect AI answers to find and then purge or repair this old and/or poorly written documentation and improve the output.
I agree - and I've noticed that these AI transformations tend to lay bare the many issues, inconsistencies, and other problems with workspace functions and data. Unfortunately the people that are usually in charge of these projects do not have the seniority or sway to actually change the broken processes or aren't on the right team to remove cruft. Usually you have to wait until a salesperson misquotes something from an AI summary before these issues get unblocked because they actually affected revenue.
I think this can only work based on a solid agent id system.
Shameless plug: I have been working on a solution for it, available at https://github.com/awebai/aweb and with a distributed, independently verifiable, and fully open id system at https://awid.ai
I wonder if this could be made to work with OpenAI's workspace agents.
> businesses are going to have issues down the line.
The AI agents coding up everything in parallel is just the latest iteration of a series of fads. Previously it was "outsource everything to low-cost Indian developers". Before that, it was "visual drag & drop rapid development". Before that, it was 4GL. Etc, etc...
Turns out that coding is more like a proof of understanding by the humans responsible.
I’m helping a client move data from dozens of spreadsheet to an aggregate one. The elegantish solution is to use Python, each run takes about 5s and 1 cal of energy. If i hadn’t helped her write the script, she as a non techie could have started with something like this tool, and it’ll take 90s and use 200 cal of energy. The numbers are fudged a bit, but still, how can this be profitable, or ethical? To say nothing of the spontaneous hallucination that sneaks in from time to time, especially when the model gets silently lobotomized.
Or she could have done it manually, and spent orders of magnitude more time and energy.
The scarce resource preventing more people from the ideal solution of using a script in your scenario is you. Most people can’t write a script, so their options are slow and “expensive” manual process, or the 100x as efficient AI. The 1000x as efficient script isn’t an option (well, until the model is good enough to know it should obviously just write the script too).
i actually like the concept of workspace agent, because i am feeling some real pain here to run long-term project while retaining context for each instance of agent. but based on the demo it seems more like for cooperation instead of preserving long-term project state: decisions made, actions taken, approvals given, history of what each agent did and why. it is then just a more convenient chatgpt entry in group chat.
another thing: this is all on OpenAI's servers. Which is fine if that's what you want. But there's a real class of user — technical, working on actual production code, security-conscious — for whom "my workspace lives on my machine, in my git repo, under my version control, works for my other non-openai tools" is a hard requirement, not a preference.
Without commenting on the product itself (I haven't tried it), the marketing copy around this release commits the same sins I have seen from Anthropic and Grok and all the rest of them.
I'm so tired of seeing these companies trivializing other people's work! Nobody's job is "edit files" and "respond to messages"! People have jobs like "find and close leads" and "reconcile accounts" and "arrange student field trips" and "make sure the hospital has enough inventory", not "generate reports" and "write code".
Editing files, producing reports, even writing code is just a byproduct. This is like the idiotic "lines of code produced" metric, but now they apply it to all of society.
They have to be generic because it's a generic tool. If they write "this tool can arrange student field trips", people might ignore thinking it has a narrow purpose.
Yes, work is being trivialized, but the symptom here isn't caused by that.
The issue is not that they are generic. You could still be generic with phraseology that actually acknowledged the contributions and ownership involved in the jobs being done. For example, you could write e.g. "monitor for outages", "manage projects", "arrange community events", "handle logistics", and so on.
But the problem is LLMs can't do those things. All they can do is "edit files" and "send messages".
While there definitely is a healthy dose of trivializing work I think once you scratch the surface the real messaging is that we can automate or optimize these parts of a current workflow to open work for higher value tasks to folks.
That sort of messaging has been done for decades with business process orchestration companies, RPA vendors, etc. All the way back to the original business software vendors like Lotus and Excel. It's only big LLM labs that adopt this tone of dismissive trivialization of other people's work.
I think I enjoyed OpenAI releases like ~1 year ago when they did video and presentation. This days with so many mini feature / releases is hard to be up to date or even figure out some use cases.
I just want a guarantee that OpenAI isn't just going to steal my ideas as I design my own agents. And if they did, I want compensation. I think of my custom skills, MCP servers, agents, etc, as intellectual property.
Now that's a good joke! Do you think any of the writers, artists, software developers, etc. whose work has been unwittingly used for training these models received any compensation whatsoever? If you are so concerned with IP, you should immediately stop using this technology.
They will compensate agent engineers just like they compensated all of the developers, artists, academics, and editors that created the data models are trained on.
So we’ve got about a year and a half max until we have AGI and OpenAI is launching a bunch of in house harnesses.
They must have some crazy shit cooking in the back rooms. So super duper top secret they can’t even announce it. Because if their public models are any hint, you would never think we were 18 months away from human level machine intelligence.
The hardest part is ensuring that shared context is maintained and it converges on a representation of reality and the people in the company.
[1] https://www.notion.com/help/custom-agents
It works, but models seem to have these insane long traces to do the most basic things. I had to create a couple of skills so they know how to properly use the thing without breaking, so they don't always try to pass the wrong parameters to it.
It also doesn't let us change a couple of things (like icons). Or, if it does, not even Opus 4.6 can figure out how to do it.
New knowledge additions are proposed when agents decide it would be relevant to retain, humans confirm/deny or create wiki modifications themselves.
"If only I had an agent which does everything at work for me". The logical continuation would be 'then I wouldn't have to work', right ?
No. Just as with coding agents, it doesn't mean less work - it means a lot more work of a different kind with the main challenge - don't loose your mind when managing the outputs from these agents.
Look at it another way - if these agents work perfectly and really increase productivity and profits and companies agentify all of their processes/development - then won't these companies essentially become extensions of OpenAI and not the other way around ?
If it helps, here's an example: Our team shares wonderful customer demos in Slack. In the past, these would get lost in Slack unless someone took the time to manually create the documentation and log it in Notion.
That meant no one really logged them and the example of good work was lost.
Now, I tag workspace agent in the slack thread, it reads the thread, puts it into the right shape, and logs it (with considerably high fidelity). Saves us time, does the job no one wanted to do, and helps new hires+tenured folks like me learn from our colleagues.
(I used to use Codex to do the same thing by sharing the slack link, but now can skip that step).
The main remaining part is the poor docx / pdf / final output but will create a skill/workflow to get around that.
Worked really well end-end!
Agents at the edge of business where they can work independently, asynchronously, is an approach that I don't feel was explored enough in business environments.
Sending your entire communication and documents to OpenAI would be a very bold choice.
I do believe that LLMs and AI provide actual value, but the "workspace" is usually the passive aggressive CYA battleground for employees to appear productive in-spite of leadership's blind-spots, ossified business practices, and "aligned" decision-making that doesn't actually fix a broken org. Maybe this release will be the one that finally challenges nepo-hires, not-invented here, and all of the other corpo crap that defines "enterprise" business.
With this said, you can use your incorrect AI answers to find and then purge or repair this old and/or poorly written documentation and improve the output.
- How you keep on top of what they are up to?
- How do they organize and coordinate?
I think this can only work based on a solid agent id system.
Shameless plug: I have been working on a solution for it, available at https://github.com/awebai/aweb and with a distributed, independently verifiable, and fully open id system at https://awid.ai
I wonder if this could be made to work with OpenAI's workspace agents.
How many more are thinking “am I next?”
(I built https://nelly.is as a solo founder without funding)
I'm either genuinely missing some key harness/whatever, or businesses are going to have issues down the line.
The AI agents coding up everything in parallel is just the latest iteration of a series of fads. Previously it was "outsource everything to low-cost Indian developers". Before that, it was "visual drag & drop rapid development". Before that, it was 4GL. Etc, etc...
Turns out that coding is more like a proof of understanding by the humans responsible.
You can't outsource responsibility.
The scarce resource preventing more people from the ideal solution of using a script in your scenario is you. Most people can’t write a script, so their options are slow and “expensive” manual process, or the 100x as efficient AI. The 1000x as efficient script isn’t an option (well, until the model is good enough to know it should obviously just write the script too).
another thing: this is all on OpenAI's servers. Which is fine if that's what you want. But there's a real class of user — technical, working on actual production code, security-conscious — for whom "my workspace lives on my machine, in my git repo, under my version control, works for my other non-openai tools" is a hard requirement, not a preference.
I'm so tired of seeing these companies trivializing other people's work! Nobody's job is "edit files" and "respond to messages"! People have jobs like "find and close leads" and "reconcile accounts" and "arrange student field trips" and "make sure the hospital has enough inventory", not "generate reports" and "write code".
Editing files, producing reports, even writing code is just a byproduct. This is like the idiotic "lines of code produced" metric, but now they apply it to all of society.
Yes, work is being trivialized, but the symptom here isn't caused by that.
But the problem is LLMs can't do those things. All they can do is "edit files" and "send messages".
Edit: To answer my own question "Workspace agents are available in research preview in ChatGPT Business, Enterprise, Edu, and Teachers plans."
Funny we landed on the same terminology. Will need to connect Stripe.
Zzz this is boring. So much for scaling up compute and data = intelligence.
So we’ve got about a year and a half max until we have AGI and OpenAI is launching a bunch of in house harnesses.
They must have some crazy shit cooking in the back rooms. So super duper top secret they can’t even announce it. Because if their public models are any hint, you would never think we were 18 months away from human level machine intelligence.