So we’ve got about a year and a half max until we have AGI and OpenAI is launching a bunch of in house harnesses.
They must have some crazy shit cooking in the back rooms. So super duper top secret they can’t even announce it. Because if their public models are any hint, you would never think we were 18 months away from human level machine intelligence.
From the announcement, I'm getting only a vague idea about what this is. It sounds like this Agent might run in some kind of sandbox with access to files? It would be nice if there were some documentation for the environment it runs in and what tools it has.
Notion, as any other thin-AI product out there, is now in Anthropic/OpenAI/Google's crosshairs. Unless one has a moat the size of SharePoint or Google Docs or OneDrive, it's just a feature away.
I know this is probably out of scope, but I'd love it as well if Notion could slowly accrete the features of Airtable... at least expose some form of programmatic access to tables!
Tried it to automate something that was on my to do list for the day. I had blocked off a few hours for this and managed to get the agent working reasonably well (85%) of the way there in < 15 mins.
The main remaining part is the poor docx / pdf / final output but will create a skill/workflow to get around that.
This is the LLM integration approach I was pitching last year to some companies. Though in my case it was strictly tied to self-hosted inference.
Agents at the edge of business where they can work independently, asynchronously, is an approach that I don't feel was explored enough in business environments.
Sending your entire communication and documents to OpenAI would be a very bold choice.
Not only are businesses already doing that - they're not even cleaning up their source material so LLMs are generating garbage outputs from the old inconsistent trash that haunts Confluence, Google Drive, and all of the other dumping grounds for enterprise ephemera. Oftentimes "AI transformation" is just a slightly better search engine that regurgitates your old strategy (that didn't work the first time) and wraps it up in new sycophantic language that C-levels use to bulldoze the budgets and timelines of actual skilled front line employees.
I do believe that LLMs and AI provide actual value, but the "workspace" is usually the passive aggressive CYA battleground for employees to appear productive in-spite of leadership's blind-spots, ossified business practices, and "aligned" decision-making that doesn't actually fix a broken org. Maybe this release will be the one that finally challenges nepo-hires, not-invented here, and all of the other corpo crap that defines "enterprise" business.
Cleaning up source material is not easy work in companies that have massive piles of it and don't exactly know which parts of it are wrong. Quite often these documents are poorly versioned and do work for something but not exactly what you're looking for.
With this said, you can use your incorrect AI answers to find and then purge or repair this old and/or poorly written documentation and improve the output.
I agree - and I've noticed that these AI transformations tend to lay bare the many issues, inconsistencies, and other problems with workspace functions and data. Unfortunately the people that are usually in charge of these projects do not have the seniority or sway to actually change the broken processes or aren't on the right team to remove cruft. Usually you have to wait until a salesperson misquotes something from an AI summary before these issues get unblocked because they actually affected revenue.
I’m helping a client move data from dozens of spreadsheet to an aggregate one. The elegantish solution is to use Python, each run takes about 5s and 1 cal of energy. If i hadn’t helped her write the script, she as a non techie could have started with something like this tool, and it’ll take 90s and use 200 cal of energy. The numbers are fudged a bit, but still, how can this be profitable, or ethical? To say nothing of the spontaneous hallucination that sneaks in from time to time, especially when the model gets silently lobotomized.
Or she could have done it manually, and spent orders of magnitude more time and energy.
The scarce resource preventing more people from the ideal solution of using a script in your scenario is you. Most people can’t write a script, so their options are slow and “expensive” manual process, or the 100x as efficient AI. The 1000x as efficient script isn’t an option (well, until the model is good enough to know it should obviously just write the script too).
> well, until the model is good enough to know it should obviously just write the script too
(I worked on this product) Agents use the Codex runtime, so would likely solve this problem by writing a script. They also have memory, so they could save that script in the memory file system to re-use in the future. I have an agent deployed internally that has a bunch of Python scripts stored in skills/memory.
Looks like ChatGPTs answer to claude managed agents, but using existing ChatGPT Business subscription and not API Keys. With one Caveat , it needs to be invoked from ChatGPT or Slack does not support invoking from APIs, so cannot embed it. Also google launched agent cli today to build own one and integrate with Gemini enterprise https://developers.googleblog.com/agents-cli-in-agent-platfo...
I think I enjoyed OpenAI releases like ~1 year ago when they did video and presentation. This days with so many mini feature / releases is hard to be up to date or even figure out some use cases.
Without commenting on the product itself (I haven't tried it), the marketing copy around this release commits the same sins I have seen from Anthropic and Grok and all the rest of them.
I'm so tired of seeing these companies trivializing other people's work! Nobody's job is "edit files" and "respond to messages"! People have jobs like "find and close leads" and "reconcile accounts" and "arrange student field trips" and "make sure the hospital has enough inventory", not "generate reports" and "write code".
Editing files, producing reports, even writing code is just a byproduct. This is like the idiotic "lines of code produced" metric, but now they apply it to all of society.
They have to be generic because it's a generic tool. If they write "this tool can arrange student field trips", people might ignore thinking it has a narrow purpose.
Yes, work is being trivialized, but the symptom here isn't caused by that.
The issue is not that they are generic. You could still be generic with phraseology that actually acknowledged the contributions and ownership involved in the jobs being done. For example, you could write e.g. "monitor for outages", "manage projects", "arrange community events", "handle logistics", and so on.
But the problem is LLMs can't do those things. All they can do is "edit files" and "send messages".
While there definitely is a healthy dose of trivializing work I think once you scratch the surface the real messaging is that we can automate or optimize these parts of a current workflow to open work for higher value tasks to folks.
That sort of messaging has been done for decades with business process orchestration companies, RPA vendors, etc. All the way back to the original business software vendors like Lotus and Excel. It's only big LLM labs that adopt this tone of dismissive trivialization of other people's work.
I just want a guarantee that OpenAI isn't just going to steal my ideas as I design my own agents. And if they did, I want compensation. I think of my custom skills, MCP servers, agents, etc, as intellectual property.
Now that's a good joke! Do you think any of the writers, artists, software developers, etc. whose work has been unwittingly used for training these models received any compensation whatsoever? If you are so concerned with IP, you should immediately stop using this technology.
They will compensate agent engineers just like they compensated all of the developers, artists, academics, and editors that created the data models are trained on.
So we’ve got about a year and a half max until we have AGI and OpenAI is launching a bunch of in house harnesses.
They must have some crazy shit cooking in the back rooms. So super duper top secret they can’t even announce it. Because if their public models are any hint, you would never think we were 18 months away from human level machine intelligence.
reply