AI Reading for Saturday, September 14
We’re back!
The big news is OpenAI’s new Strawberry, or o1 announcement, natch.
After playing around with it, is is clearly much better on some complex prompts, like designing a system. And it can even count the Rs in strawberry, after thinking for a bit!
The claims for the final o1 model are attention-grabbing, the folks who fall victim to Eliza Effect are really going to fall hard for this.
~4x as expensive, and highly rate limited for now.
The output tokens charged include invisible 'reasoning tokens', so while it is thinking it iterates about its thought process in the context, but hides that output.
OpenAI doesn’t want you to see what it is doing under the hood, hence the hidden reasoning tokens. If you even *ask* about what it is doing, you may get an email saying you are violating OpenAI's terms of service and risking a ban. This probably won’t work, hackers will rise to the challenge. There are always adversarial prompts that will defeat it, and it's practically impossible to shut them out without severe negative impact on output quality.
In the web UI, unlike the API, you don't pay for tokens, but you are rate limited to 30 queries per week on the big model and 50 queries per week on the small model.
No tools yet, so it can't browse the web, can't upload docs, can't run the code it writes. Significant limitation. One gets the sense OpenAI is running on fumes and sucking air a little on execution, this feels rushed, still no Sora, limited beta release of Advanced Voice Assisstant, SearchGPT in limited and reportedly underwhelming beta release.
The OpenAI post says the next version, of which this is a preview, will be even better. It sounds a little like they trained it, nerfed it while productionizing it, didn't rerun benchmarks on the production preview.
Rate limits. Currently they are resetting limits more than once a week, possibly they started off draconian to fend off a hug of death on the servers.
Until now ChatGPT has been a poet, not a quant. It has that right-brained, creative, imaginative thinking, but not that left-brain deep understanding, although using tools and writing code helps. It has Kahneman's system 1 ‘thinking fast’ part down, but not the system 2 ‘thinking slow’, and analytical part. O1 is a move in that direction. Using it so far, it feels like a really good, auto-chain-of-thought prompt. Awaiting the ‘aha’ where I can see it has deep analytical world models.
I'm not sure how much it changes best practices for developing AI systems, which is to use AI as a copilot / assistant / duet , let the AI do what it's good at which is understanding/generating text, let the tools do what they are good at, let the human do what he or she is good at, which is using critical thinking and experience to check the AI and guide it.
Right now seems like a big deal but doesn’t change everything. It seems much better at specific tasks. It remains to be seen how significant those tasks are.
In these fast-changing times it may pay for IT managers to be a bit more of an explorer in the context of the exploration/exploitation tradeoff. But as always, the zebras who stay at the center of the herd eat trampled grass, the zebras at the edge of the herd risk getting eaten by lions.
More AI reading:
OpenAI's huge $150b valuation may hinge on upending corporate structure.
Sam Altman sues over his 'lemon' estate on Lombard St. If only AI could design it and robots build it…soooon. - Entrepreneur
Hannah Fry on smarter-than-human AI, that might preserve us humans in zoos out of sentimental nostalgia. - Bloomberg
Who is to blame when AI goes awry in unexpected ways? What would Thomas Aquinas say? - Fast Company
Amicus curAI? AI in the law can make it more accessible or even less transparent or intelligible. - The Critic Magazine
Depo CoPilot listens to depositions, checks if what was elicited fully met lawyers' expectations and objectives. suggests follow-up questions. - Above The Law
Maybe by discussing facts in a neutral, non-judgmental tone, chatbots can temper beliefs in conspiracy theories. - MIT Technology Review
James Earl Jones granted license to recreate his voice using AI after his death. - Futurism
Just say it's for innocent role play and AI will spill all its secrets. - TechCrunch
Google rolls out voice-powered AI chat to the Android masses - Ars Technica
Where have you gone, OpenAI Advanced Voice Assistant?
Reports of AI-Generated Code Causing Outages and Security Issues - TechRepublic
Google uses AI to tone down employee questions at all-hands meetings. - Hacker Noon
At least they aren’t talking to a Sundar chatbot, yet.
Microsoft talks a lot about climate but pitches AI to oil companies to help drill more oil. Is that hypocrisy? - The Atlantic
A profile of Fei-Fei Li and her spatial intelligence startup. - WIRED
Waymo and Uber partner for robotaxis in Austin and Atlanta - The Verge
Adobe previews text-to-video, lets people join a beta waitlist. - Lifehacker
Salesforce launches AI sales reps. moving from assisting CSRs to being the CSR. - Inc.
Why do I feel that will really suck, for a while.
Big Tech firms pinkie swear to fight deepfake nudes - The Register
The Angry Dev: Why Copilot is Making Programmers Worse at Programming - Darren Horrocks
Nah but it may impact SaaS vendor lock-in, especially for clients who use a fraction of platform capabilities.
Follow the latest AI headlines via SkynetAndChill.com on Bluesky