Skip links
OpenAI DevDay 2025 event recap, featuring the conference name and year, summarising ChatGPT platform announcements including Apps SDK, AgentKit, Codex, GPT-5 Pro, and Sora 2 for digital marketing agencies

OpenAI DevDay 2025 Event Recap: For Digital Marketing

OpenAI DevDay 2025 recap covering ChatGPT Apps SDK, AgentKit, and new AI models announced at the San Francisco developer conference

OpenAI’s third annual DevDay at Fort Mason in San Francisco on 6 October felt different from previous years. Less research showcase, more commercial proposition. The company unveiled tools that transform ChatGPT from a conversational interface into something closer to a platform for building modern applications. Sam Altman opened with numbers that are difficult to ignore: ChatGPT now serves 800 million people weekly, up from 100 million just two years ago. Four million developers are building on the platform, which has doubled since 2023. The API processes 6 billion tokens per minute, a thirtyfold increase from 200 million in 2023.

For digital marketing teams, the event signalled that AI integration is moving from experimental to operational faster than many agencies have prepared for.

ChatGPT Becomes a Distribution Channel

The Apps SDK was the headline announcement, and it genuinely changes the landscape for customer interactions. Third-party developers can now build full applications that run directly inside ChatGPT conversations. This creates a fundamentally new distribution channel. Rather than driving users to websites or mobile apps, brands can deliver interactive experiences within the conversational flow where 800 million people already spend their time.

Think about what this means in practice. A travel brand displays bookable properties with pricing and availability directly in chat. An e-commerce retailer renders product catalogues with add-to-basket functionality. A media company serves personalised content recommendations with inline playback. Launch partners, including Booking.com, Canva, Expedia, Figma, Spotify, Zillow, and Coursera, demonstrate this is more than theoretical. These are live integrations handling real transactions.

The technical side builds on Anthropic’s Model Context Protocol, which OpenAI has extended to support component rendering with HTML. This allows developers to design both logic and interface in a single framework whilst maintaining cross-platform compatibility. From a business perspective, the value proposition is straightforward: develop once, deploy to an audience the size of Instagram’s early peak. The apps went live on 6 October to Free, Go, Plus, and Pro users outside the EU.

For agencies, this opens several service opportunities. Brands will need a strategy for conversational commerce, implementation expertise for app development, and optimisation capabilities to drive engagement within the ChatGPT ecosystem. The technical requirements include HTTPS endpoints with Server-Sent Events or Streamable HTTP transport, OAuth 2.1 authentication, and dynamic client registration. These are established web development standards, which means most technical teams can implement Apps SDK integrations without requiring entirely new skill sets.

The security considerations deserve attention, though. When you’re building apps that connect to multiple data sources simultaneously, the attack surface expands considerably. Clear data handling policies, regular security audits, and compliance with GDPR or industry-specific requirements become non-negotiable.

Nick Turley, OpenAI’s head of ChatGPT, described the trajectory: over the next six months, ChatGPT will progress “from an app that is really useful into something that feels a little bit more like an operating system.” That’s the pitch, anyway. Whether it actually becomes the AI-first platform OpenAI envisions remains to be seen, but the infrastructure is now in place for developers to find out.

Figma Integration Moves Markets

Among the initial ChatGPT apps, Figma’s integration stood out enough to send its stock soaring. Sam Altman highlighted it during the keynote, and investors responded by driving Figma’s share price up as much as 15 percent intraday, closing 7 percent higher. So what does the integration actually do?

Essentially, it links ChatGPT with Figma’s collaborative design tools, allowing users to generate diagrams and visuals from a conversation. You can convert ChatGPT conversations directly into FigJam diagrams (Figma’s online whiteboard product). Brainstorming text becomes flowcharts or wireframes in real time. Conversely, ChatGPT can also suggest using Figma when a visual aid might help, seamlessly handing off part of the task to Figma’s interface and then continuing the dialogue.

This tight coupling of a design platform with an AI chatbot hints at how creative work may become more conversational and assisted by AI. A product manager could sketch out an idea in text and have ChatGPT–Figma translate it into a draft design, which can then be refined collaboratively. For non-designers, this lowers the barrier to creative planning considerably. The boundary between discussing an idea and designing it is disappearing.

The stock market reaction tells you something about the perceived value of early platform integration. By being a launch partner, Figma gains credibility and a foothold in OpenAI’s app ecosystem, potentially expanding its reach beyond traditional design teams into broader brainstorming and ideation contexts. The move comes as its competitor Adobe pushes its own AI features, so integrating with ChatGPT helps Figma maintain competitive positioning.

For marketing teams, this suggests a case study worth studying. AI platforms can drive significant user engagement to services that integrate early, creating new growth channels. The AI handles the generative heavy lifting or tedious setup, whilst humans focus on higher-level direction and refinement.

AgentKit: Building AI Agents Gets Easier

AgentKit tackles something most agencies have already discovered: building AI agents is still more difficult than it should be. The toolkit bundles four components designed to reduce development time and improve reliability.

Agent Builder provides a visual drag-and-drop canvas for multi-agent workflows. This allows developers to orchestrate complex sequences of AI actions with a clear visual overview, enabling faster iteration and collaboration across teams. Until now, creating an AI agent required stitching together various components (prompts, model calls, external tool APIs, custom logic). AgentKit streamlines this considerably.

ChatKit offers an embeddable chat UI with custom branding capabilities. If you build an AI agent with OpenAI’s models, ChatKit provides the front-end components to integrate a ChatGPT-like chat experience into your product, skinned to your brand and needs.

The Connector Registry centralises APIs and data sources with administrative controls. This makes it easier to plug an agent into the software and data it needs, whilst maintaining security and admin control. It’s a hub to manage integrations with services like Dropbox, Google Drive, Slack, internal databases, and third-party MCP-based tools.

Enhanced Evals for Agents adds step-by-step trace grading, datasets, and automated prompt optimisation. Developers can measure and improve an agent’s performance with reinforcement learning fine-tuning options.

The live demonstration at DevDay made the value proposition clear. An OpenAI engineer built a complete DevDay assistant with two functional agents in under eight minutes. Whilst live demonstrations are always optimised for impact, the early adopter metrics suggest these efficiency gains actually translate to production environments. Canva integrated a support agent with ChatKit in under an hour, saving two weeks of engineering work. Klarna’s support agent now handles two-thirds of all customer service tickets. One organisation reported a 50 percent reduction in development time and a 30 percent accuracy improvement using the new Evals features.

Box, Canva, and Evernote are already using AgentKit internally. For marketing applications, several use cases become immediately more viable.

Customer service automation becomes practical when you can build branded chatbots that handle common enquiries whilst maintaining brand voice and escalating complex queries to human agents. The ability to integrate with existing CRM systems, ticketing platforms, and knowledge bases creates a unified customer experience rather than yet another disconnected tool.

Lead qualification is another area where this makes sense. Deploy agents that engage with prospects, ask qualifying questions, and route high-intent leads to sales teams. The Connector Registry allows secure integration with marketing automation platforms, ensuring lead data flows into existing workflows rather than creating manual export-import processes.

Content workflows could benefit as well, though this is where agencies will need to be most careful about quality control. You can create agents that assist with content research, generate first drafts, optimise for SEO, and maintain brand guidelines. Multi-agent systems can coordinate research, writing, editing, and approval processes within a single workflow. Whether that actually saves time or just creates a different kind of overhead remains to be seen in practice.

Campaign management is perhaps the most immediately practical application. Develop agents that monitor campaign performance, identify anomalies, generate reports, and suggest optimisations based on historical data. Integration with Google Analytics, Meta Ads, and other platforms provides a lot of related data. This is the kind of repetitive analysis work that genuinely benefits from automation.

AgentKit carries no additional fees beyond standard API pricing, making it accessible for agencies and clients of various sizes. The Connector Registry is rolling out in beta to API users, ChatGPT Enterprise, and Education customers, providing IT teams with secure configuration settings, output review capabilities, and cloud environment management through the Global Admin Console. For agencies managing multiple client accounts, these administrative controls are essential for maintaining security and compliance standards.

Codex: The AI Coding Assistant Goes Mainstream

Codex, OpenAI’s software engineering agent, reached general availability with features that suggest the company is serious about enterprise adoption. Sam Altman noted that “almost all new code written at OpenAI today is written by Codex users,” a striking testament to how effective AI-assisted coding has become internally.

The agent operates on three models: codex-1 (optimised from o3), GPT-5-Codex (with dynamic “thinking time” adjustment for complex problems), and codex-mini-latest (based on o4-mini at £1.20 per million input tokens). It writes features, fixes bugs, runs tests, and proposes pull requests, all within isolated cloud sandboxes.

The development timeline from research preview in May to production readiness by October is notably fast. The generally available version includes Slack integration, allowing teams to assign tasks directly from conversations. There’s a TypeScript SDK for integration into CI/CD pipelines. Administrative tools provide usage dashboards and workspace management for organisations juggling multiple projects.

For marketing agencies, Codex addresses several operational bottlenecks that technical teams will recognise immediately. Website and landing page development could be accelerated considerably. Internal adoption metrics at OpenAI show 70 percent more pull requests merged per week, which suggests productivity gains could be achieved in reality. Companies like Cisco, Duolingo, and Instacart are early adopters of the advanced coding AI. Cisco reduced code review times by up to 50 percent. Virgin Atlantic’s development team reports that Codex generates clean updates from simple pull request comments.

Marketing technology integration is another practical application. Building and maintaining integrations between marketing platforms, CRM systems, analytics tools, and data warehouses is the kind of work that’s essential but rarely exciting. The ability to generate clean code from natural language descriptions reduces the technical barrier for marketing technologists who understand what needs connecting but don’t necessarily want to spend hours writing API integrations.

Automated reporting benefits as well. Developing custom reporting dashboards and data pipelines that consolidate metrics across multiple platforms is straightforward in concept but often tedious to implement. Codex can write scripts that extract, transform, and load data from APIs, maintaining these integrations as platforms inevitably change their specifications.

Technical SEO implementation is perhaps less obvious but potentially useful. Generating structured data markup, creating XML sitemaps, implementing canonical tags, and building tools for custom technical SEO audits all require attention to web standards. The model’s understanding of current best practices means you’re less likely to end up with deprecated markup or non-compliant implementations.

The model served over 40 trillion tokens in its first three weeks after GPT-5-Codex launch, which points to real production usage rather than experimental testing. From 20 October, Codex cloud tasks will count towards usage limits. This marks the shift from preview to metered commercial service, which is fair enough but requires agencies to do proper capacity planning and budget allocation. The efficiency gains often justify the costs, particularly for teams managing multiple client projects with recurring technical requirements. But “often” doesn’t mean “always”, and testing with real workloads before committing to production usage continues to make sense.

Four New Models That Actually Matter

OpenAI introduced four models optimised for different use cases and price points, each with specific applications for marketing operations.

GPT-5 Pro: Enterprise Reasoning

GPT-5 Pro targets enterprise workloads requiring high precision and deep reasoning. Priced at £12 per million input tokens and £96 per million output tokens, it represents noticeable cost savings compared to the previous o1-pro at £120 and £480. For marketing applications, GPT-5 Pro excels at complex strategic analysis, competitive research synthesis, and sophisticated content planning that requires nuanced understanding of brand positioning and market dynamics. Whether it actually outperforms o3-pro as claimed remains to be tested in practice, but the pricing alone makes it worth evaluating for high-stakes work.

Sora 2: Video Generation Goes Mainstream

Sora 2 is where things get interesting from a content production perspective. The model produces up to 90 seconds of 4K video with synchronised dialogue and sound effects, featuring physically accurate motion and persistent world state across multi-shot sequences. It supports detailed camera direction and handles realistic, cinematic, and anime styles. OpenAI describes it as “more physically accurate, realistic, and more controllable than prior systems.”

The marketing applications are fairly straightforward. Generate short-form video content for Instagram Reels, TikTok, and YouTube Shorts without relying on stock videos or traditional video production resources. Create product demonstrations and feature explainer videos from text descriptions, reducing both time and cost whilst enabling rapid iteration based on performance data. Transform campaign concepts and storyboards into video mockups for client presentations, allowing stakeholders to evaluate creative directions before committing to full production budgets.

There’s also a “Cameos” feature that allows insertion of specific individuals into generated scenes following one-time identity verification. The potential for personalised video in account-based marketing campaigns is obvious, though there will be some trial and error around what actually feels compelling versus what crosses into uncanny valley territory.

Mattel’s partnership demonstrates commercial viability, with designers using Sora 2 to transform sketches into toy concepts. The API preview enables programmatic access, supporting workflow automation and integration with existing content management systems. Worth noting that Sora 2 went live the same day as the announcement, which is unusual for OpenAI and suggests they’re confident in its production readiness.

This opens the door for applications like dynamic video content creation, game development, or marketing, where an app can generate bespoke video clips on demand. Imagine generating concept commercials or visualising toy designs via AI video. For creativity, individuals and small businesses could produce quality videos without big budgets. But it also intensifies concerns around deepfakes and misinformation if not used responsibly. OpenAI appears aware of this, rolling out features in Sora (the consumer app) for more user control and content moderation.

gpt-4o-mini-realtime-preview: Voice Commerce Gets Viable

The voice model, gpt-4o-mini-realtime-preview (commonly called gpt-realtime-mini), is a native speech-to-speech model that processes audio directly without text conversion. It costs 70 percent less than advanced voice models but still manages to maintain expressiveness. Pricing sits at roughly 6p per minute for input and 24p per minute for output.

Marketing applications include voice commerce, where customers can make purchases and resolve issues through natural voice interactions. Voice-activated content that responds dynamically to listener questions. Customer service deployments that handle routine enquiries with natural conversation flow. Another potential use case would be accessibility features that ensure marketing content reaches audiences with visual impairments or reading difficulties.

The model supports WebRTC and WebSocket connections, includes two exclusive voices (Cedar and Marin), and integrates Session Initiation Protocol for phone calling. The latency improvements over traditional multi-step pipelines create noticeably more natural conversational experiences, which matters for customer-facing applications.

gpt-image-1-mini: Scaled Visual Production

gpt-image-1-mini is priced at 80 percent less than the large gpt-image-1, at £6.40 per million tokens. High-quality images cost roughly 2.6 to 3.8p per image. Medium quality runs approximately 1.2p, low quality around half a penny. Generation times sit at approximately 20 seconds.

For marketing teams managing substantial image requirements across social media, display advertising, email campaigns, and web content, the cost reduction enables scaled production of visual assets. The ability to generate variations quickly supports A/B testing of creative concepts. The transparent background option facilitates integration into various design contexts. Agencies can use gpt-image-1-mini for concepting and iteration, reserving the more expensive gpt-image-1 for final production assets requiring higher quality. That tiered approach to quality versus cost is probably how most teams will end up using these models in practice.

Infrastructure Investment: The AMD Partnership

OpenAI announced a strategic partnership with AMD to deploy 6 gigawatts of AMD Instinct GPUs, with AMD receiving a warrant for up to 160 million shares. That’s approximately 10 percent of the company, which gives you some sense of the scale of infrastructure investment required. For context, that’s an enormous amount of computing power on the order of the output of several large power plants, highlighting the voracious compute needs of advanced AI models.

The arrangement is financially hefty: AMD expects the partnership to generate tens of billions of pounds in annual revenue, and over £80 billion across four years. News of this alliance sent AMD’s stock rocketing over 34 percent in one day (its biggest jump in nine years), adding about £64 billion to AMD’s market value.

This partnership ties OpenAI to one of Nvidia’s chief rivals in the AI chip space. Nvidia currently dominates AI hardware (and had itself agreed to invest in OpenAI recently), but OpenAI’s massive commitment to AMD’s forthcoming MI series GPUs indicates a desire to diversify its hardware base. In effect, OpenAI is ensuring it has enough cutting-edge chips to power future models. By securing a stake in AMD, it could benefit if AMD’s AI market share grows.

For agencies relying on OpenAI’s platform for client deliverables, improved infrastructure stability reduces the operational risk of API unavailability. One of the limiting factors for AI adoption has been the scarcity and expense of GPU compute. By locking in 6GW of compute, OpenAI is gearing up to train larger models (think GPT-6 and beyond) and serve more complex queries, which could translate to better AI capabilities available via its API and ChatGPT. It might also reduce the risk of service capacity crunches.

Performance Tiers That Actually Matter

The priority processing announcement matters for anyone running production systems. GPT-5 API requests run 40 percent faster on the priority processing tier compared to the standard tier. There’s a new Service Health Dashboard that provides real-time monitoring of uptime, request time, token velocity, and time to first token. Priority processing delivers SLA-backed, predictably low latency even during peak demand, with Enterprise access and premium per-token pricing. The alternative “flex” tier offers 50 percent cheaper processing with increased latency for o3, o4-mini, and gpt-5 models.

For agencies managing client campaigns with time-sensitive requirements, the performance tiers enable appropriate service level selection. Standard tier suffices for batch processing and non-urgent tasks. Priority tier ensures reliable performance for real-time applications like chatbots, voice interactions, and live customer service. Flex tier provides cost savings for development environments and testing workflows where latency tolerance is higher. In practice, most agencies will probably use a combination of all three tiers depending on specific workload requirements.

The Service Health Dashboard is one of those features that seems minor until you actually need it. When API performance degrades, being able to quickly identify whether issues stem from OpenAI’s infrastructure or internal systems reduces diagnostic time considerably. For agencies with service level agreements, this visibility supports more accurate performance reporting.

Market response to DevDay announcements has been positive, at least in the short term. HubSpot’s share price increased nearly 7 percent following mentions of its AgentKit implementation. Coursera’s stock rose over 6 percent during ChatGPT integration demonstrations.

What This Actually Means for Marketing Agencies

The announcements at DevDay 2025 suggest several considerations for marketing agencies evaluating AI integration, though it’s worth approaching this with some healthy scepticism alongside the enthusiasm.

OpenAI’s evolution from a research organisation to a platform provider does reduce adoption risk. With 800 million users, 4 million developers, and established enterprise customers, the platform demonstrates commercial viability beyond experimental implementations. That said, platform maturity doesn’t automatically translate to marketing ROI. The technology exists, but a successful application still requires strategy, implementation expertise, and realistic expectations about what AI can and cannot do effectively.

The Apps SDK, AgentKit, and Codex create several marketing service opportunities. Clients will require strategy development for conversational commerce, technical implementation of custom applications, and ongoing optimisation as the platform evolves. Agencies that develop expertise early gain a competitive advantage in an emerging market. However, it’s also worth remembering that we’ve been here before with other platforms that promised revolution and rather just delivered evolution. Early adopter advantage exists, but so does early adopter risk.

Whilst new models offer significant price reductions compared to predecessors, production usage at scale can be more expensive than initial estimates suggest. Developer feedback indicates that multimodal implementations can be “three times more expensive than anticipated, despite pricing cuts.” Testing with limited traffic before full deployment is essential for avoiding surprises. Monitor token consumption patterns closely and adjust implementation accordingly.

Security and compliance become more complex as AI agents access multiple data sources simultaneously. For anyone managing client data across CRM systems, marketing platforms, and analytics tools, this includes clear data handling policies, regular security audits, and compliance with GDPR or industry-specific requirements.

OpenAI competes directly with Anthropic’s Claude, Google’s AI tools, and Microsoft’s GitHub Copilot. The Apps SDK’s foundation on Anthropic’s Model Context Protocol suggests increasing interoperability, which potentially reduces platform lock-in risk. That’s encouraging, though cross-platform compatibility in practice often proves more complicated than in theory.

An Ecosystem That Evolves

OpenAI’s DevDay 2025 announcements collectively paint a picture of a company expanding its reach on all fronts. It’s not just launching a new model here or a feature there, but building an entire ecosystem. On the software side, turning ChatGPT into an app platform and launching AgentKit means OpenAI wants to be the foundation upon which AI-powered applications are built and delivered, much like iOS for mobile apps or Windows for desktop software.

This creates opportunities for developers (new channels, tools, and revenue streams) whilst giving users a more integrated AI experience. On the model side, OpenAI is pushing the envelope in both capabilities (GPT-5 for coding, Sora 2 for video) and accessibility (making these available in products and APIs, not just research). On the hardware side, it’s securing the raw power needed to fuel those ambitions at scale.

For consumers, these developments promise AI that is more useful, ubiquitous, and integrated into the apps we already use (and perhaps some new ones). Ask your chatbot to book travel, sketch a UI, or even generate a short video, and it can oblige by tapping specialised services behind the scenes.

For the tech industry at large, OpenAI’s moves signal both consolidation and competition. The company is solidifying its role as a platform player, not just an AI model provider. This could concentrate AI activity around OpenAI (and by extension Microsoft, its major backer and partner), raising questions about ecosystem lock-in and control, much as big tech platforms have in the past. At the same time, the partnership with AMD and support for open standards like MCP show that OpenAI is willing to collaborate and shape industry standards, possibly to avoid bottlenecks and regulatory scrutiny.

Practical Next Steps

For agencies considering OpenAI’s platform following DevDay, a measured approach makes sense. Start by identifying specific client challenges that align with the announced capabilities. Customer service automation, content production workflows, and technical marketing operations represent practical starting points with measurable ROI. Avoid the temptation to find problems that fit the solution. Focus on genuine operational bottlenecks that these tools might address more effectively than current approaches.

Begin with contained pilot projects that deliver value quickly whilst limiting exposure. A ChatGPT app for a single client, an AgentKit-powered customer service bot, or Codex integration for landing page development provides learning opportunities without betting the entire agency strategy on unproven implementations. Learn what works in practice rather than what sounds impressive in theory.

Establish a clear understanding of token consumption patterns for intended use cases before committing to production deployment. Test with representative data volumes. Ensure budget allocations reflect actual usage rather than theoretical estimates. The gap between expected and actual costs can be substantial, particularly for implementations that process large volumes of content or maintain persistent conversational context.

Conduct a thorough assessment of data handling requirements, particularly for implementations that connect multiple client systems. Establish clear policies for data access, storage, and transmission that comply with relevant regulations and client security standards. Document these policies clearly and review them regularly as implementations evolve.

Establish metrics for measuring implementation success beyond anecdotal evidence. This includes technical performance (response time, error rates, availability), business outcomes (cost savings, efficiency gains, customer satisfaction), and comparative analysis against previous solutions. Without baseline measurements, determining whether implementations actually deliver value becomes largely subjective.

The DevDay announcements represent a substantial evolution in OpenAI’s commercial offering. For marketing agencies, the strategic question is not whether to integrate AI capabilities but how to do so effectively while managing costs, security, and complexity. Several new tools now exist for production deployment. The challenge lies in applying them to deliver measurable client value.

The shift from research showcase to production platform is complete. What happens next depends on whether agencies can move from experimentation to operational integration without losing sight of what actually drives results for clients. The technology is ready. The question is whether marketing teams are.

Leave a comment

Explore
Drag