Even as Meta’s Llama 3 rapidly gains popularity as one of the most-used large language models (LLMs), OpenAI, the pioneer of generative AI, is enhancing its competitive edge by introducing new enterprise-grade features for its GPT-4 Turbo LLM and other models.
OpenAI announced an expansion of enterprise-focused features for its API customers, enriching its Assistants API and introducing new tools that improve security, administrative control, and cost management.
Olivier Godement, Head of Product for API at OpenAI, stated in a recent media interview, “When talking to developers and businesses about meaningful work for AI models, OpenAI still has the lead. We always welcome competition—it drives improvement for everyone.”
Private Link and Enhanced Security Features
A key security enhancement is the introduction of Private Link, which facilitates secure communication between Microsoft’s Azure cloud service and OpenAI. This reduces customer data exposure to the open internet in API queries.
OpenAI's security framework also includes SOC 2 Type II certification, single sign-on (SSO), AES-256 data encryption at rest, TLS 1.2 encryption in transit, and role-based access controls. Additionally, OpenAI has implemented native Multi-Factor Authentication (MFA) to strengthen access controls amid rising compliance requirements.
For healthcare organizations needing HIPAA compliance, OpenAI offers Business Associate Agreements along with a zero data retention policy for qualifying API clients.
Upgraded Assistants API for Enhanced File Management
One of OpenAI’s significant enterprise offerings is its Assistants API, now upgraded to enhance file retrieval capabilities with the new ‘file_search’ feature, capable of managing up to 10,000 files per assistant—a 500X increase from the previous limit of just 20 files. This upgrade includes functionalities like parallel queries, improved reranking, and query rewriting.
OpenAI has also introduced streaming capabilities for real-time conversational responses, allowing models such as GPT-4 Turbo to generate outputs as tokens are created, rather than waiting for complete responses. The API now includes new ‘vector_store’ objects for improved file management and offers detailed control over token usage, aiding cost management.
Projects Feature for Tailored Access Control
The newly introduced Projects feature allows organizations to manage roles and API keys at a project level. This enables enterprise customers to set permissions, control model availability, and establish usage limits to avoid unexpected costs, simplifying project management.
For instance, different teams within an enterprise can work with specific AI models on separate projects without any risk of data co-mingling. Miqdad Jaffer from OpenAI explained, “Projects allow you to sequester your resources and members into individualized projects, providing distinct reporting and enhanced control over access, security, and costs.”
Cost Management Features for Scalable AI Operations
OpenAI is also introducing new cost management features to help organizations scale AI operations efficiently. These include discounted rates for consistent token usage and a 50% cost reduction for asynchronous workloads via the new Batch API, which promises faster response times and higher rate limits.
To use the Batch API, customers must submit their tokens in a single request, with results typically delivered within 24 hours, though OpenAI states responses can be as quick as 10-20 minutes for those who don’t require immediate feedback.
As OpenAI continues to strengthen its offerings with a focus on enterprise-grade security, administrative control, and cost management, these updates aim to provide a more seamless experience for businesses, directly countering the rise of Llama 3 and open models from competitors like Mistral, which may require additional setup effort for enterprises.