The PlaceOS Podcast

Why OpenAI Hasn't Released Multi-modal


Listen Later

  • OpenAI recently introduced a beta version of voice control, gaining traction through YouTubers who showcased how the AI interacts using voice and visuals.
  • This voice control allows users to combine audio commands with camera input, enabling the AI to process what's happening in the environment and respond accordingly.
  • However, it’s not yet available to the public, sparking curiosity about why such an advanced feature remains limited to demos.
  • One major reason is OpenAI's challenge in ensuring the AI doesn't produce harmful content, such as dangerous instructions for illegal activities.
  • Additionally, the processing power needed to support such features is immense, raising concerns about costs and operational overhead.
  • While some speculate OpenAI could offer it as a premium paid service, this option hasn't materialized, leaving potential users wondering about the rollout strategy.
  • Beyond technical constraints, there's intrigue around the financial model—what level of paid subscriptions would break even on the computational demands?
  • OpenAI may be profiting from basic features, but the high cost of developing and training more advanced AI models keeps them in the red.
  • Meanwhile, competitors like Meta's LLaMA 3 and Anthropic's Claude are pushing forward, and in some cases, rivaling OpenAI's performance in specific tasks.
  • As consumers await these tools to be integrated into everyday products, the anticipation builds, particularly for applications in meeting rooms and smart buildings.
...more
View all episodesView all episodes
Download on the App Store

The PlaceOS PodcastBy PlaceOS

  • 5
  • 5
  • 5
  • 5
  • 5

5

1 ratings