OpenAI planning to develop audio AI hardware
Published January 2, 2026
OpenAI plans to announce a new audio language model in the first quarter of 2026, a step toward building an audio-based physical hardware device, according to a report in The Information.
Citing sources familiar with the plans, The Information reported that OpenAI has combined multiple engineering, product, and research teams into one initiative to improve its audio models. Researchers at the company reportedly believe these models lag behind text-based ones in accuracy and speed.
The report noted that few ChatGPT users opt for the voice interface, with most preferring text. The company hopes better audio models could shift user behavior, enabling deployment in a wider range of devices like cars.
OpenAI plans to release a family of physical devices in the coming years, starting with an audio-focused one expected to ship in about a year. Internal discussions have covered forms like smart speakers and glasses, with an emphasis on audio over screen-based interfaces.
Competitors including Google, Meta, and Amazon have also increased R&D efforts on voice-and-audio interfaces, such as Meta’s push into smart glasses. This follows a previous boom in voice assistants like Alexa, Google Assistant, and Siri, which were popular with casual tech consumers but had significant limitations.
Some AI product developers, including former Apple design lead Jony Ive, have stated they believe voice-controlled products are less addictive than screen-based ones, though they have not generally cited reliable evidence.



