OpenAI has announced two new features for ChatGPT that are currently rolling out to Plus and Enterprise users. This allows the AI chatbot to respond to voice input with speech instead of text and analyze images.
ChatGPT is now able to talk and see. The well-known chatbot from OpenAI has received additional functions that customers Plus or Enterprise plans will stay Will be delivered within the next two weeks. The editorial team already has access to voice functions, but image processing is still missing.
Five votes for the ChatGPT answer
Since the launch of app versions for Android and iOS, ChatGPT has been able to transcribe mobile voice input. Now chatbots can Also responds to spoken voice input. To do this, “Voice Conversation” must be enabled in the app’s settings. Five different voices can be selected for ChatGPT replies. OpenAI has worked with professional voice actors for this purpose. Meanwhile, Spotify has announced that it will automate the feature Translating the podcast To be used.
With image processing, ChatGPT does what Microsoft’s Bing chatbot, based on GPT 4, has been able to do for a long time. Users can upload single or multiple images and analyze them. When uploading an image, it is possible to identify a specific area of the image that the chatbot should focus on. OpenAI cites examples of this function as evaluating the contents of a refrigerator for a complex graphic or recipes.
Image processing should be performed Both in app and web interface To protect function privacy, OpenAI has taken technical measures when it comes to analyzing or saying something about the people in the image. New functions will be distributed to other user groups later, which probably means the general (free) version.
formula
For reasons of better readability, masculine, feminine and different language forms (m/f/d) are not used at the same time. All personal names apply equally to all genders.