{"id":172801,"date":"2023-09-26T00:22:44","date_gmt":"2023-09-26T05:22:44","guid":{"rendered":"https:\/\/lifeboat.com\/blog\/2023\/09\/chatgpt-can-now-speak-listen-and-process-images-openai-says"},"modified":"2023-09-26T00:22:44","modified_gmt":"2023-09-26T05:22:44","slug":"chatgpt-can-now-speak-listen-and-process-images-openai-says","status":"publish","type":"post","link":"https:\/\/lifeboat.com\/blog\/2023\/09\/chatgpt-can-now-speak-listen-and-process-images-openai-says","title":{"rendered":"ChatGPT can now \u2018speak,\u2019 listen and process images, OpenAI says"},"content":{"rendered":"<p><\/p>\n<p><iframe style=\"display: block; margin: 0 auto; width: 100%; aspect-ratio: 4\/3; object-fit: contain;\" src=\"https:\/\/www.youtube.com\/embed\/tmFFd8fMqxk?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; encrypted-media; gyroscope;\n   picture-in-picture\" allowfullscreen><\/iframe><\/p>\n<p>OpenAI\u2019s ChatGPT can now \u201csee, hear and speak,\u201d or, at least, understand spoken words, respond with a synthetic voice and process images, the company <a href=\"https:\/\/openai.com\/blog\/chatgpt-can-now-see-hear-and-speak\" target=\"_blank\">announced<\/a> Monday.<\/p>\n<p>The update to the chatbot \u2014 OpenAI\u2019s biggest since the <a href=\"https:\/\/www.cnbc.com\/2023\/03\/14\/openai-announces-gpt-4-says-beats-90percent-of-humans-on-sat.html\">introduction of GPT-4<\/a> \u2014 allows users to opt into voice conversations on ChatGPT\u2019s mobile app and choose from five different synthetic voices for the bot to respond with. Users will also be able to share images with ChatGPT and highlight areas of focus or analysis (think: \u201cWhat kinds of clouds are these?\u201d).<\/p>\n<p>The changes will be rolling out to paying users in the next two weeks, OpenAI said. While voice functionality will be limited to the iOS and Android apps, the image processing capabilities will be available on all platforms.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>OpenAI\u2019s ChatGPT can now \u201csee, hear and speak,\u201d or, at least, understand spoken words, respond with a synthetic voice and process images, the company announced Monday. The update to the chatbot \u2014 OpenAI\u2019s biggest since the introduction of GPT-4 \u2014 allows users to opt into voice conversations on ChatGPT\u2019s mobile app and choose from five [\u2026]<\/p>\n","protected":false},"author":359,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[6],"tags":[],"class_list":["post-172801","post","type-post","status-publish","format-standard","hentry","category-robotics-ai"],"_links":{"self":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/172801","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/users\/359"}],"replies":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/comments?post=172801"}],"version-history":[{"count":0,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/172801\/revisions"}],"wp:attachment":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/media?parent=172801"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/categories?post=172801"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/tags?post=172801"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}