ChatGPT offers chat and image processing features

chatgpt offers chat and image processing features 651712f886257 | Dang Ngoc Duy

According to OpenAI , ChatGPT can interact with voice in the style of one of five default characters. Besides, AI can also process images when users input them. The company said the chat feature is an important addition to get more people to interact and use ChatGPT .

“That’s our big challenge,” said Peter Deng, OpenAI’s Vice President of Consumer Products. “One of the hardest jobs is to use the technology we have to make it simpler to reach the next 300-400 million users.”

ChatGPT logo displayed on a smartphone model. Photo: Reuters

ChatGPT logo displayed on a smartphone model. Photo: Reuters

In the new update, ChatGPT can debate with users by voice, tell stories or speak out loud the content they enter in text. According to the test used and evaluated by the Washington Post , the responses had a “generally more conversational tone than popular virtual assistants like Google Assistant, Alexa or Siri.”

In a blog post on September 25, OpenAI emphasized that the new chat feature “opens the door to many innovative and accessibility-focused applications”.

With the image feature, users can take photos of everything around them, then ask ChatGPT to analyze the photos to “fix problems that prevent the oven from starting, see what foods are left in the refrigerator to cook.” meal plans, or complex graph analysis for work-related data”. This feature is similar to Alphabet’s Google Lens.

It is expected that the new update on ChatGPT will be available to users in the next two weeks, but only for Plus and Enterprise package subscribers.

OpenAI was founded in 2016, while ChatGPT was launched late last year and quickly went viral thanks to its ability to answer questions naturally. The new update is expected to greatly enhance the power of this AI, becoming a counterweight to traditional virtual assistants.

“The addition of voice and video capabilities moves ChatGPT further on the path to becoming a truly multimodal model: a chatbot that can ‘see’ and ‘hear’ the world, as well as respond with voice and images next to text,” Washington Post commented.

Meanwhile, AI researchers evaluate multimodal modeling as the next stage of competition in the field of artificial intelligence. They are expected to be applied to many areas of life, such as on smartphones, TVs, cars, and smart speakers.

On September 25, Amazon also said it had signed an agreement to invest four billion USD in an AI startup called Anthropic. This is the largest deal in the field of artificial intelligence since Microsoft poured more than 10 billion USD into OpenAI. Last week, Amazon also added a chatbot “chat” feature for Alexa speakers, but received limited reviews.

Bao Lam (according to Washington Post, Reuters )

48 | Dang Ngoc Duy

Leave a Reply

en_USEN