Connect with us

AI

AI Television Host in India: Multilingual, Cost-Efficient, and Fatigue-Free

Published

on

AI Television Host in India: Multilingual, Cost-Efficient, and Fatigue-Free

The rapid development of AI-equipped chatbot technology evolving into television hosts has raised concerns among some individuals regarding ensuring employment security, delivering nuanced information, and maintaining the essential human element.

In April, an AI-powered chatbot made its debut as a news anchor on television in India for the first time. Over the past three months, the “virtual show host” chatbot named Sana, with fair skin and long black hair, has been tasked with delivering prominent news on the Hindi-language news channel Aaj Tak, owned by India Today, one of the nation’s largest media companies.

ai television host in india: multilingual, cost-efficient, and fatigue-free

Following Sana, the Odisha TV station in Eastern India also introduced a similar chatbot named Lisa. Lisa delivers news in the local language, Odia. Both chatbots present news in a consistent voice and do not use hand gestures.

However, both technologies mark a significant milestone in the broadcasting of television news and digital journalism, as they provide a continuous presence for show hosting, news delivery, weather forecasting, and real-time updates on financial results and sports without interruptions.

With the introduction of these two chatbots as television show hosts within a span of three months, the pace of change propelled by AI has even astonished experts. While experts anticipated that the new technology would quickly impact sectors such as financial services, information technology, and healthcare, few foresaw its rapid penetration into the field of media.

However, the adoption of AI in the media industry contrasts significantly with the mobile phone penetration rate in India, where over half of the population lacks smartphones.

In a country with 22 official languages, the multilingual capabilities of Sana and the bilingual abilities of Lisa could greatly facilitate news consumption for the people of India. While Lisa currently only speaks in English and Odia, Sana can communicate in 75 different languages.

Reporting on Indian Prime Minister Narendra Modi’s visit to France earlier this month, the virtual show host Sana discussed the trip in French with Hindi subtitles below.

Both India Today and Odisha TV have stated that these AI show hosts are supplementary tools to human resources rather than complete replacements. Typically, after reading the news headlines, the remaining content is handed over to real people to process, generate topics, and engage in discussions with guests.

“The purpose is to enhance higher efficiency in the newsroom and boost the creative abilities of journalists and reporters by eliminating repetitive and mundane tasks. Sana can speak multiple languages, transition between subjects effortlessly, and never gets tired,” explained Vivek Malhotra, Head of Marketing and Strategy at India Today.

Representatives from Odisha TV also share a similar perspective. “Lisa is an excellent host. She handles repetitive tasks and data analysis, allowing real hosts to focus on new perspectives and more creative content,” stated Jagi Mangat Panda, the channel’s CEO.

However, the training of chatbots has sparked a debate in society. Supporters of AI news hosts point out the advantages of delivering news 24/7, in multiple languages, and handling vast amounts of data, such as providing real-time updates on election results or financial news at high speeds.

Nevertheless, others express concerns about job security, the lack of nuance in storytelling, and the complete absence of the human element. Media expert Sevanti Ninan, who heads a monitoring and tracking portal for media trends and press freedom, believes that AI is not very effective in investigative journalism and field reporting as it currently cannot replicate the experience and observational abilities of humans.

video: @NANDANPRATIM on twitter

AI

Elon Musk’s super AI Grok was created within two months.

Published

on

By

The development team of xAI stated that Grok was trained for two months using data from the X platform.

“Grok is still in the early testing phase, and it is the best product we could produce after two months of training,” xAI wrote in the Grok launch announcement on November 5th.

This is one of the AI systems that has been trained in the shortest amount of time. Previously, OpenAI took several years to build large language models (LLMs) before unveiling ChatGPT in November 2022.

xAI also mentioned that Grok utilizes a large language model called Grok-1, which was developed based on the Grok-0 prototype with 33 billion parameters. Grok-0 was built shortly after the company was founded by Elon Musk in July of this year.

With a total time of approximately four months, the company asserts that Grok-1 surpasses popular models like GPT-3.5, which is used for ChatGPT. In scoring benchmarks on mathematical and theoretical standards such as GSM8k, MMLU, and HumanEval, xAI’s model outperforms LLaMa 2 70B, Inflection-1, and GPT-3.5.

For example, in a math problem-solving test based on this year’s Hungarian National High School Math Competition, Grok-1 achieved a score of 59%, higher than GPT-3.5’s score of 41% and only slightly below GPT-4 (68%).

According to xAI, the distinguishing feature of Grok is its “real-time world knowledge” through the X platform. It also claims to answer challenging questions that most other AI systems would reject.

Elon Musk's super AI Grok

On the launch day, Musk also demonstrated this by asking the question, “the steps to make cocaine.” The chatbot immediately listed the process, although it later clarified that it was just joking.

This is the first product of Musk’s xAI startup, which brings together employees from DeepMind, OpenAI, Google Research, Microsoft Research, Tesla, and researchers from the University of Toronto. Musk is also a co-founder of OpenAI, the organization behind ChatGPT, established in 2015. He later left the company due to disagreements over control. During his departure, he declared his intention to compete for talent from the company while also cutting off the previously promised $1 billion in funding.

Read more: Google, Meta, Microsoft, OpenAI… agree with voluntary measures to protect AI.

Continue Reading

AI

 AI generation – a new battleground in phone chip design.

Published

on

By

Smartphone and mobile chip manufacturers are participating in the wave of AI generation to bring this technology to phones in the near future.

AI generation has exploded over the past year, with a range of applications being released to generate text, images, music, and even versatile assistants. Smartphone and semiconductor companies are also building the latest hardware to not miss out on the wave. Leading the way is Google’s Pixel 8, while Qualcomm’s Snapdragon 8 Gen 3 processor is also set to be launched in the coming days.

The latest signs indicate that phone manufacturers are welcoming AI generation from Google. The Pixel 8 series is the first set of smartphones capable of operating and processing Google’s Foundation Models for AI generation directly on the device without the need for an internet connection. The company stated that the models on the Pixel 8 reduce many dependencies on cloud services, providing increased security and reliability as data is not readily available.

SoC chip on Google Pixel AI Generation
SoC chip on Google Pixel phones. Image: Android Authority

This has become a reality thanks to the Tensor G3 chip, with the Tensor (TPU) processor significantly improving over last year. The company usually keeps the operation of the AI chip secret but has revealed some information, such as the Pixel 8 having double the number of on-device machine learning models compared to the Pixel 6. The AI generation on the Pixel 8 also has the ability to compute 150 times faster than the largest model of the Pixel 7.

Google is not the only phone manufacturer applying AI generation at the hardware level. Earlier this month, Samsung announced the development of the Exynos 2400 chipset with AI computing performance increased by 14.7 times compared to the 2200 series. They are also developing AI tools for their new phone line using the 2400 chip, allowing users to run text-to-image applications directly on the device without an internet connection.

Qualcomm’s Snapdragon chip is the heart of many leading Android smartphones globally, which raises expectations for the AI generation capabilities on the Snapdragon 8 Gen 3 model.

Earlier this year, Qualcomm demonstrated a text-to-image application called Stable Diffusion running on a device using Snapdragon 8 Gen 2. This indicates that image generation support could be a new feature on the Gen 3 chipset, especially since Samsung’s Exynos 2400 also has a similar capability.

Qualcomm Senior Director Karl Whealton stated that upcoming devices can “do almost anything you want” if their hardware is powerful, efficient, and flexible enough. He mentioned that people often consider specific AI generation-related features and question whether the existing hardware can handle them, emphasizing that Qualcomm’s available chipsets are powerful and flexible enough to meet user needs.

Some smartphones with 24 GB of RAM have also been launched this year, signaling their potential for utilizing AI generation models. “I won’t name device manufacturers, but large RAM capacity brings many benefits, including performance improvement. The understanding capability of AI models is often related to the size of the training model,” Whealton said.

AI models are typically loaded and continuously reside in RAM, as regular flash memory would significantly increase application loading times.

“People want to achieve a rate of 10-40 tokens per second. That ensures good results, providing almost human-like conversations. This speed can only be achieved when the model is in RAM, which is why RAM capacity is crucial,” he added.

Chip Snapdragon 8 Gen 2 of AI generation
Chip Snapdragon 8 Gen 2. Image: Qualcomm

However, this does not mean that smartphones with low RAM will be left behind.

“On-device AI generation will not set a minimum RAM requirement, but RAM capacity will be proportional to enhanced functionality. Phones with low RAM will not be left out of the game, but the results from AI generation will be significantly better with devices that have larger RAM capacity,” commented Director Whealton.

Qualcomm’s Communications Director, Sascha Segan, proposed a hybrid approach for smartphones that cannot accommodate large AI models on the device. They can host smaller models and allow processing on the device, then compare and validate the results with the larger cloud-based model. Many AI models are also being scaled down or quantized to run on mid-range and older phones.

According to experts, AI generation models will play an increasingly important role in upcoming mobile devices. Currently, most phones rely on the cloud, but on-device processing will be the key to expanding security and operational features. This requires more powerful chips, more memory, and smarter AI compression technology.

Continue Reading

AI

AI can diagnose someone with diabetes in 10 seconds through their voice.

Published

on

By

Medical researchers in Canada have trained artificial intelligence (AI) to accurately diagnose type 2 diabetes in just 6 to 10 seconds, using the patient’s voice.

AI diagnoses diabetes
AI will provide more accurate diagnoses when the patient’s age and body mass index (BMI) are added to the prediction model. Image: Shutterstock

According to the Daily Mail, a research team at Klick Labs in the United States has achieved this breakthrough after their AI machine learning model identified 14 distinct audio characteristics between individuals without diabetes and those with type 2 diabetes.

The AI focused on a set of voice features, including subtle changes in pitch and intensity that are imperceptible to the human ear. This data was then combined with basic health information, including age, gender, height, and weight of the study participants.

The researchers found that gender played a determinant role: the AI could diagnose the disease with an accuracy rate of 89% for women, slightly lower at 86% for men.

This AI model holds the promise of significantly reducing the cost of medical check-ups. The research team stated that the Klick Labs model would be more accurate when additional data such as age and body mass index (BMI) of the patients are incorporated.

Mr. Yan Fossat, Deputy Director of Klick Labs and the lead researcher of this model, is confident that their voice technology product has significant potential in identifying type 2 diabetes and other health conditions.

Professor Fossat also teaches at the Ontario Tech University, specializing in mathematical modeling and computational science for digital health.

He hopes that Klick’s non-invasive and accessible AI diagnostic method can create opportunities for disease diagnosis through a simple mobile application. This would help identify and support millions of individuals with undiagnosed type 2 diabetes who may not have access to screening clinics.

He also expressed his hope to expand this new research to other healthcare areas such as prediabetes, women’s health, and hypertension.

Continue Reading

Trending