AI technology fosters robotics transformation

2024. 1. 2. 15:03
글자크기 설정 파란원을 좌우로 움직이시면 글자크기가 변경 됩니다.

이 글자크기로 변경됩니다.

(예시) 가장 빠른 뉴스가 있고 다양한 정보, 쌍방향 소통이 숨쉬는 다음뉴스를 만나보세요. 다음뉴스는 국내외 주요이슈와 실시간 속보, 문화생활 및 다양한 분야의 뉴스를 입체적으로 전달하고 있습니다.

A robot tour guide using Spot integrated with Chat GPT and other AI models. [Courtesy of Boston Dynamics]
Artificial intelligence (AI) is expanding its role in the real world beyond the domain of knowledge worker activities, such as text and image generation, and is taking over specific tasks in particular fields.

According to multiple industry sources, Boston Dynamics, a U.S. engineering and robotics design company, is experimenting with integrating ChatGPT application programming interface (API) into its robotic dog Spot, making it a tour guide by attaching a speaker to the robot and converting text into speech. Moving the gripper also gives the impression that the robot is speaking.

Matt Klingensmith, the principal software engineer at Boston Dynamics, said that the company provided a script to the robot and combined it with cameras. It is designed to perceive its surroundings and answer questions, he added.

The robots are suitable for internal tour purposes as they are already familiar with the surroundings, and visitors need only to follow them.

A research team led by Professor Shim Hyun-chul of the Korea Advanced Institute of Science and Technology (KAIST) has embarked on developing a humanoid pilot via robots without completely converting the entire aircraft into autonomous operation.

Pibot has jointed arms and fingers that move the flight control device and monitors the aircraft’s internal and external conditions via cameras during operation. It is 165 centimeters tall and weighs around 65 kilograms, made possible by combining AI with a physical robot. Pibot reads and understands flight manuals in the form of computer files, then sits in the cockpit and operates the aircraft.

The robot is designed to perform the entire process, including starting, taxiing, takeoff, landing, and cruising, in a flight simulator.

The research team’s current goal is to control actual commercial aircraft. The robot pilot can respond immediately because it is integrated with ChatGPT, a large language model (LLM), and can extract information from flight control manuals and emergency procedures as soon as an incident occurs, finding the safest flight path.

The technological competition to apply hyper-scale AI to robotics for use in the actual physical world is fierce, with attempts to move beyond AI technology confined to digital spaces as well as inducing machines to perceive and act like humans.

This trend is possible due to the rapid advancement of AI models, with the AI robot market expected to show explosive growth moving forward. According to analysis firm NextMSC, the AI robot market is projected to grow at an annual average of 32.95 percent, reaching $184.7 billion in 2030 from $95.7 billion in 2021.

The race for dominance in the field of advanced robot brains is also intense, and just as the development of in-house LLMs became active after the emergence of ChatGPT, companies have started building their own AI models for robots.

Toyota Research Institute defined its training technique for enhancing robots via AI, which was disclosed in September 2023, as a large behavior model for robots. The institute announced that its self-developed LMB would teach robots 1,000 new skills by the end of 2024.

Google also unveiled the Robotics Transformer 2 (RT-2), an AI model for robots, in July 2023. It is an improved version of RT-1, which was introduced the year before, specializing in robots that can understand commands without programming or separate training.

With RT-1, engineers had to individually program tasks like picking up and moving objects or opening drawers. On the other hand, RT-2 learns technology on its own based on visual information on the internet, using images and text.

Nvidia Corp. has introduced Eureka, an AI agent that automatically generates algorithms to train robots. Combining the natural language capabilities of OpenAI’s LLM GPT-4 with reinforcement learning, it helps robots learn complex skills.

The company claims that robots equipped with Eureka can perform around 30 tasks, such as turning a pen with fingers like a human, opening draws and cabinets, throwing and catching a ball, and using scissors.

Copyright © 매일경제 & mk.co.kr. 무단 전재, 재배포 및 AI학습 이용 금지

이 기사에 대해 어떻게 생각하시나요?