National Yang Ming Chiao Tung University

10/09/2024 | Press release | Distributed by Public on 10/09/2024 02:44

AI as the Catalyst for the Fourth Industrial Revolution – Insights from Professor Tien-Fu Chen, Vice Dean of the College of Computer Science

By CIO Taiwan
Translated by Chance Lai

______
AI technology, hailed as the driving force behind the Fourth Industrial Revolution, is reshaping human life globally. OpenAI's release of ChatGPT at the end of 2022 sparked a revolution in AI, introducing a tool that can answer questions, translate, summarize, and even code. Its plus versions can generate images and videos, marking the dawn of the AI 2.0 era. This breakthrough has ignited worldwide interest among businesses in generative AI (GenAI), pushing AI computing power to become synonymous with national competitiveness.
According to the "2022-2023 Global Computing Power Index" report by IDC, the global AI computing power market is projected to grow from $19.5 billion in 2022 to $34.66 billion by 2026. The generative AI market is set to expand from $820 million in 2022 to $10.99 billion by 2026, increasing its share of the AI computing market from 4.2% to 31.7%. This surge is poised to drive innovation across manufacturing, finance, education, and healthcare industries.
Professor Tien-Fu Chen, Vice Dean of the College of Computer Science (CCS) at National Yang Ming Chiao Tung University (NYCU), notes that governments worldwide are racing to build national AI computing centers to support industry and research. The AI supercomputer "Taiwania 2," developed by the National Center for High-Performance Computing (NCHC) and local enterprises, serves this purpose in Taiwan. However, despite rapid expansion, AI projects still face significant challenges due to the dominance of NVIDIA in the GPU market, where high costs and limited chip supply have led to a global chip shortage.

NCHC Teams Up with AI National Team to Build Energy-Efficient AI Cloud Data Centers (Photo credit: NVIDIA )

As companies like AMD, Intel, Meta, Microsoft, and Google develop their own AI hardware architectures to optimize efficiency and cut costs, the AI chip shortage is expected to ease, reducing the cost of building AI platforms.

The Three Pillars of AI: Data, Algorithms, and Computing Power

Since ChatGPT's remarkable debut, companies have sought to integrate GenAI into their operations to boost efficiency, improve product quality, and reduce operational costs. MediaTek's GenAI platform, MediaTek DaVinci, exemplifies this trend. Initially designed for internal use to enhance productivity, it has since evolved into a platform for external businesses, drawing dozens of companies from various sectors such as technology, finance, telecommunications, and education into its ecosystem.

As competition intensifies, adopting GenAI is no longer optional but essential. MediaTek's DaVinci platform has already seen a 96% penetration rate within the company, with 88% of employees acknowledging its productivity boost in technical documentation, analysis, and software development.

For companies that hesitate, the risk of falling behind is real. Companies must ensure access to vast and varied datasets, advanced AI algorithms, and sufficient computing power to launch AI projects effectively. Without these, the benefits of AI may take over a year to materialize, lagging behind the fast-paced industry.

Photo credit: MediaTek

Conversely, companies who ignore this technology and maintain a conservative, wait-and-see attitude risk being eliminated from the market. However, initiating a generative AI project within a company means building a large language model (LLM) to support the computing demands of the project. Before doing so, the company must have the three key elements: data, algorithms, and computing power.

Professor Tien-Fu Chen pointed out that, like any AI project, a company must have sufficient data, including a large and diverse dataset, to meet AI training and inference requirements. Additionally, the company needs an appropriate and advanced AI algorithm to make the most of this valuable data. Finally, it must have an AI computing platform capable of supporting the project; otherwise, the initial AI training and inference could take several months, and the project's benefits might not be realized for over a year, which is already out of step with current industry trends.

Generative AI: A Must for Enhancing Productivity

First, companies must prioritize data preparation. However, many companies need to use effective methods during the data organization process, resulting in data that is not entirely usable. When data is incomplete, it often affects the outcomes of AI model training. Therefore, before starting AI model training, data cleaning is essential to ensure the overall quality of the data.

That said, if this complex task relies solely on manual processes, it can be time-consuming and labor-intensive. Fortunately, GenAI technology can handle raw documents, and numerous open-source data-cleaning tools are available. These AI-powered data cleaning tools save significant time and resources and help maintain data quality, making them indispensable tools for companies pushing AI projects forward.

Image generated by ChatGPT

Professor Tien-Fu Chen stated that open-source tools are now moving towards a No-Code development approach, significantly reducing the time and difficulty in building GenAI solutions. This greatly enhances the quality of AI projects. However, knowledge extraction still relies on the expertise and experience of senior employees to improve the quality and effectiveness of GenAI projects. He suggests that companies form cross-departmental teams, with senior employees sharing their industry knowledge and experience to assist AI engineers in data cleaning tasks, which can significantly contribute to the success of AI projects.

For companies that have previously implemented Business Intelligence (BI) projects, their experience can be extended to GenAI initiatives, offering substantial benefits in improving business operations. A company's knowledge classification quality is closely linked to the effectiveness of AI models. Well-classified data from previous efforts not only shortens the time and cost of implementing GenAI projects but also helps establish a corporate knowledge base and enhance employee productivity in the long run.


Leveraging Open-Source Large Language Models to Accelerate Project Timelines

In the early stages, companies typically focused their AI projects on applications such as image recognition and defect detection, aiming to reduce errors, improve product quality, and lower operational costs. However, today's GenAI projects are becoming key drivers for enhancing efficiency and powering future business growth.

Large enterprises see this technology as critical in boosting their market competitiveness. However, many need to pay more attention to the fact that training LLMs is highly complex, time-consuming, and costly. Most large companies begin by training their AI models, but completing the entire process-including inference, testing, and implementation-can take 1-2 years before yielding results.

The demand for computing power during the AI model training process is immense, requiring significant financial investment that only resource-rich companies can afford. "When developing GenAI models, companies should invest in training and inference separately, rather than attempting to do everything in-house, as it is extremely time-consuming and often fails to deliver the expected results," explains Professor Chen. "With the explosion of GenAI, many pre-trained AI models are now available on the market. Companies can leverage these models, combine them with internal data, and perform inferences on smaller-scale AI platforms, shortening AI project timelines. Most medium to large enterprises also have sufficient budgets to support this approach."

Currently, widely used open-source large language models include BLOOM, LLaMA3, and Mistral, which have already been broadly adopted. These models come in various parameter sizes, allowing companies to choose models based on their specific use cases. It is only sometimes necessary to use models with 175B parameters or more. Instead, companies can select models with appropriate parameters for fine-tuning or combine multiple smaller models using a mixture-of-experts (MOE) approach. This strategy reduces infrastructure costs and shortens AI model inference times.

Leveraging External AI Experts to Minimize Learning Errors and Accelerate Progress

Although open-source LLMs are emerging rapidly, companies can still establish their internal platforms for large language models. Recognizing the high level of interest in AI projects, businesses often need help with challenges such as a shortage of AI talent and limited resources. To address these issues, public cloud providers offer a variety of large language models for users to choose from.

Numerous information service companies on the market also provide one-stop solutions, including platform setup and consulting services. These solutions benefit small and medium-sized enterprises with limited budgets and human resources, allowing them to implement AI projects affordably and leverage AI technology to enhance productivity and competitiveness.

Professor Tien-Fu Chen pointed out that compared to the early stages of AI technology development, the emergence of ChatGPT, with its remarkable ability to respond fluidly and provide vast knowledge across various fields, has rapidly increased interest in GenAI among businesses. Companies now see GenAI as a way to strengthen their operational capabilities. However, as previously mentioned, the success of AI projects depends on the effective integration of three key factors: data, AI algorithms, and computing power. Ensuring that GenAI delivers the expected benefits relies heavily on how well these elements are coordinated, which could pose a significant challenge for businesses.

Considering market trends and companies' future development, many information service providers now offer one-stop AI project solutions. Companies can take advantage of external resources by starting with a consultancy approach using professional AI teams to implement projects on a small scale or in non-core departments. The most significant advantage of collaborating with external AI experts is that it allows companies to avoid substantial upfront investments before fully understanding GenAI technology, enabling more accurate assessments of potential benefits.

Additionally, most external experts have extensive experience in successful AI implementations, allowing businesses to gain insights into the finer details of project execution and identify areas of improvement, which can be highly beneficial for future AI initiatives.

Training Employees as AI Experts: A Challenge Companies Must Address

According to a report published by IDC, global spending on AI reached $432.8 billion in 2022, covering software, hardware, and AI systems. By 2026, global AI spending is expected to grow to $300 billion, with a compound annual growth rate (CAGR) of 26.5%, demonstrating the rapidly increasing demand for AI applications globally.

In this environment, the world is experiencing a surge in demand for AI talent. To address the challenge of recruiting skilled external talent, many companies are partnering with external research institutions to train their employees to be AI experts.

One of the most notable examples comes from several large electronic product manufacturers that have successfully used AI technology to drive digital transformation. To meet the demands of this transformation, these companies began collaborating with the Taiwan AI Academy in 2018, offering AI training programs for their employees with a focus on research and supply chain optimization.

This substantial investment has brought significant operational benefits. For example, some companies have adopted the concept of "machine doctors," using AI for visual inspections and "listening" diagnostics to assess the health of equipment and its components.

In intelligent healthcare, the accumulation of big data is the foundation of everything. Only with a rich and high-quality database can AI be endowed with predictive and decision-making abilities, ultimately achieving the effectiveness of precision medicine. (Image generated by ChatGPT)

Professor Tien-Fu Chen believes that the talent cultivation strategies adopted by leading intelligent manufacturing companies can be valuable lessons for other businesses, as they are part of an essential process. When companies recruit AI talent externally, they often face challenges such as a lack of industry knowledge and familiarity with the company's unique characteristics. These employees may need time to adapt and learn, and there is always the risk of being poached by other companies. Therefore, businesses must invest in training their employees to become AI experts.

Of course, some companies are concerned about new employees gaining access to sensitive information. However, this can be addressed by implementing a practical data access policy that balances security with the company's future growth.

It is undeniable that implementing GenAI projects helps improve work efficiency. Chen suggests that business leaders should not expect drastic changes overnight but focus on achieving incremental improvements of 10% to 20%. By making gradual changes, companies can find the most suitable transformation strategy. NYCU's myLLM industry-academia alliance platform offers a great starting point for businesses. Companies can quickly establish their LLMs and then evaluate whether to further invest in developing their own GenAI solutions to achieve more excellent digital transformation results and benefits.

Please refer to the original text (Mandarin) for details.