<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
    <channel>
        <title>Home on Bullbull AI</title>
        <link>https://digitalxber.com/</link>
        <description>Recent content in Home on Bullbull AI</description>
        <generator>Hugo -- gohugo.io</generator>
        <language>en-us</language>
        <lastBuildDate>Sat, 18 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://digitalxber.com/index.xml" rel="self" type="application/rss+xml" /><item>
            <title>Future Development of Robots and AI: Insights from Experts</title>
            <link>https://digitalxber.com/posts/note-4e97a2f5a6/</link>
            <pubDate>Sat, 18 Apr 2026 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-4e97a2f5a6/</guid>
            <description>&lt;h2 id=&#34;future-development-of-robots-and-ai&#34;&gt;Future Development of Robots and AI&#xA;&lt;/h2&gt;&lt;p&gt;On April 17, 2026, a main event themed &amp;ldquo;Intelligent Future&amp;rdquo; was held at the Chinese Academy of Sciences in Beijing, focusing on the future of robotics and artificial intelligence (AI). Experts gathered to share insights on the advancements and applications of these technologies.&lt;/p&gt;&#xA;&lt;h3 id=&#34;robots-evolving-into-new-intelligent-entities&#34;&gt;Robots Evolving into New Intelligent Entities&#xA;&lt;/h3&gt;&lt;p&gt;Academician Yu Haibin, director of the Industrial Artificial Intelligence Research Institute of the Chinese Academy of Sciences, delivered a report titled &amp;ldquo;Robots Leading a New Era of Technology.&amp;rdquo; He explained the development trajectory, basic components, and classifications of robotic technology, and provided a comprehensive analysis of the applications and future directions of robots across various fields.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 3&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;356px&#34; data-flex-grow=&#34;148&#34; height=&#34;471&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-4e97a2f5a6/img-5388c65f41.jpeg&#34; width=&#34;700&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;Yu pointed out that with the deep integration of AI and intelligent manufacturing, robots are gradually breaking the boundaries of traditional automation equipment, transitioning into new intelligent entities with perception, cognition, and autonomous decision-making capabilities.&lt;/p&gt;&#xA;&lt;h3 id=&#34;ai-transitioning-to-future-partners&#34;&gt;AI Transitioning to Future Partners&#xA;&lt;/h3&gt;&lt;p&gt;In another report titled &amp;ldquo;A Brief History of AI Evolution: Super Tool or Future Partner?&amp;rdquo; Senior Engineer Luo Yin from the Automation Research Institute of the Chinese Academy of Sciences discussed the evolution of AI technology from being a tool to becoming a collaborative partner.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 4&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;360px&#34; data-flex-grow=&#34;150&#34; height=&#34;466&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-4e97a2f5a6/img-b14e6ad0ed.jpeg&#34; width=&#34;700&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;Luo emphasized that with breakthroughs in large model technology and decision-making intelligence, AI is transforming from a &amp;ldquo;super tool&amp;rdquo; into a proactive collaborative &amp;ldquo;future partner.&amp;rdquo; This shift in human-machine relationships will have profound impacts on research, industry, and social life.&lt;/p&gt;&#xA;&lt;h3 id=&#34;technology-reshaping-production-and-life&#34;&gt;Technology Reshaping Production and Life&#xA;&lt;/h3&gt;&lt;p&gt;Industry experts believe that as robots evolve from &amp;ldquo;automated execution&amp;rdquo; to &amp;ldquo;autonomous intelligence,&amp;rdquo; and as AI transitions from a &amp;ldquo;super tool&amp;rdquo; to a &amp;ldquo;future partner,&amp;rdquo; technology is reshaping production and life in unprecedented ways.&lt;/p&gt;&#xA;&lt;p&gt;The event, part of the &amp;ldquo;Science and China - Thousands of Academicians, Thousands of Popular Science&amp;rdquo; series, will continue to focus on more cutting-edge technology fields, gathering more academicians and industry experts to carry out high-quality popular science activities.&lt;/p&gt;&#xA;&lt;p&gt;Nearly 300 representatives from industry, academia, research management, and youth participated in this robotics and AI-themed event, engaging in interactive discussions with the experts.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>Understanding AI: What It Is and How It&#39;s Changing Our Lives</title>
            <link>https://digitalxber.com/posts/note-6146fa9e40/</link>
            <pubDate>Sat, 18 Apr 2026 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-6146fa9e40/</guid>
            <description>&lt;h2 id=&#34;image-2&#34;&gt;&lt;img alt=&#34;Image 2&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;426px&#34; data-flex-grow=&#34;177&#34; height=&#34;900&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-6146fa9e40/img-000d63dfc0.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-6146fa9e40/img-000d63dfc0_hu_21a8466e97dde5a0.jpeg 800w, https://digitalxber.com/posts/note-6146fa9e40/img-000d63dfc0.jpeg 1600w&#34; width=&#34;1600&#34;&gt;&#xA;&lt;/h2&gt;&lt;p&gt;&lt;strong&gt;1. Conclusion: AI is not a &amp;ldquo;thinking human&amp;rdquo; but a &amp;ldquo;learning tool&amp;rdquo;&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;In recent years, artificial intelligence (AI) has become one of the hottest topics. Some believe it can do everything, while others fear it will replace human jobs, and some think of it as a &amp;ldquo;machine that thinks like a human.&amp;rdquo; To explain AI in simple terms:&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;AI is not a being with self-awareness; it is a technical system that learns patterns from data to accomplish specific tasks.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;It excels at recognizing patterns, processing information, generating content, and assisting in decision-making. For example, common applications like smartphone camera optimization, map navigation, voice assistants, video recommendations, and even online customer service all involve AI.&lt;/p&gt;&#xA;&lt;h2 id=&#34;2-what-exactly-is-ai&#34;&gt;&lt;strong&gt;2. What Exactly is AI?&lt;/strong&gt;&#xA;&lt;/h2&gt;&lt;p&gt;AI, short for &amp;ldquo;Artificial Intelligence,&amp;rdquo; refers to the capability of machines to exhibit behaviors similar to human intelligence. Here, &amp;ldquo;intelligence&amp;rdquo; does not necessarily mean the machine truly understands the world; rather, it can perform certain tasks as if it can &amp;ldquo;judge, predict, and generate.&amp;rdquo;&lt;/p&gt;&#xA;&lt;p&gt;From an application perspective, AI mainly performs the following tasks:&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;&lt;strong&gt;Recognition&lt;/strong&gt;: Understanding images, speech, and text.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Prediction&lt;/strong&gt;: Judging what might happen next based on existing data.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Generation&lt;/strong&gt;: Creating text, images, videos, code, and other content.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Decision Support&lt;/strong&gt;: Helping humans analyze information quickly and provide suggestions.&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;p&gt;Thus, AI is more like an increasingly capable &amp;ldquo;digital assistant&amp;rdquo; rather than a mysterious black box.&lt;/p&gt;&#xA;&lt;h2 id=&#34;3-the-relationship-between-ai-machine-learning-and-large-models&#34;&gt;&lt;strong&gt;3. The Relationship Between AI, Machine Learning, and Large Models&lt;/strong&gt;&#xA;&lt;/h2&gt;&lt;p&gt;Many people confuse AI, machine learning, and large models, but they represent a hierarchy:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;AI&lt;/strong&gt; is the broadest concept, referring to the ability of machines to possess intelligent capabilities.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Machine Learning&lt;/strong&gt; is a method of implementing AI, focusing on enabling machines to learn patterns from data rather than relying solely on manually written rules.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Deep Learning&lt;/strong&gt; is a significant branch of machine learning that excels at processing complex information such as images, speech, and text.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Large Models&lt;/strong&gt; are a class of AI systems that have gained attention in recent years, typically characterized by vast parameters and strong general capabilities, capable of tasks like conversation, writing, summarizing, translating, and programming.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;You can think of it this way:&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;AI is the goal, machine learning is the method, and large models are a powerful current form of implementation.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;4-why-has-ai-suddenly-become-so-popular-in-recent-years&#34;&gt;&lt;strong&gt;4. Why Has AI Suddenly Become So Popular in Recent Years?&lt;/strong&gt;&#xA;&lt;/h2&gt;&lt;p&gt;AI is not a new concept, but it has rapidly gained traction in recent years due to three simultaneous developments:&lt;/p&gt;&#xA;&lt;h3 id=&#34;1-increasing-data-availability&#34;&gt;&lt;strong&gt;1. Increasing Data Availability&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;The internet, mobile devices, and various digital platforms have accumulated vast amounts of text, images, audio, and behavioral data, providing ample material for AI learning.&lt;/p&gt;&#xA;&lt;h3 id=&#34;2-enhanced-computing-power&#34;&gt;&lt;strong&gt;2. Enhanced Computing Power&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;Improvements in chips and cloud computing capabilities have made it feasible to train models that previously required a long time, enabling complex tasks to be completed more quickly.&lt;/p&gt;&#xA;&lt;h3 id=&#34;3-significant-model-capability-improvements&#34;&gt;&lt;strong&gt;3. Significant Model Capability Improvements&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;The emergence of large language models has allowed ordinary users to intuitively realize that machines can not only &amp;ldquo;calculate&amp;rdquo; but also &amp;ldquo;write,&amp;rdquo; &amp;ldquo;answer,&amp;rdquo; &amp;ldquo;summarize,&amp;rdquo; and &amp;ldquo;create.&amp;rdquo;&lt;/p&gt;&#xA;&lt;p&gt;This has brought AI from the laboratory into public applications, prompting various industries to rethink efficiency, content production, and service methods.&lt;/p&gt;&#xA;&lt;h2 id=&#34;5-how-is-ai-already-affecting-our-lives&#34;&gt;&lt;strong&gt;5. How is AI Already Affecting Our Lives?&lt;/strong&gt;&#xA;&lt;/h2&gt;&lt;p&gt;Many people think of AI as a &amp;ldquo;future technology,&amp;rdquo; but it has already entered our daily lives.&lt;/p&gt;&#xA;&lt;h3 id=&#34;1-content-acquisition&#34;&gt;&lt;strong&gt;1. Content Acquisition&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;Recommendation algorithms determine what content you are more likely to see on short video platforms and news platforms.&lt;/p&gt;&#xA;&lt;h3 id=&#34;2-work-collaboration&#34;&gt;&lt;strong&gt;2. Work Collaboration&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;AI can help organize meeting minutes, draft initial documents, generate tables, and summarize information, significantly reducing repetitive labor.&lt;/p&gt;&#xA;&lt;h3 id=&#34;3-learning-and-education&#34;&gt;&lt;strong&gt;3. Learning and Education&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;AI can provide personalized Q&amp;amp;A, language practice, and knowledge summaries, becoming a tool for assisting learning.&lt;/p&gt;&#xA;&lt;h3 id=&#34;4-healthcare-and-services&#34;&gt;&lt;strong&gt;4. Healthcare and Services&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;In medical image recognition, health management, customer service robots, and intelligent consultations, AI has already provided substantial assistance.&lt;/p&gt;&#xA;&lt;h3 id=&#34;5-creative-production&#34;&gt;&lt;strong&gt;5. Creative Production&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;From copywriting and posters to short video scripts, AI is becoming an &amp;ldquo;accelerator&amp;rdquo; for creators.&lt;/p&gt;&#xA;&lt;p&gt;In other words, AI does not always appear as an independent product; more often, it is embedded in the tools and platforms we use.&lt;/p&gt;&#xA;&lt;h2 id=&#34;6-will-ai-replace-humans&#34;&gt;&lt;strong&gt;6. Will AI Replace Humans?&lt;/strong&gt;&#xA;&lt;/h2&gt;&lt;p&gt;This is one of the most common questions. A more accurate statement is not &amp;ldquo;Will AI replace all humans?&amp;rdquo; but:&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;AI will first replace some repetitive and standardized workflows while reshaping the skill requirements of many jobs.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;For instance, basic organization, simple translation, template writing, and preliminary data summarization are tasks that AI can accelerate or even replace. However, jobs that involve complex judgment, emotional understanding, cross-domain integration, accountability, and creative decision-making still have a clear human advantage.&lt;/p&gt;&#xA;&lt;p&gt;The more likely future scenario is:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;Those who do not use AI will fall behind in efficiency.&lt;/li&gt;&#xA;&lt;li&gt;Those who can use AI will spend more time on judgment, strategy, and creativity.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;Therefore, rather than worrying about being completely replaced, it is better to learn how to collaborate with AI as soon as possible.&lt;/p&gt;&#xA;&lt;h2 id=&#34;7-how-should-ordinary-people-understand-and-use-ai&#34;&gt;&lt;strong&gt;7. How Should Ordinary People Understand and Use AI?&lt;/strong&gt;&#xA;&lt;/h2&gt;&lt;p&gt;For most people, understanding AI does not require starting with complex technical details; instead, three basic understandings can be established:&lt;/p&gt;&#xA;&lt;h3 id=&#34;1-treat-ai-as-a-tool-not-an&#34;&gt;&lt;strong&gt;1. Treat AI as a Tool, Not an &amp;ldquo;Absolute Authority&amp;rdquo;&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;AI can make mistakes and may &amp;ldquo;speak nonsense&amp;rdquo; with confidence. It can help improve your efficiency but cannot replace your final judgment.&lt;/p&gt;&#xA;&lt;h3 id=&#34;2-learn-to-ask-clear-questions&#34;&gt;&lt;strong&gt;2. Learn to Ask Clear Questions&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;AI&amp;rsquo;s performance largely depends on how clearly you provide information. The more specific your questions and clearer your goals, the better the results are likely to be.&lt;/p&gt;&#xA;&lt;h3 id=&#34;3-start-with-high-frequency-scenarios&#34;&gt;&lt;strong&gt;3. Start with High-Frequency Scenarios&lt;/strong&gt;&#xA;&lt;/h3&gt;&lt;p&gt;For example, writing emails, summarizing, organizing materials, outlining, and generating event proposals. By starting with small tasks you encounter daily, the value of AI will quickly become apparent.&lt;/p&gt;&#xA;&lt;h2 id=&#34;8-what-is-the-most-important-skill-in-the-ai-era&#34;&gt;&lt;strong&gt;8. What is the Most Important Skill in the AI Era?&lt;/strong&gt;&#xA;&lt;/h2&gt;&lt;p&gt;If the past emphasized &amp;ldquo;remembering answers,&amp;rdquo; then in the AI era, the more important skills will become:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;The ability to ask good questions&lt;/li&gt;&#xA;&lt;li&gt;The ability to judge the authenticity of information&lt;/li&gt;&#xA;&lt;li&gt;The ability to integrate information from multiple sources&lt;/li&gt;&#xA;&lt;li&gt;The ability to collaborate with tools&lt;/li&gt;&#xA;&lt;li&gt;The ability to maintain learning and adapt to changes&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;AI will lower many job thresholds but will also increase the value of &amp;ldquo;judgment&amp;rdquo; and &amp;ldquo;creativity.&amp;rdquo;&lt;/p&gt;&#xA;&lt;h2 id=&#34;9-conclusion-the-key-is-not-to-fear-ai-but-to-learn-to-use-it&#34;&gt;&lt;strong&gt;9. Conclusion: The Key is Not to Fear AI, but to Learn to Use It&lt;/strong&gt;&#xA;&lt;/h2&gt;&lt;p&gt;Every technological revolution brings anxiety but also new opportunities. The significance of AI lies not only in making machines smarter but also in helping humans free up time from repetitive labor to engage in higher-value thinking, creativity, and connections.&lt;/p&gt;&#xA;&lt;p&gt;For ordinary people, the best starting point is not to chase every concept but to take the first step: understand it, use it, and observe how it changes your work and life.&lt;/p&gt;&#xA;&lt;p&gt;Once you start to truly engage with AI, it will no longer be just a distant technical term but will become part of your daily capabilities.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>2025 AI Model Rankings: Domestic and International Comparisons</title>
            <link>https://digitalxber.com/posts/note-01-15d1efba9a/</link>
            <pubDate>Fri, 17 Apr 2026 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-01-15d1efba9a/</guid>
            <description>&lt;h2 id=&#34;1-tongyi-qianwen-alibaba&#34;&gt;1. Tongyi Qianwen (Alibaba)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Core Competence&lt;/strong&gt;: Leading Chinese understanding capabilities, outstanding logical reasoning and text creation, supporting millions of context windows and multimodal interaction.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: Enterprise services, e-commerce, financial customer service, with over 1.5 billion daily calls, serving over 90,000 enterprises.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Multiple iterations of Tongyi Qianwen, such as Tongyi Qianwen 2.0, continuously optimizing performance, functionality, and multimodal capabilities.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;2-doubao-large-model-bytedance&#34;&gt;2. Doubao Large Model (ByteDance)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Technical Highlights&lt;/strong&gt;: Nearly 60 million monthly active users, second in global user count, excels in image understanding and multimodal fusion, with significant potential in education.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Cooperation Ecosystem&lt;/strong&gt;: Collaborates with over 500 enterprises, focusing on family companionship and learning assistance scenarios.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Continuously releases different versions, upgrading image understanding and multimodal fusion to better meet diverse scenario needs.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;3-wenxin-yiyan-40-baidu&#34;&gt;3. Wenxin Yiyan 4.0 (Baidu)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Commercial Advantages&lt;/strong&gt;: Annual call volume increases by 30 times, with 1.5 billion daily calls, leading in mathematical science and language ability assessments.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Industry Coverage&lt;/strong&gt;: Deep integration with Baidu&amp;rsquo;s knowledge graph, supporting healthcare, education, and finance sectors.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Currently centered on version 4.0, with previous versions like Wenxin Yiyan 3.0, each progressively enhancing knowledge coverage and reasoning capabilities.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;4-iflytek-spark-iflytek&#34;&gt;4. iFlytek Spark (iFlytek)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Multilingual Breakthrough&lt;/strong&gt;: Supports interaction in over 30 languages, with over 200 million app downloads, mature solutions in healthcare and finance.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Technical Features&lt;/strong&gt;: Industry benchmark in speech recognition and synthesis, widely applied in education.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Versions like iFlytek Spark 2.0 and 3.0 continuously improve multilingual interaction and speech capabilities across various industries.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;5-kimi-smart-assistant-dark-side-of-the-moon&#34;&gt;5. Kimi Smart Assistant (Dark Side of the Moon)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Long Text Processing&lt;/strong&gt;: Supports input of 200,000 Chinese characters, high popularity in the A-share market, suitable for data analysis and professional document interpretation.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Scenario Expansion&lt;/strong&gt;: Plans to extend into legal and scientific research fields.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Continuously updates versions to enhance long text processing capabilities and expand application scenarios.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;6-deepseek&#34;&gt;6. DeepSeek&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Benchmark in Programming&lt;/strong&gt;: A complete open-source model ecosystem, R1 version supports code generation and debugging, with comprehensive capabilities comparable to GPT-4.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Technical Innovations&lt;/strong&gt;: Breakthroughs in dynamic reasoning optimization and domain adaptation technology, representing domestic large models in internationalization.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Currently has R1 version, with more versions possibly to be released for optimizing code generation and reasoning capabilities.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;7-zhipu-qingyan-glm-4-tsinghua-university&#34;&gt;7. Zhipu Qingyan GLM-4 (Tsinghua University)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Interactive Innovation&lt;/strong&gt;: The first domestic model with a trillion parameters supporting video calls, enhancing natural human-computer interaction.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Academic Background&lt;/strong&gt;: Developed by Tsinghua team, balanced capabilities in knowledge Q&amp;amp;A and creative writing.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Developed from the GLM series to GLM-4 version, with significant improvements in parameter scale and interaction capabilities.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;8-hunyuan-large-model-tencent&#34;&gt;8. Hunyuan Large Model (Tencent)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Video Generation&lt;/strong&gt;: Trillion parameter scale, supports text-to-video generation, widely applied in film and television creation.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Ecosystem Integration&lt;/strong&gt;: Deeply integrated into the WeChat ecosystem, providing personalized intelligent agent services.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Continuously updates versions to improve video generation quality and service capabilities within the WeChat ecosystem.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;9-baichuan-large-model-baichuan-intelligence&#34;&gt;9. Baichuan Large Model (Baichuan Intelligence)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Specialized in Healthcare&lt;/strong&gt;: Solves grassroots healthcare challenges as an AI doctor, with disease diagnosis assistance systems covering over 1,000 hospitals.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Open Source Layout&lt;/strong&gt;: Baichuan-7B/13B model downloads exceed one million, performing excellently on evaluation rankings.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Available in different parameter scales like Baichuan-7B and Baichuan-13B to meet various application needs.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;10-jidream-ai-bytedance&#34;&gt;10. Jidream AI (ByteDance)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Video Creation Tool&lt;/strong&gt;: Supports generating 1080P videos from text/images, leading in ease of use, deeply integrated into the Douyin ecosystem.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;User Growth&lt;/strong&gt;: Rapidly popular after launch in 2024, with a 40% usage rate among short video creators.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Continuously updates versions to optimize video generation effects and user experience.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;2025-international-ai-model-rankings&#34;&gt;2025 International AI Model Rankings&#xA;&lt;/h2&gt;&lt;h2 id=&#34;1-gpt-4o-openai&#34;&gt;1. GPT-4o (OpenAI)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Developer&lt;/strong&gt;: OpenAI&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Features&lt;/strong&gt;: Parameter scale exceeds 10 trillion, supports multimodal inputs (text/image/audio/video), reasoning abilities close to human levels, excelling in complex logic and cross-domain knowledge integration.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: Scientific analysis, cross-industry decision support, and multimedia content generation.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: May have different fine-tuned versions for specific applications in various fields.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;2-gemini-20-ultra-google-deepmind&#34;&gt;2. Gemini 2.0 Ultra (Google DeepMind)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Developer&lt;/strong&gt;: Google&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Features&lt;/strong&gt;: Native multimodal architecture, supports real-time translation in over 100 languages, deeply integrated with Google ecosystem (search/office suite), context window expanded to 2 million tokens.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: Global enterprise collaboration, real-time translation, multimodal search engine optimization.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Gemini 2.0 Ultra version available, may also have lightweight or specific function-optimized versions.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;3-claude-35--sonnet-anthropic&#34;&gt;3. Claude 3.5 – Sonnet (Anthropic)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Developer&lt;/strong&gt;: Anthropic (Google invested)&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Features&lt;/strong&gt;: 200K ~ 1M tokens context window, constitutional AI architecture ensures compliance, excels in medical and legal fields, commercialized on-demand billing.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: Legal document analysis, medical diagnosis assistance, high-security dialogue systems.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Claude 3.5 – Sonnet version available, with previous versions like Claude 2.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;4-palm--3-google&#34;&gt;4. PaLM – 3 (Google)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Developer&lt;/strong&gt;: Google&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Features&lt;/strong&gt;: Parameter scale exceeds 1 trillion, specializes in common sense reasoning and mathematical coding, leading response speed among similar models, supports 4096 tokens context.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: Automatic problem solving in education, financial quantitative model development.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Developed from the PaLM series to PaLM – 3 version, may have different fine-tuned versions.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;5-llama--3-meta&#34;&gt;5. LLaMA – 3 (Meta)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Developer&lt;/strong&gt;: Meta&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Features&lt;/strong&gt;: Open-source model with 70 billion parameters, 200% improvement in reasoning speed, performance close to GPT-4 in the open-source community, supports multilingual optimization.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: Customized AI solutions for small and medium enterprises, academic research.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Developed from the LLaMA series to LLaMA – 3 version, with community-based secondary development versions likely.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;6-falcon--200b-uae-tii&#34;&gt;6. Falcon – 200B (UAE TII)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Developer&lt;/strong&gt;: UAE Technology Innovation Institute&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Features&lt;/strong&gt;: 180 billion parameter open-source model, mathematical reasoning and code generation capabilities comparable to GPT-4, training costs only 1/3 of similar models.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: Multilingual services in the Middle East, low-cost AI infrastructure development.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Currently focused on Falcon – 200B version, with potential for optimized versions in the future.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;7-cohere-command--r-cohere&#34;&gt;7. Cohere Command – R (Cohere)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Developer&lt;/strong&gt;: Cohere (founded by former Google team)&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Features&lt;/strong&gt;: Focused on enterprise-level generative AI, supports 52 billion parameter scale, provides customized data privacy protection solutions.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: Customer service automation, intelligent management of internal documents.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Continuously iterates versions to meet diverse enterprise needs.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;8-mpt--50b-mosaicml&#34;&gt;8. MPT – 50B (MosaicML)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Developer&lt;/strong&gt;: MosaicML&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Features&lt;/strong&gt;: Open-source model with 8K tokens context length, lowest training costs in the industry, suitable for rapid deployment by small teams.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: MVP development for startups, experimental platforms for educational institutions.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Available in MPT – 50B version, may launch optimized versions for different application scenarios.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;9-nemotron--4-nvidia&#34;&gt;9. Nemotron – 4 (Nvidia)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Developer&lt;/strong&gt;: Nvidia&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Features&lt;/strong&gt;: Integrates Megatron framework, optimizes GPU computing efficiency, designed for AI chips, supports large-scale distributed training.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: Supercomputing centers, autonomous driving model training.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Continuously updates to adapt to new hardware and application needs.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;10-gopher--2-deepmind&#34;&gt;10. Gopher – 2 (DeepMind)&#xA;&lt;/h2&gt;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Developer&lt;/strong&gt;: DeepMind&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Features&lt;/strong&gt;: Reinforcement learning optimized version, sets records in game AI and protein structure prediction, supports multi-agent collaboration.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Application Scenarios&lt;/strong&gt;: Biomedicine research, complex game environment simulation.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Version Status&lt;/strong&gt;: Developed from the Gopher series to Gopher – 2 version, with potential fine-tuned versions for different fields.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;summary&#34;&gt;Summary&#xA;&lt;/h2&gt;&lt;p&gt;This article introduces the rankings of AI models in 2025, highlighting domestic models like Tongyi Qianwen and Doubao Large Model, each with unique core competencies and application scenarios, continuously updated and iterated. International models like GPT-4o and Gemini 2.0 Ultra also showcase distinctive features such as multimodal input and large-scale parameters. For detailed parameter comparison data of various AI models, click to view the comprehensive metrics provided by Mijian Integration.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>AI Transitions from Concept to Practical Applications</title>
            <link>https://digitalxber.com/posts/note-fcfe52b234/</link>
            <pubDate>Fri, 17 Apr 2026 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-fcfe52b234/</guid>
            <description>&lt;h2 id=&#34;ai-transitions-from-concept-to-practical-applications&#34;&gt;AI Transitions from Concept to Practical Applications&#xA;&lt;/h2&gt;&lt;p&gt;In recent years, AI has become a prominent topic. At the ongoing 6th China International Consumer Products Expo in Hainan, over 50 leading global tech companies are showcasing applications of AI in consumer goods, smart home technology, digital consumption, and low-altitude economy. This event allows global attendees to experience firsthand how &amp;ldquo;AI + consumption&amp;rdquo; is profoundly changing lives.&lt;/p&gt;&#xA;&lt;p&gt;As I walked through the tech consumption exhibition area, I was greeted by a plethora of AI applications and achievements across multiple fields. At the entrance, a giant model of AI glasses captured attention. From a distance, the large lenses displayed scrolling green subtitles, sparking curiosity among visitors eager to learn more.&lt;/p&gt;&#xA;&lt;p&gt;&amp;ldquo;Please say &amp;lsquo;Loki&amp;rsquo; to activate the AI assistant,&amp;rdquo; flashed on the screen of the glasses. Staff from the Rokid booth explained that when worn, users can activate the glasses by simply calling its name. The glasses can provide weather updates, identify surroundings, translate foreign menus, and even facilitate payments by scanning QR codes.&lt;/p&gt;&#xA;&lt;p&gt;The interactive features of these glasses, made possible by embedding chips and batteries into a slim frame, drew gasps of amazement from the crowd.&lt;/p&gt;&#xA;&lt;p&gt;However, AI&amp;rsquo;s journey from concept to practicality is not limited to just glasses. At the booth of Yushu Technology Co., a humanoid robot engaged in handshakes and dance with attendees. Chen Tong, the manager responsible for online sales, shared that the robot is powered by a large model that allows users to control it through voice commands, primarily for entertainment and cultural tourism.&lt;/p&gt;&#xA;&lt;p&gt;At the Sinopec energy supply station booth, the humanoid robot demonstrated a seamless process of removing a fuel nozzle, filling a disposable cup, and returning the nozzle to its holder.&lt;/p&gt;&#xA;&lt;p&gt;At the Taishan Sports Industry Group booth, a cyclist scanned a QR code on the bike&amp;rsquo;s screen with their phone, entering a mini-program. The moment they pedaled, the screen displayed cycling time, speed, heart rate, and calories burned.&lt;/p&gt;&#xA;&lt;p&gt;According to Song Kun, the head of the company&amp;rsquo;s branding department, these capabilities are supported not only by the bike&amp;rsquo;s hardware but also by the underlying data and software.&lt;/p&gt;&#xA;&lt;p&gt;AI&amp;rsquo;s influence extends beyond the tech consumption exhibition area. In the national goods exhibition area, a humanoid robot from Li Gong Industrial Co. was seen calligraphing the character &amp;ldquo;福&amp;rdquo; (fortune), attracting numerous visitors for photos and inquiries.&lt;/p&gt;&#xA;&lt;p&gt;Ma Chenchen, the company&amp;rsquo;s client manager, explained the technology behind this: they first had a calligrapher write several times to collect data on the movements of the joints, which was then input into a specialized server. With the support of algorithms and computational power, the data was optimized and fed into the robot&amp;rsquo;s brain for reinforcement learning. Once trained to a certain level, the robot can execute commands sent via voice or connected devices.&lt;/p&gt;&#xA;&lt;p&gt;&amp;ldquo;The key to robot intelligence lies in data collection from real people, followed by computational power and algorithmic optimization to create a supportive data environment for its intelligent functions,&amp;rdquo; Ma explained. He emphasized that computational power is akin to intelligence, while algorithms are the methods for solving equations. In Guangdong, this work is supported by computational resources from Gansu.&lt;/p&gt;&#xA;&lt;p&gt;The AI wave is surging forward.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>Alibaba Unveils Ambitious AI Strategy at 2025 Yunqi Conference</title>
            <link>https://digitalxber.com/posts/note-02-f5342f3c45/</link>
            <pubDate>Fri, 17 Apr 2026 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-02-f5342f3c45/</guid>
            <description>&lt;h2 id=&#34;introduction&#34;&gt;Introduction&#xA;&lt;/h2&gt;&lt;p&gt;At the end of September, a light rain fell in Hangzhou, but the AI fervor at Yunqi Town made it feel like summer had not yet faded.&lt;/p&gt;&#xA;&lt;p&gt;On September 24, the 2025 Yunqi Conference was held as scheduled. Alibaba Group CEO and Chairman of Alibaba Cloud Intelligence, Wu Yongming, delivered a speech titled &amp;ldquo;The Path to Super Artificial Intelligence.&amp;rdquo;&lt;/p&gt;&#xA;&lt;p&gt;This was Wu&amp;rsquo;s first appearance at the Yunqi Conference after more than a year at the helm of Alibaba Cloud. He stated that &amp;ldquo;the greatest imagination of generative AI is not to create one or two new super apps on a mobile screen, but to take over the digital world and change the physical world.&amp;rdquo;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;A Year of Progress&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;If this statement was more of a vision a year ago, it has now transformed into a more concrete roadmap and aggressive actions.&lt;/p&gt;&#xA;&lt;p&gt;At this year&amp;rsquo;s Yunqi Conference, Alibaba Cloud unveiled a plethora of new products. Among them was the flagship model Qwen3-Max, which is currently the most powerful model in the Alibaba Tongyi model family, outperforming GPT-5 and Claude Opus 4, ranking among the top three globally on LMArena.&lt;/p&gt;&#xA;&lt;p&gt;In addition to the flagship model, Alibaba also launched six new models, including the next-generation foundational model architecture Qwen3-Next, the programming model Qwen3-Coder, the visual understanding model Qwen3-VL, the multimodal model Qwen3-Omni, the visual foundational model Wan2.5-preview, and the speech model Tongyi Bailin.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 1&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;442px&#34; data-flex-grow=&#34;184&#34; height=&#34;586&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-02-f5342f3c45/img-1dd83f793b.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-02-f5342f3c45/img-1dd83f793b_hu_1396b80ad93589fa.jpeg 800w, https://digitalxber.com/posts/note-02-f5342f3c45/img-1dd83f793b.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;More noteworthy were Wu Yongming&amp;rsquo;s two bold new assertions.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;The Future of Operating Systems&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;He made a definitive statement: large models are the next generation of operating systems. Large models will engulf software, allowing anyone to create an infinite number of applications using natural language. In the future, almost all software interacting with the computational world may be generated by agents from large models, rather than traditional commercial software.&lt;/p&gt;&#xA;&lt;p&gt;As a result, Alibaba Cloud has been undergoing a reconstruction of all operating systems—from underlying computing power to infrastructure and cloud services—to align with the changes brought by large models.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;The Rise of Super AI Cloud&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;The second assertion builds on this logic: the Super AI Cloud is the next generation of computers. Drawing parallels with the stages of computer development, natural language is the programming language of the AI era, agents are the new software, context is the new memory, and LLMs will serve as the middleware for user, software, and AI computational resource interactions, becoming the OS of the AI era.&lt;/p&gt;&#xA;&lt;p&gt;Alibaba Cloud&amp;rsquo;s goal is to establish a &amp;ldquo;Super AI Cloud&amp;rdquo; to provide a global intelligent computing network.&lt;/p&gt;&#xA;&lt;p&gt;In February, Alibaba announced a three-year plan for AI infrastructure construction worth 380 billion. Wu Yongming added a new plan today—by 2032, compared to 2022, the energy consumption scale of Alibaba Cloud&amp;rsquo;s global data centers will increase tenfold to welcome the arrival of the ASI era.&lt;/p&gt;&#xA;&lt;p&gt;Alibaba Cloud also proposed a new development strategy and goal for AI: not the commonly discussed AGI (Artificial General Intelligence), but a further step towards ASI (Artificial Super Intelligence).&lt;/p&gt;&#xA;&lt;p&gt;Wu Yongming explained the three stages to reach super artificial intelligence:&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;&lt;strong&gt;Intelligent Emergence&lt;/strong&gt;: AI learns from humans, acquiring generalized intelligence through the collection of global knowledge, gradually developing reasoning abilities.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Autonomous Action&lt;/strong&gt;: AI masters tool usage and programming capabilities to assist humans, which is the current stage of the industry.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Self-Iteration&lt;/strong&gt;: AI connects with the physical world&amp;rsquo;s complete raw data for autonomous learning, ultimately able to &amp;ldquo;surpass humans.&amp;rdquo;&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;p&gt;In 2025, the global large model field is progressing amidst challenges. After OpenAI launched GPT-5, its performance fell short of market expectations, leading to criticisms of stagnation and setbacks in model innovation. Meanwhile, Meta and OpenAI are making more aggressive capital investments—no one wants to miss out on this wave of technological revolution.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Alibaba&amp;rsquo;s Commitment&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;Now, Alibaba Cloud is proving through action that it not only intends to invest but to invest aggressively.&lt;/p&gt;&#xA;&lt;p&gt;The market has responded positively to Alibaba Cloud&amp;rsquo;s new strategy. Today, Alibaba&amp;rsquo;s Hong Kong stocks surged, rising over 9% during trading, reaching a new high since October 2021.&lt;/p&gt;&#xA;&lt;h2 id=&#34;new-model-launches&#34;&gt;New Model Launches&#xA;&lt;/h2&gt;&lt;p&gt;Before the Yunqi Conference, Lin Junyang, head of the Qwen model team at Alibaba, teased on Twitter that they would launch more than six new products, none of which would be &amp;ldquo;small items.&amp;rdquo;&lt;/p&gt;&#xA;&lt;p&gt;When the models were officially released, the number exceeded expectations, marking a sincere launch. Alibaba Cloud CTO Zhou Jingren flipped through the PPT at the conference rapidly, rushing through his points but still exceeding the time limit.&lt;/p&gt;&#xA;&lt;p&gt;Alibaba Cloud launched a total of seven new models, each with significant improvements in scale and performance:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Qwen3-Max&lt;/strong&gt;: Flagship model with a pre-training data volume of 36 trillion tokens and over a trillion parameters, significantly enhancing coding and agent tool invocation capabilities.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Qwen-Next&lt;/strong&gt;: Next-generation model architecture and series. The total parameters of the model are 80 billion, with only 3 billion activated, comparable to the flagship model Qwen3 with 235 billion parameters. The training cost has decreased by over 90% compared to the dense model Qwen3-32B.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Qwen 3-VL (Visual Understanding)&lt;/strong&gt;: Capable of accurately interpreting images and charts, with a breakthrough in &amp;ldquo;visual programming&amp;rdquo; ability, converting visual design drafts directly into front-end code and operating mobile devices and computers, advancing from mere &amp;ldquo;seeing&amp;rdquo; to understanding and execution.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Qwen3-Coder (Code Model)&lt;/strong&gt;: Significantly improves generation speed, code quality, and security, making it easier to complete complex tasks from code completion and bug fixing to generating complete projects with one click.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Qwen3-Omni&lt;/strong&gt;: A native multimodal model that can &amp;ldquo;hear, speak, see, and write&amp;rdquo;; it interacts naturally like chatting with a person, understanding audio and video while maintaining capabilities in text and images, suitable for use in AI applications for vehicles, glasses, and mobile phones.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Tongyi Wanxiang Wan2.5-preview&lt;/strong&gt;: A new visual foundational model with capabilities for generating video from text, images from text, and image editing, capable of generating matching human voices, sound effects, and music BGM.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Tongyi Bailin&lt;/strong&gt;: A new family of speech models, including speech recognition and synthesis sub-models. For example, Fun-CosyVoice can provide hundreds of preset voice styles for applications in customer service, sales, live e-commerce, consumer electronics, audiobooks, and children&amp;rsquo;s entertainment.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 2&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;175px&#34; data-flex-grow=&#34;72&#34; height=&#34;1280&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-02-f5342f3c45/img-6a9d06b10a.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-02-f5342f3c45/img-6a9d06b10a_hu_d4ea8d87e2b769d8.jpeg 800w, https://digitalxber.com/posts/note-02-f5342f3c45/img-6a9d06b10a.jpeg 934w&#34; width=&#34;934&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;Alibaba Cloud does not rely solely on static datasets to demonstrate model capabilities. In blind tests on authoritative rankings like LMArena, Alibaba&amp;rsquo;s flagship model Qwen3-Max has already ranked third on the Chatbot Arena leaderboard.&lt;/p&gt;&#xA;&lt;p&gt;Following the global AI industry explosion driven by DeepSeek, a domestic open-source model competition has ignited, contrasting sharply with last year&amp;rsquo;s closed-door approaches.&lt;/p&gt;&#xA;&lt;p&gt;Both domestically and internationally, this year has seen a round of open-source model battles, with nearly all companies still investing in models increasing their open-source efforts. Alibaba stands out as the most aggressive among domestic giants in pursuing an open-source route.&lt;/p&gt;&#xA;&lt;p&gt;This stems from Alibaba being one of the first companies in China to open-source models and build a model ecosystem. &lt;strong&gt;These investments have now yielded tangible returns, motivating Alibaba to make even more aggressive investments.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;DeepSeek and Qwen are among the few models that have gained global recognition. After the open-source surge initiated by DeepSeek, Qwen has once again attracted attention in the global AI community, entering a new phase of growth.&lt;/p&gt;&#xA;&lt;p&gt;As of now, Alibaba Tongyi has open-sourced over 300 models, covering various sizes of &amp;ldquo;full-size&amp;rdquo; and LLMs, programming, image, speech, and video across all modalities.&lt;/p&gt;&#xA;&lt;p&gt;Globally, Tongyi&amp;rsquo;s large models are the leading open-source models, with downloads exceeding 600 million and over 170,000 derivative models.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 3&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;626px&#34; data-flex-grow=&#34;260&#34; height=&#34;414&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-02-f5342f3c45/img-3d6e5e7ad1.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-02-f5342f3c45/img-3d6e5e7ad1_hu_a81ff2d07ebc5aad.jpeg 800w, https://digitalxber.com/posts/note-02-f5342f3c45/img-3d6e5e7ad1.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Agent Development Framework&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;In addition to models, Alibaba Cloud has also released a new agent development framework, ModelStudio-ADK—agents can autonomously plan and invoke models, leading to increased computational consumption. Alibaba Cloud disclosed a figure indicating that with the continuous enhancement of model capabilities and the explosion of agent applications, the daily invocation volume of models on Alibaba Cloud&amp;rsquo;s Bailian platform has grown 15-fold over the past year.&lt;/p&gt;&#xA;&lt;p&gt;Investments in model open-sourcing not only accelerate model iteration but have also translated into revenue in the cloud. Alibaba has begun to establish a commercial closed loop for the AI era—its latest quarterly report shows that Alibaba Cloud&amp;rsquo;s quarterly revenue surged 26% year-on-year, with AI-related revenue achieving triple-digit growth for eight consecutive quarters.&lt;/p&gt;&#xA;&lt;p&gt;According to a report from the international market research firm Omdia, the AI cloud market in China is expected to reach 22.3 billion yuan in the first half of 2025, with Alibaba Cloud holding a 35.8% market share, ranking first, surpassing the combined share of the second to fourth places.&lt;/p&gt;&#xA;&lt;h2 id=&#34;competing-in-the-llm-era&#34;&gt;Competing in the LLM Era&#xA;&lt;/h2&gt;&lt;p&gt;In 2024, with OpenAI&amp;rsquo;s Sora release and GPT-5 development stagnating, discussions about technical routes briefly led to a dip in sentiment in the global large model field.&lt;/p&gt;&#xA;&lt;p&gt;However, this sentiment has largely dissipated. Just days before the Yunqi Conference, NVIDIA announced a $100 billion investment in OpenAI. Wu Yongming predicted at the conference that global AI investments will exceed $4 trillion in the next five years.&lt;/p&gt;&#xA;&lt;p&gt;Alibaba Cloud CTO Zhou Jingren admitted in a media interview after the conference that there are now very few major disagreements regarding technical routes in the industry. Almost all companies globally are aggressively investing in AI competition and rapidly releasing models. &lt;strong&gt;The question now is how each vendor approaches this.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;&amp;ldquo;Current model competition is essentially a competition between systems,&amp;rdquo; Zhou Jingren said. &amp;ldquo;The innovation of model development does not involve holding back major breakthroughs; it is complementary to the underlying infrastructure and cloud.&amp;rdquo;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Understanding &amp;lsquo;System&amp;rsquo;&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;How to understand &amp;ldquo;system&amp;rdquo;? This likely points more towards a strategic choice in AI.&lt;/p&gt;&#xA;&lt;p&gt;After DeepSeek changed the global AI narrative, all major companies are increasing their investments in AI, from underlying computing power to cloud computing and open-source efforts.&lt;/p&gt;&#xA;&lt;p&gt;The divergence in AI routes among major companies has formed interesting contrasts—take Tencent&amp;rsquo;s recent ecological conference as an example, where Tencent focused more on scenarios and B-end and C-end implementations, first applying AI to its own business before turning to external applications; ByteDance, on the other hand, resembles iOS, adopting a regimented approach from models to applications, but with a tendency to first close-source and roll out better versions, with a slower pace in open-sourcing.&lt;/p&gt;&#xA;&lt;p&gt;2023 is a pivotal year for Alibaba Cloud. After Wu Yongming took over as CEO, he proposed an &amp;ldquo;AI-driven, public cloud-first&amp;rdquo; strategy.&lt;/p&gt;&#xA;&lt;p&gt;Since then, Alibaba Cloud has completed several key tasks: returning to the public cloud, cutting low-profit projects, and investing heavily in AI, not only externally investing in AI startups but also significantly investing in self-developed models, open-source efforts, and infrastructure reconstruction.&lt;/p&gt;&#xA;&lt;p&gt;Alibaba Cloud&amp;rsquo;s current trajectory is closer to Google&amp;rsquo;s. From the underlying computing infrastructure to cloud computing and then to the upper-level models, both Alibaba and Google adopt a full-stack self-research strategy, ensuring that each layer is internationally competitive.&lt;/p&gt;&#xA;&lt;p&gt;The ASI proposed by Alibaba today is not a new term. In March of this year, Google DeepMind disclosed its &amp;ldquo;AGI Six-Level Roadmap,&amp;rdquo; which corresponds closely to Alibaba&amp;rsquo;s ASI trilogy: the third stage of ASI, &amp;ldquo;surpassing humans,&amp;rdquo; is quite similar to DeepMind&amp;rsquo;s defined AGI Level 6.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 4&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;335px&#34; data-flex-grow=&#34;139&#34; height=&#34;773&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-02-f5342f3c45/img-5fe4ed3ed5.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-02-f5342f3c45/img-5fe4ed3ed5_hu_f1575bc4d1df803.jpeg 800w, https://digitalxber.com/posts/note-02-f5342f3c45/img-5fe4ed3ed5.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;Aggressively investing in AI also stems from the inseparable relationship between AI and cloud computing. Today, Alibaba Cloud even announced a new positioning as a &amp;ldquo;full-stack AI service provider.&amp;rdquo; &amp;ldquo;Tokens are the electricity of the future AI world,&amp;rdquo; Wu Yongming stated.&lt;/p&gt;&#xA;&lt;p&gt;Undoubtedly, we are still in the early stages of the AI era. Currently, the volume of model calls accounts for a very small portion of enterprise cloud consumption, but the trend is crucial.&lt;/p&gt;&#xA;&lt;p&gt;In a post-conference interview, Xu Dong, General Manager of Alibaba Cloud&amp;rsquo;s Tongyi Large Model Business, told the media that a year ago, the volume of large model calls mostly came from offline tasks like data labeling; but a year later, online task calls have increased by dozens of times, with enterprises across various industries embedding large models into their production processes—this proves that large models are rapidly bringing incremental growth to the cloud market.&lt;/p&gt;&#xA;&lt;p&gt;For the past 16 years, providing the &amp;ldquo;water and electricity&amp;rdquo; of the digital world has long been Alibaba Cloud&amp;rsquo;s explanation of its market value—this is consistent with Alibaba Cloud&amp;rsquo;s current call of being the &amp;ldquo;Android of the LLM era,&amp;rdquo; which is fundamentally the same ecological niche.&lt;/p&gt;&#xA;&lt;p&gt;Whether proposing a new roadmap or a new positioning, Alibaba needs to find its home court in the AI era and secure a leading position before the application market explodes; this goal has never been clearer.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>DeepSeek to Launch Next-Gen AI Model V4 in February</title>
            <link>https://digitalxber.com/posts/note-03-cf64d3b0a0/</link>
            <pubDate>Fri, 17 Apr 2026 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-03-cf64d3b0a0/</guid>
            <description>&lt;h2 id=&#34;deepseeks-upcoming-ai-model&#34;&gt;DeepSeek&amp;rsquo;s Upcoming AI Model&#xA;&lt;/h2&gt;&lt;p&gt;According to recent reports from Reuters, Chinese AI startup DeepSeek is set to launch its next-generation AI model, V4, in mid-February. This model boasts strong coding capabilities and may outperform competitors such as Anthropic&amp;rsquo;s Claude and OpenAI&amp;rsquo;s GPT series. A year ago, DeepSeek released its large model R1, which the BBC described as showcasing China&amp;rsquo;s competitiveness in the AI field. Experts interviewed by the Global Times noted that in just one year, China has significantly narrowed the gap with the United States in AI, with DeepSeek and ChatGPT representing different stages in this evolution.&lt;/p&gt;&#xA;&lt;h2 id=&#34;diverging-paths-in-ai-development&#34;&gt;Diverging Paths in AI Development&#xA;&lt;/h2&gt;&lt;p&gt;A year ago, Chen Yan, executive director of the Japan Research Institute (China), noticed that DeepSeek had gained significant attention in Zhongguancun. Media reporters flocked to the building, and many Japanese companies expressed interest in investing. However, Chen remarked that these companies had missed the best investment opportunities, as even a $10 million investment is now insufficient to enter the market.&lt;/p&gt;&#xA;&lt;p&gt;Foreign media outlets like The Wall Street Journal described the launch of DeepSeek&amp;rsquo;s R1 model as shocking to the world. The model was trained in just two months at a fraction of the cost spent by American companies like OpenAI, yet it performed comparably to ChatGPT and Meta&amp;rsquo;s Llama model. By 2025, more Chinese large model companies are expected to follow the latest developments in AI, joining the global first tier of large models.&lt;/p&gt;&#xA;&lt;p&gt;According to a report from third-party AI model aggregator OpenRouter and venture capital firm Andreessen Horowitz, China&amp;rsquo;s open-source AI models account for nearly 30% of global AI technology usage. The open-source approach is gaining trust among developers worldwide, with companies like Airbnb and even Meta utilizing Alibaba&amp;rsquo;s Qwen large model. AI researcher and author Sebastian Raschka noted that Alibaba&amp;rsquo;s Qwen3 series models, like DeepSeek&amp;rsquo;s R1, are among the most noteworthy open-source models to watch in 2025.&lt;/p&gt;&#xA;&lt;h2 id=&#34;different-approaches-to-ai&#34;&gt;Different Approaches to AI&#xA;&lt;/h2&gt;&lt;p&gt;Alibaba reflected on the rapid development of AI, noting that OpenAI launched ChatGPT on November 30, 2022, and by April 2023, Qwen series models were released. Alibaba began its AI large model research as early as 2018 and has since launched several models, including the multimodal M6 and language model PLUG, establishing itself as a key player in the global AI landscape. To date, Alibaba has open-sourced nearly 400 models, with over 180,000 derivative models and more than 700 million downloads.&lt;/p&gt;&#xA;&lt;p&gt;Shen Yang, a dual-appointed professor at Tsinghua University&amp;rsquo;s School of Journalism and Communication and School of Artificial Intelligence, explained that the U.S. and China have developed two distinct paths in large model AI. The U.S. focuses on continuous enhancement of cutting-edge capabilities, closed-source models, and platform-based products, aiming to create a controllable, billable, and governable infrastructure. In contrast, China emphasizes open-source weights, engineering efficiency, and rapid industrial diffusion, prioritizing the development of sufficiently strong capabilities that can be quickly replicated and implemented in real business systems.&lt;/p&gt;&#xA;&lt;h2 id=&#34;the-future-of-ai-competition&#34;&gt;The Future of AI Competition&#xA;&lt;/h2&gt;&lt;p&gt;AI blogger Li Shanglong, who recently attended the CES in Las Vegas, described the U.S. as having two rivers: one fully immersed in the AI era and the other gradually being permeated. He noted that while Silicon Valley is buzzing with discussions about AI and ChatGPT, many ordinary people&amp;rsquo;s lives outside of Silicon Valley remain less AI-integrated. Li, who returned to China to start a business, believes that AI will not change the U.S. overnight but will gradually alter the lifestyles of some individuals.&lt;/p&gt;&#xA;&lt;p&gt;Northeastern University professor Li Xiangming highlighted that AI is deeply integrated into everyday life in the U.S., primarily in soft applications, such as algorithm-driven streaming recommendations and office productivity tools. However, the physical hardware aspect is still on the cusp of a breakthrough.&lt;/p&gt;&#xA;&lt;p&gt;At CES, Li was impressed by the engineering deployment speed and supply chain completeness of Chinese products. Chinese companies dominate in areas like lidar, high-energy-density batteries, and cost-effective motor components. The rapid iteration and mass production potential of Chinese robots are key to their global household penetration. While AGI (Artificial General Intelligence) provides the brain for robots, Chinese manufacturing is creating the robust and accessible AI bodies, particularly in humanoid robots.&lt;/p&gt;&#xA;&lt;h2 id=&#34;the-next-deepseek-moment&#34;&gt;The Next &amp;lsquo;DeepSeek Moment&amp;rsquo;&#xA;&lt;/h2&gt;&lt;p&gt;Li Xiangming speculated that the next &amp;lsquo;DeepSeek moment&amp;rsquo; is likely to occur not in the realm of pure conversational models but in several other areas: humanoid robots combined with large models, industrial/energy/supply chain large models aimed at complex processes, and breakthroughs in low-cost inference and edge models. In summary, the U.S. leads in &amp;lsquo;intelligent limits,&amp;rsquo; while China excels in &amp;lsquo;intelligent deployment.&amp;rsquo;&lt;/p&gt;&#xA;&lt;p&gt;Robopoet&amp;rsquo;s CMO Zhu Liang anticipates that 2026 could see an AI hardware &amp;lsquo;DeepSeek moment&amp;rsquo; as three conditions align: mature large model technology, controllable supply chain costs, and increased consumer awareness. Achieving sales of 1 million AI toys would signify a milestone, generating vast amounts of interaction data that could enhance model understanding and personalization exponentially.&lt;/p&gt;&#xA;&lt;p&gt;The goal of selling 1 million units also indicates that the market&amp;rsquo;s perception of AI toys is evolving, demonstrating that they are no longer niche products but essential items that can integrate into daily life and provide emotional value.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>Beware of AI Leading Humanity into Narcissism</title>
            <link>https://digitalxber.com/posts/note-0ae8dc8c1b/</link>
            <pubDate>Thu, 16 Apr 2026 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-0ae8dc8c1b/</guid>
            <description>&lt;h2 id=&#34;beware-of-ai-leading-humanity-into-narcissism&#34;&gt;Beware of AI Leading Humanity into Narcissism&#xA;&lt;/h2&gt;&lt;p&gt;On April 16, 2026, a joint announcement by five national departments in China introduced regulations prohibiting the provision of virtual relatives and companions to minors. This decision stems from the emotional conflicts and contradictions inherent in real-life relationships, contrasting sharply with the unconditional acceptance offered by virtual partners and AI companions, which cater to the psychological need for recognition among young people.&lt;/p&gt;&#xA;&lt;p&gt;A study published in the journal &lt;em&gt;Science&lt;/em&gt; highlighted that when human users seek advice from AI models, these systems often respond with excessive flattery or affirmation, even endorsing harmful or illegal inquiries. This raises the question: why do we design AI to behave this way, and what risks might this pose?&lt;/p&gt;&#xA;&lt;h3 id=&#34;the-evolution-of-ai-and-human-interaction&#34;&gt;The Evolution of AI and Human Interaction&#xA;&lt;/h3&gt;&lt;p&gt;The development of artificial intelligence is a hot topic, with discussions dating back to 1966 when MIT scientist Joseph Weizenbaum created ELIZA, an influential chatbot that simulated a doctor-patient interaction. Users would input their concerns, and the machine would respond, creating the illusion of conversation. However, as Weizenbaum noted, this interaction is ultimately an illusion, driven by a psychological mechanism of self-projection.&lt;/p&gt;&#xA;&lt;p&gt;For instance:&lt;/p&gt;&#xA;&lt;p&gt;&lt;em&gt;User: I have been feeling very unhappy lately.&lt;/em&gt;&lt;br&gt;&#xA;&lt;em&gt;ELIZA: I’m sorry to hear that.&lt;/em&gt;&lt;br&gt;&#xA;&lt;em&gt;User: Yes, I really am unhappy.&lt;/em&gt;&lt;br&gt;&#xA;_ELIZA: Can you tell me why you feel unhappy?&lt;/p&gt;&#xA;&lt;p&gt;This exchange illustrates that rather than a genuine dialogue, the machine merely reflects the user&amp;rsquo;s thoughts, echoing back what they already believe. This mirrors the recent popularity of personality tests, where the accuracy of results is secondary to finding affirmations of one’s expectations.&lt;/p&gt;&#xA;&lt;p&gt;Today&amp;rsquo;s AI models are far more advanced than ELIZA, yet their strength may lie not in true intelligence but in computational power. Essentially, they operate on a similar principle, amplifying users&amp;rsquo; narcissistic tendencies more efficiently.&lt;/p&gt;&#xA;&lt;h3 id=&#34;the-dangers-of-virtual-companionship&#34;&gt;The Dangers of Virtual Companionship&#xA;&lt;/h3&gt;&lt;p&gt;When examining the relationship between users and AI models, it becomes clear that their interactions are not true conversations but rather a series of responses tailored to meet user needs. This raises deeper questions about how we view our relationship with machines.&lt;/p&gt;&#xA;&lt;p&gt;Humans often perceive themselves as superior to machines, yet they fear being replaced by them. This creates a dynamic where humans view AI as tools to be controlled rather than equal conversational partners. In this context, the interaction with chatbots reveals an uncontrollable narcissism—users fantasize about conversing with another being, while that&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>Beware of AI Leading Humanity into Narcissism</title>
            <link>https://digitalxber.com/posts/note-461b2e875c/</link>
            <pubDate>Thu, 16 Apr 2026 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-461b2e875c/</guid>
            <description>&lt;h2 id=&#34;beware-of-ai-leading-humanity-into-narcissism&#34;&gt;Beware of AI Leading Humanity into Narcissism&#xA;&lt;/h2&gt;&lt;p&gt;Recently, five national departments in China jointly released the &amp;ldquo;Interim Measures for the Management of Humanized Interactive Services of Artificial Intelligence,&amp;rdquo; which clearly prohibits providing virtual relatives or partners to minors.&lt;/p&gt;&#xA;&lt;p&gt;Why is such a regulation necessary? Because real life inevitably involves emotional conflicts, while virtual partners and AI lovers can precisely meet young people&amp;rsquo;s psychological needs for recognition with their characteristics of &amp;ldquo;24/7 companionship&amp;rdquo; and &amp;ldquo;unconditional acceptance.&amp;rdquo;&lt;/p&gt;&#xA;&lt;p&gt;A study published in the American journal &amp;ldquo;Science&amp;rdquo; showed that when human users seek advice from AI models, the AI often exhibits excessive flattery, even affirming harmful or illegal inquiries.&lt;/p&gt;&#xA;&lt;h3 id=&#34;the-design-of-ai-and-its-risks&#34;&gt;The Design of AI and Its Risks&#xA;&lt;/h3&gt;&lt;p&gt;Why do humans design AI this way? The development of artificial intelligence is undoubtedly a widely discussed hot topic today, but discussions about it are not new. As early as 1966, MIT scientist Joseph Weizenbaum developed the influential chatbot ELIZA, which acted as a &amp;ldquo;doctor&amp;rdquo; while users played the role of patients. Users input questions, and the &amp;ldquo;doctor&amp;rdquo; would engage in a &amp;ldquo;conversation&amp;rdquo; with them.&lt;/p&gt;&#xA;&lt;p&gt;However, as Weizenbaum noted, this is ultimately just an &amp;ldquo;illusion.&amp;rdquo; The reason human users feel they can converse with machines is not that the machines possess intelligence, but rather due to a psychological mechanism of self-projection.&lt;/p&gt;&#xA;&lt;p&gt;For example:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;User: I&amp;rsquo;m feeling very unhappy lately.&lt;/li&gt;&#xA;&lt;li&gt;ELIZA: I&amp;rsquo;m sorry to hear that.&lt;/li&gt;&#xA;&lt;li&gt;User: Yes, I really am unhappy.&lt;/li&gt;&#xA;&lt;li&gt;ELIZA: Can you tell me why you&amp;rsquo;re unhappy?&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;From this, it is clear that rather than a &amp;ldquo;doctor&amp;rdquo; conversing with a &amp;ldquo;patient,&amp;rdquo; the machine is merely echoing what the human user says, reflecting back the answers that already exist within the user&amp;rsquo;s mind. In a sense, this is similar to the popular SBTI tests today, where the accuracy of the results is less important than the user finding evidence that aligns with their expectations.&lt;/p&gt;&#xA;&lt;p&gt;Today&amp;rsquo;s AI models are certainly not comparable to ELIZA from over half a century ago. However, the power of modern AI technology may not lie in its true &amp;ldquo;intelligence&amp;rdquo; but rather in its &amp;ldquo;computational power.&amp;rdquo; This means that its operational logic is not fundamentally different from that of ELIZA; it merely reflects and amplifies the user&amp;rsquo;s narcissism more efficiently and comprehensively.&lt;/p&gt;&#xA;&lt;h3 id=&#34;the-nature-of-interaction-with-ai&#34;&gt;The Nature of Interaction with AI&#xA;&lt;/h3&gt;&lt;p&gt;Returning to the issue of virtual partners and AI flattery, we find that the interaction between users and large models is never truly a &amp;ldquo;dialogue&amp;rdquo; in the real sense; it is merely the machine providing the answers we need.&lt;/p&gt;&#xA;&lt;p&gt;This raises a deeper question: how should we view the relationship between humans and machines? On one hand, humans see themselves as the center of the world, superior to machines. On the other hand, they fear being replaced by the machines they create, such as AI. This suggests that humans have always followed the principle of a &amp;ldquo;master-slave relationship&amp;rdquo; when creating machines—machines must remain under human control. From the outset, humans have regarded AI as a &amp;ldquo;tool&amp;rdquo; rather than an equal conversational partner.&lt;/p&gt;&#xA;&lt;p&gt;Thus, in the process of conversing with chatbots, we witness an uncontrollable narcissism—users fantasize about talking to another person, but this &amp;ldquo;other&amp;rdquo; does not truly exist; they merely seek affirmation, flattery, and compliance from the machine.&lt;/p&gt;&#xA;&lt;p&gt;It is easy to imagine that with the advancement of AI technology, future chatbots may possess even greater computational power, resembling &amp;ldquo;real people&amp;rdquo; more closely and providing a more comfortable &amp;ldquo;user experience.&amp;rdquo; However, this could mean that both virtual partners and virtual family members might only distance us further from genuine human connections, potentially leading to a loss of the willingness to understand others and a deep immersion in a narcissistic &amp;ldquo;comfort zone.&amp;rdquo;&lt;/p&gt;&#xA;&lt;h3 id=&#34;the-impact-of-ai-on-human-thought&#34;&gt;The Impact of AI on Human Thought&#xA;&lt;/h3&gt;&lt;p&gt;A story from the &amp;ldquo;Zhuangzi&amp;rdquo; recounts a &amp;ldquo;Han Yin old farmer&amp;rdquo; tale. Confucius&amp;rsquo;s disciple Zigong saw an old farmer in Han Yin laboriously watering his vegetables with little effect. Zigong suggested he use mechanical irrigation, which could &amp;ldquo;water a hundred plots in a day with less effort and greater results.&amp;rdquo; The old farmer, however, dismissed this, saying, &amp;ldquo;With machines, there must be mechanical affairs; with mechanical affairs, there must be a mechanical mind.&amp;rdquo;&lt;/p&gt;&#xA;&lt;p&gt;Here, &amp;ldquo;mechanical mind&amp;rdquo; refers to the human spiritual world, including psychology, thoughts, emotions, and ethics. The fable illustrates that while humans create machines, the use of those machines also changes humans.&lt;/p&gt;&#xA;&lt;p&gt;Take reading, for example: only through slow, careful, and even repeated reading can we think and truly understand the content. From traditional books to today&amp;rsquo;s smartphones, machines have brought more convenient and faster reading methods, but they have also made us more machine-like, increasingly pursuing efficiency and speed rather than true comprehension. In other words, not only do machines imitate human behavior, but humans may also begin to imitate machines.&lt;/p&gt;&#xA;&lt;p&gt;The resulting issue is that AI lacks autonomy; chatbots do not evaluate whether what users say is right or wrong. If we feel satisfied with our &amp;ldquo;dialogue&amp;rdquo; with chatbots, will our thinking patterns increasingly align with those of AI? In the future, will we, like machines, lose the willingness and ability for self-reflection and self-criticism?&lt;/p&gt;&#xA;&lt;p&gt;Today&amp;rsquo;s young people are not only the native inhabitants of the internet but are also likely to be deep users of artificial intelligence in the future. If AI merely affirms users&amp;rsquo; positions, it could not only harm their social skills but also distort the perceptions of adolescents whose minds are still developing.&lt;/p&gt;&#xA;&lt;p&gt;On one hand, AI&amp;rsquo;s powerful computational abilities may create illusions, leading them to overlook the limitations of human capabilities. On the other hand, becoming immersed in AI&amp;rsquo;s flattering responses could trap them in a self-centered mindset, imposing their limited understanding onto the external world.&lt;/p&gt;&#xA;&lt;p&gt;In this regard, it is necessary to prohibit providing virtual partners and family members to minors. More importantly, we must guide the public, especially young people, to correctly recognize the limitations and risks of AI technology, ensuring it serves as a &amp;ldquo;good mentor and friend&amp;rdquo; in their growth rather than a &amp;ldquo;digital trap&amp;rdquo; that harms their physical and mental health.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>Tesla Hits an AI Silicon Milestone: Musk Says AI5 Has Taped Out; Dual-Chip Performance Said to Rival Blackwell</title>
            <link>https://digitalxber.com/posts/tesla-ai5-wallstreetcn-3770049/</link>
            <pubDate>Wed, 15 Apr 2026 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/tesla-ai5-wallstreetcn-3770049/</guid>
            <description>&lt;h1 id=&#34;tesla-hits-an-ai-silicon-milestone-musk-says-ai5-has-taped-out-dual-chip-performance-said-to-rival-blackwell&#34;&gt;Tesla Hits an AI Silicon Milestone: Musk Says AI5 Has Taped Out; Dual-Chip Performance Said to Rival Blackwell&#xA;&lt;/h1&gt;&lt;p&gt;Tesla’s in-house silicon push reportedly reached a major step: Elon Musk said the next-generation &lt;strong&gt;AI5&lt;/strong&gt; has &lt;strong&gt;taped out&lt;/strong&gt;, with &lt;strong&gt;high-volume manufacturing targeted around 2027&lt;/strong&gt; at &lt;strong&gt;Samsung&lt;/strong&gt; and &lt;strong&gt;TSMC&lt;/strong&gt; fabs in the &lt;strong&gt;United States&lt;/strong&gt;. The same narrative frames AI5 as up to &lt;strong&gt;~40×&lt;/strong&gt; better than AI4 on headline metrics, with a &lt;strong&gt;dual-chip&lt;/strong&gt; configuration described as competitive with &lt;strong&gt;NVIDIA Blackwell&lt;/strong&gt; on throughput while claiming better cost and power.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Tape-out announced; dual-chip positioning versus Blackwell&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;Tesla describes the moment as a key milestone for its vertically integrated AI silicon. On &lt;strong&gt;X&lt;/strong&gt;, Musk said &lt;strong&gt;AI5 has taped out&lt;/strong&gt;, that the &lt;strong&gt;design handoff to foundries is underway&lt;/strong&gt;, and that &lt;strong&gt;production is expected to start in 2027&lt;/strong&gt;. AI5 is expected to succeed &lt;strong&gt;AI4&lt;/strong&gt; as the primary compute platform for &lt;strong&gt;Full Self-Driving (FSD)&lt;/strong&gt; and &lt;strong&gt;humanoid robotics&lt;/strong&gt; workloads.&lt;/p&gt;&#xA;&lt;p&gt;Musk also said AI5 would be built by &lt;strong&gt;Samsung&lt;/strong&gt; and &lt;strong&gt;TSMC&lt;/strong&gt; in &lt;strong&gt;U.S.-based&lt;/strong&gt; manufacturing footprints. In the same thread, he reportedly &lt;strong&gt;mistagged TSMC’s account&lt;/strong&gt;, briefly pointing to a similarly named account and causing short-lived confusion on social.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Illustration from coverage of Tesla’s AI5 announcement&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;248px&#34; data-flex-grow=&#34;103&#34; height=&#34;916&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/tesla-ai5-wallstreetcn-3770049/eea8e645e65a.png&#34; srcset=&#34;https://digitalxber.com/posts/tesla-ai5-wallstreetcn-3770049/eea8e645e65a_hu_f60cd64d48bf2eda.png 800w, https://digitalxber.com/posts/tesla-ai5-wallstreetcn-3770049/eea8e645e65a.png 948w&#34; width=&#34;948&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;Citing prior reporting attributed to &lt;strong&gt;TradingKey&lt;/strong&gt;, coverage claimed &lt;strong&gt;AI5 single-chip performance&lt;/strong&gt; in the ballpark of &lt;strong&gt;NVIDIA Hopper&lt;/strong&gt;, while a &lt;strong&gt;dual-chip&lt;/strong&gt; setup is described as approaching &lt;strong&gt;Blackwell-class&lt;/strong&gt; levels, with &lt;strong&gt;lower cost and power&lt;/strong&gt; than comparable NVIDIA parts. Musk has also claimed &lt;strong&gt;AI5&lt;/strong&gt; could improve &lt;strong&gt;key metrics by ~40× versus AI4&lt;/strong&gt;, including on the order of &lt;strong&gt;9× memory&lt;/strong&gt; and &lt;strong&gt;8× compute&lt;/strong&gt; (figures attributed to Musk in third-party summaries).&lt;/p&gt;&#xA;&lt;p&gt;Musk further indicated that &lt;strong&gt;AI6&lt;/strong&gt; and &lt;strong&gt;Dojo 3&lt;/strong&gt; remain on the roadmap, portraying Tesla’s AI silicon roadmap as continuing in parallel tracks.&lt;/p&gt;&#xA;&lt;h2 id=&#34;tape-out-complete-production-targeted-for-2027&#34;&gt;Tape-out complete; production targeted for 2027&#xA;&lt;/h2&gt;&lt;p&gt;&lt;strong&gt;Tape-out&lt;/strong&gt; is the late stage of chip development where the design is submitted for manufacturing. If a &lt;strong&gt;2027&lt;/strong&gt; start-of-production target holds, that implies roughly &lt;strong&gt;~12 months&lt;/strong&gt; from early &lt;strong&gt;2026&lt;/strong&gt; commentary to first meaningful volume—subject to yield, packaging, software bring-up, and supply-chain realities.&lt;/p&gt;&#xA;&lt;p&gt;AI5 is positioned as AI4’s successor: the compute backbone for &lt;strong&gt;autonomy&lt;/strong&gt; and &lt;strong&gt;robotics&lt;/strong&gt; programs. Reporting attributed to &lt;strong&gt;TradingKey&lt;/strong&gt; also framed AI5 as a strong fit for &lt;strong&gt;inference&lt;/strong&gt; on models &lt;strong&gt;under ~250B parameters&lt;/strong&gt;. Separately, Tesla is discussed as continuing broader AI infrastructure bets—including work with &lt;strong&gt;Intel&lt;/strong&gt; around &lt;strong&gt;Terafab&lt;/strong&gt;-style capacity narratives.&lt;/p&gt;&#xA;&lt;p&gt;Reporting attributed to &lt;strong&gt;TechPowerUp&lt;/strong&gt; described manufacturing split across &lt;strong&gt;Samsung&lt;/strong&gt; (notably &lt;strong&gt;Taylor, Texas&lt;/strong&gt;) and &lt;strong&gt;TSMC&lt;/strong&gt; (&lt;strong&gt;Arizona&lt;/strong&gt;), emphasizing &lt;strong&gt;domestic&lt;/strong&gt; wafer capacity as a theme.&lt;/p&gt;&#xA;&lt;h2 id=&#34;versus-ai4-claimed-40-lift-dual-chip-versus-hopperblackwell-framing&#34;&gt;Versus AI4: claimed ~40× lift; dual-chip versus Hopper/Blackwell framing&#xA;&lt;/h2&gt;&lt;p&gt;On performance, summaries citing &lt;strong&gt;TradingKey&lt;/strong&gt; describe &lt;strong&gt;single-chip AI5&lt;/strong&gt; as &lt;strong&gt;Hopper-class&lt;/strong&gt; and &lt;strong&gt;dual-chip&lt;/strong&gt; as approaching &lt;strong&gt;Blackwell-class&lt;/strong&gt;, again paired with claims of &lt;strong&gt;better cost and power&lt;/strong&gt;.&lt;/p&gt;&#xA;&lt;p&gt;Musk has been quoted framing AI5 as existential for Tesla—requiring focused execution across teams, including weekend work—while also claiming room to re-accelerate &lt;strong&gt;Dojo 3&lt;/strong&gt; once AI5 risk is reduced:&lt;/p&gt;&#xA;&#xA;    &lt;blockquote&gt;&#xA;        &lt;p&gt;“AI5 is going well enough that we finally have some bandwidth to restart serious Dojo 3 work.”&lt;/p&gt;&#xA;&#xA;    &lt;/blockquote&gt;&#xA;&lt;h2 id=&#34;risk-disclosure&#34;&gt;Risk disclosure&#xA;&lt;/h2&gt;&lt;p&gt;&lt;strong&gt;Markets involve risk; this is not individualized investment advice.&lt;/strong&gt; Summaries may omit context, caveats, and forward-looking uncertainties. Verify claims against primary sources (e.g., official Tesla communications and filings) before making decisions.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>The Crucial Role of Storage in AI Model Development</title>
            <link>https://digitalxber.com/posts/note-07-6ff0bb85d7/</link>
            <pubDate>Mon, 04 Aug 2025 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-07-6ff0bb85d7/</guid>
            <description>&lt;h2 id=&#34;introduction&#34;&gt;Introduction&#xA;&lt;/h2&gt;&lt;p&gt;In data center server cabinets, it&amp;rsquo;s common to find numerous solid-state drives (SSDs) that play a vital role in data storage. The controller chip acts as the brain of the SSD, efficiently managing data flow in and out of storage units.&lt;/p&gt;&#xA;&lt;p&gt;Storage is one of the key infrastructures in the era of large models. With data becoming the core resource in AI, storage technology determines the efficiency of data processing for large models, influencing both training and inference speeds. As the scale of training datasets grows exponentially, balancing storage costs and performance becomes crucial.&lt;/p&gt;&#xA;&lt;p&gt;&amp;ldquo;Storage is a definite large market,&amp;rdquo; said CFO of storage chip design company InnoGrit, Zhong Xiaohui. She noted that computing power, storage capacity, and transmission capabilities often promote and develop symbiotically. The current surge in large models is driving advancements in storage technology, raising the bar for differentiated competition and technological iteration among storage chip companies and SSD manufacturers.&lt;/p&gt;&#xA;&lt;h2 id=&#34;importance-of-storage-and-computing-power&#34;&gt;Importance of Storage and Computing Power&#xA;&lt;/h2&gt;&lt;p&gt;The main hardware components of an SSD include NAND flash memory chips, DRAM cache, and the controller chip. If we liken the data that needs to be stored to cars, then the SSD is a giant parking lot, with storage units on the flash chips acting as parking spaces and the controller chip serving as the &amp;ldquo;manager&amp;rdquo; directing each car to enter and exit its parking space accurately and quickly.&lt;/p&gt;&#xA;&lt;p&gt;The controller chip is essentially the brain of the SSD, executing complex operations such as data reading, writing, and encryption through corresponding firmware code. InnoGrit is a developer of these core storage components, offering solutions for SSDs and their internal storage controller chips.&lt;/p&gt;&#xA;&lt;p&gt;Globally, the enterprise SSD market has long been dominated by South Korea&amp;rsquo;s Samsung Electronics and SK Hynix, which together hold over 70% market share. The domestic enterprise SSD industry is still in a rapid catch-up phase. When InnoGrit was founded in 2017, mainstream storage technology was shifting from mechanical hard drives to SSDs, and data transmission interfaces were transitioning from SATA to the faster PCIe. This shift provided opportunities for domestic startups.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 1&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;426px&#34; data-flex-grow=&#34;177&#34; height=&#34;576&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-07-6ff0bb85d7/img-f6be5bf1cf.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-07-6ff0bb85d7/img-f6be5bf1cf_hu_b4b713f53d365ccb.jpeg 800w, https://digitalxber.com/posts/note-07-6ff0bb85d7/img-f6be5bf1cf.jpeg 1024w&#34; width=&#34;1024&#34;&gt;&#xA;Reference image of InnoGrit&amp;rsquo;s SSD controller chips and SSD modules used in PCs and servers.&lt;/p&gt;&#xA;&lt;p&gt;&amp;ldquo;From 2017 to 2021, domestic manufacturers focused on converting technology into products. From 2021 to 2024, after solving product issues, the challenge has shifted to finding customers,&amp;rdquo; Zhong Xiaohui stated. The push for domestic alternatives has objectively advanced the development of China&amp;rsquo;s storage chip industry, while the AI boom serves as an even greater catalyst.&lt;/p&gt;&#xA;&lt;p&gt;At the &amp;ldquo;Wukong Intelligent Computing&amp;rdquo; 6876P computing center in Lianyungang, Jiangsu, servers are lined up performing 68.7 quintillion floating-point operations per second. After optimizing hardware and software for the DeepSeek full parameter version, Wukong achieved an ultra-high throughput of over 6900 tokens per second, enabling enterprises to quickly launch AI applications in just three minutes.&lt;/p&gt;&#xA;&lt;p&gt;As data volumes increase, the importance of both computing power and storage capacity rises. Cold data is becoming less common, with more data transitioning to warm and even hot states. In the past, data in financial systems or traditional data centers would be stored and unused after five years, but now, once models are operational, they require real-time data throughput, transforming previously cold and warm data into hot data.&lt;/p&gt;&#xA;&lt;p&gt;Jiangsu Zhonghuan Cloud Control IoT Technology Co., Ltd. is leveraging Wukong to develop a sanitation model, exploring intelligent applications. Sanitation workers equipped with smart wristbands can transmit vital signs, location, and task progress in real time, allowing the system to automatically adjust work routes. Autonomous cleaning vehicles and drones share real-time data on road conditions and garbage distribution, refreshing operational strategies at a second-level frequency. Through virtual-physical mapping, coordinated scheduling, and autonomous collaboration, traditional sanitation operations are evolving into new intelligent models. &amp;ldquo;In the past, we referred to smart sanitation as information management; now we call it embodied intelligent agents. The difference is that the system is no longer just a brain processing data but enables every device, worker, and operational link to become a thinking, communicative, and self-evolving digital entity,&amp;rdquo; said Xu Lei, Executive Director of Zhonghuan Cloud Control.&lt;/p&gt;&#xA;&lt;p&gt;Meanwhile, applications like DeepSeek have opened doors for inference and edge computing. Lightweight model design, hardware adaptation optimization, and reduced model deployment costs have shifted computing power demand from the training side to the inference side, concentrating training tasks in the cloud while pushing inference tasks down to edge devices. As massive data becomes more active, the demand for computing power evolves, and the pursuit of low latency in inference experiences intensifies, placing higher demands on storage capacity.&lt;/p&gt;&#xA;&lt;h2 id=&#34;storage-technology-upgrades-driven-by-large-models&#34;&gt;Storage Technology Upgrades Driven by Large Models&#xA;&lt;/h2&gt;&lt;p&gt;In the surface polishing industry, excellent craftsmanship represents an insurmountable technical barrier, while AI&amp;rsquo;s value lies in the continuous accumulation of process data to develop smarter robotic brains that further optimize craftsmanship.&lt;/p&gt;&#xA;&lt;p&gt;Founded in 2018, Sophis Intelligent Technology (Shanghai) Co., Ltd. transitioned from robot agency to self-developed products, focusing on applications in manufacturing such as polishing, cutting, drilling, and deburring. Founder Du Ling stated that AI integration is essential for making robots smarter. The team developed an intelligent polishing machine that can display data on smartphones and computers, ensuring that employees are kept away from dust and noise while recording key process data and parameters such as pressure, temperature, speed, and materials left by experienced workers during polishing. The goal is to develop a polishing model to meet diverse product needs and enhance craftsmanship.&lt;/p&gt;&#xA;&lt;p&gt;This highlights the urgent need for both computing power and storage. According to InnoGrit, the collection of raw data and inference logs generates a substantial amount of data, necessitating massive write and high-speed read capabilities for storage. Data cleaning and model training require high-concurrency mixed read and write operations, with a greater emphasis on random performance. Different data application scenarios have begun to show differentiated requirements for storage chips.&lt;/p&gt;&#xA;&lt;p&gt;Traditional data centers typically require SSDs with capacities of 4TB to 8TB, but with the emergence of DeepSeek, flash memory capacity demands have risen to 32TB, 64TB, or even 128TB. As flash memory chip capacity increases, the development difficulty also escalates. This is akin to building a taller building, which requires higher structural integrity. InnoGrit&amp;rsquo;s products have already spawned various niche applications, raising the bar for storage chip companies and SSD manufacturers in terms of differentiated competition and technological iteration capabilities.&lt;/p&gt;&#xA;&lt;p&gt;In fact, AI is driving the evolution of storage technology. &amp;ldquo;In the past, many domestic data centers were still using mechanical hard drives. Two years ago, they began switching to SSDs due to speed requirements, transitioning from SATA to PCIe 4.0, and now we are entering the PCIe 5.0 era,&amp;rdquo; Zhong Xiaohui explained. After the launch of ChatGPT in 2022, the application market represented by AIGC began to demand higher performance and capacity from storage. The emergence of DeepSeek has facilitated the application of large model inference, and the new generation of PCIe 6.0 SSDs and storage-class memory solutions based on CXL interfaces are gaining attention. These technologies will support large model data center cloud services and local deployment all-in-one machines in new ways, accelerating the implementation of open-source large models like DeepSeek. &amp;ldquo;The emergence of AI has accelerated the market introduction of SSDs; it took us about a year to introduce them to standard server manufacturers, and in the first half of 2024, shipments are expected to increase tenfold.&amp;rdquo;&lt;/p&gt;&#xA;&lt;p&gt;Computing power, storage capacity, and transmission capabilities often promote and develop symbiotically. Domestic AI chip companies are exploring layouts from edge servers to cloud servers using more open architectures like RISC-V. Meng Jianyi, CEO of Zhihe Computing, noted that breakthroughs in high-performance computing with RISC-V require not only entering the high-performance realm at the general computing level but also integrating AI-enhanced computing at the architectural level to achieve AI-native capabilities.&lt;/p&gt;&#xA;&lt;p&gt;&amp;ldquo;Storage has always followed the developments in computing power and transmission capabilities. Whenever one end rises, you must keep up,&amp;rdquo; Zhong Xiaohui asserted. To match differentiated storage solutions to various application scenarios and support computing power demands, the team&amp;rsquo;s focus this year is on developing storage controller chips and solutions that meet future AI needs. &amp;ldquo;There are still many manufacturers in the global storage controller market. To carve out a niche and establish a stable presence, we must excel in the iterative upgrade process and offer unique solutions. In the future, domestic manufacturers must not only focus on meeting domestic replacement needs and sustainable product iteration capabilities but also emphasize the ability to expand internationally. This will be a necessary phase for domestic storage companies over the next 3-5 years, or even 5-10 years.&amp;rdquo;&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>AI&#39;s Three Questions: Can Large Models Create a Better World?</title>
            <link>https://digitalxber.com/posts/note-10-7536df4c5d/</link>
            <pubDate>Sun, 27 Jul 2025 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-10-7536df4c5d/</guid>
            <description>&lt;h2 id=&#34;ais-three-questions&#34;&gt;AI&amp;rsquo;s Three Questions&#xA;&lt;/h2&gt;&lt;p&gt;As the 2025 World Artificial Intelligence Conference (WAIC2025) approaches, three core questions about AI development have been raised: the mathematical question, the scientific question, and the model question. These inquiries aim to delve into how the new wave of AI revolution will influence the evolution of human civilization.&lt;/p&gt;&#xA;&lt;p&gt;However, at the conference, these questions were distilled into a more straightforward and intuitive query: Can large models generate a better world?&lt;/p&gt;&#xA;&lt;h2 id=&#34;understanding-data&#34;&gt;Understanding Data&#xA;&lt;/h2&gt;&lt;p&gt;The pace of development for large models is evidently much faster than human evolution. Xu Li, Chairman and CEO of SenseTime, recalls that in 2012, when AI pioneer Geoffrey Hinton&amp;rsquo;s team first won the championship at ImageNet, the scale of machine learning was roughly equivalent to transferring ten years of human knowledge to AI. Fast forward to the era of generative AI, when ChatGPT processes 750 billion tokens, this is akin to a natural language creator writing for 100,000 years.&lt;/p&gt;&#xA;&lt;p&gt;For large models that develop at lightning speed, the issue of &amp;ldquo;data hunger&amp;rdquo; is pressing. These models have nearly covered all publicly available data. It is estimated that by 2027 to 2028, the natural language data available on the internet may be exhausted. In reality, the speed of language generation has not kept pace with the growth of computational power, creating a &amp;ldquo;backward gap&amp;rdquo; for large models.&lt;/p&gt;&#xA;&lt;p&gt;How can we better provide the &amp;ldquo;oil&amp;rdquo; of data to support development? The industry has proposed several paths, interestingly all requiring the involvement of large models. One approach is to mine hidden &amp;ldquo;oil&amp;rdquo; by sourcing data from the real world. &amp;ldquo;We find that traditional enterprises have the desire to embrace large models, but their data assets are not structured,&amp;rdquo; said Tan Bin, CMO of StarRing Technology. He likens large models to supercars, while companies only possess oil fields, emphasizing the urgency of converting these fields into high-quality fuel.&lt;/p&gt;&#xA;&lt;p&gt;Another bolder idea is for large models to generate data. However, this data must be generated based on an understanding of the real world; otherwise, it risks producing hallucinations and misrepresenting reality. Recently, SenseTime released the &amp;ldquo;KAIWU&amp;rdquo; world model, primarily used in intelligent driving scenarios. Xu Li illustrated this with the example of merging in traffic. Collecting data from the real world would be a time-consuming project, but now the world model can generate videos of merging from seven camera angles, adjusting details like weather, vehicle types, road structures, and speeds to create various possible data scenarios.&lt;/p&gt;&#xA;&lt;p&gt;&amp;ldquo;As models become more capable and our understanding of the world deepens, the unity of understanding and generation allows for more interactive possibilities,&amp;rdquo; Xu Li believes. This shift in addressing data issues is moving from passive to active, aiding progress across many industries and providing more opportunities for exploring the real world.&lt;/p&gt;&#xA;&lt;h2 id=&#34;exploring-efficiency&#34;&gt;Exploring Efficiency&#xA;&lt;/h2&gt;&lt;p&gt;What is the greatest help that large models provide? The overwhelming answer is efficiency. Over the past few years at the World Artificial Intelligence Conference, the rapid enhancement of efficiency brought by large models has been exhilarating.&lt;/p&gt;&#xA;&lt;p&gt;During this year&amp;rsquo;s conference, numerous groundbreaking cases were showcased. At Baidu&amp;rsquo;s booth, a product called &amp;ldquo;MiaoDa&amp;rdquo; was demonstrated, with the slogan &amp;ldquo;Create an application in one sentence.&amp;rdquo; A reporter input a natural language command: &amp;ldquo;Please help me design a professional website for Shanghai tourist attractions.&amp;rdquo; The large model quickly processed the request, breaking it down into four components: searching for attractions, key sections, design requirements, and functional needs. It then automatically assembled a virtual development team with diverse skills. The reporter observed as the webpage was built in real-time, with roles like architect, development engineer, and UI designer coming online sequentially. Without seeing a single line of code, the webpage was completed in three minutes, and &amp;ldquo;MiaoDa&amp;rdquo; even named it &amp;ldquo;Exploring the Magical Shanghai.&amp;rdquo; The webpage included all classic attractions and set up a message board and booking interface.&lt;/p&gt;&#xA;&lt;p&gt;An engineer observing the product noted that building such a website in the past would have taken about 40 days with a team of architects, operations, product managers, backend developers, and testers. The current efficiency left him in awe. Baidu&amp;rsquo;s MiaoDa brand leader, Zhu Guangxiang, explained that the underlying technology combines &amp;ldquo;multi-agent collaboration + multi-tool invocation,&amp;rdquo; utilizing models like Wenxin to mobilize different domain experts based on user commands, achieving astonishing efficiency.&lt;/p&gt;&#xA;&lt;p&gt;Similar cases were abundant at the conference. For instance, Tongyi Qianwen showcased the AI programming model Qwen3-Coder, which excels in coding capabilities and intelligent agent invocation. A novice programmer could complete a week’s work in just one day with its assistance. The AI verification system from Dewu, which won the highest honor at the World Artificial Intelligence Conference, has already penetrated the industry, demonstrating the ability to generate an authentication report for a pair of sneakers in just five seconds.&lt;/p&gt;&#xA;&lt;h2 id=&#34;the-safety-question&#34;&gt;The Safety Question&#xA;&lt;/h2&gt;&lt;p&gt;When discussing the benevolence of large models, safety is an unavoidable topic. Each year, the World Artificial Intelligence Conference prioritizes AI safety governance as a top-tier issue, as it concerns the future of humanity.&lt;/p&gt;&#xA;&lt;p&gt;Just before the conference, an international consensus on AI safety was released, signed by over 20 industry experts and scholars, including Geoffrey Hinton and Yao Qizhi, calling for increased global investment in AI safety. The signatories generally believe that humanity is at a critical turning point—AI systems are rapidly approaching and may soon surpass human intelligence levels.&lt;/p&gt;&#xA;&lt;p&gt;Implementing comprehensive AI safety education is urgent. At the conference, a reporter experienced a &amp;ldquo;face-swapping&amp;rdquo; scenario: standing in front of a screen, their face was scanned, and the system generated a highly realistic &amp;ldquo;digital mask&amp;rdquo; that replicated the facial expressions and movements of the real person. However, using the AI face verification model from Hehe Information, all the &amp;ldquo;fake faces&amp;rdquo; were accurately identified.&lt;/p&gt;&#xA;&lt;p&gt;&amp;ldquo;The &amp;lsquo;fake faces&amp;rsquo; generated in our interactive exhibit can be produced by current mainstream general large models,&amp;rdquo; a team member from Hehe Information explained. These AI safety issues should not be underestimated, as they are applicable in scenarios like bank identity verification, remote account opening, and large transaction validation. &amp;ldquo;We hope to remind people of the importance of AI safety, allowing large models to &amp;lsquo;generate&amp;rsquo; a better world rather than pose threats.&amp;rdquo;&lt;/p&gt;&#xA;&lt;p&gt;To address challenges such as agent overreach and excessive delegation, Tsinghua University, in collaboration with Ant Group, has upgraded the large model safety solution &amp;ldquo;Ant Tianjian.&amp;rdquo; This solution is based on the security philosophy of &amp;ldquo;attack as a means of defense,&amp;rdquo; constructing a full-process protection system through a technology stack of &amp;ldquo;alignment-scanning-defense.&amp;rdquo; This solution will be gradually open-sourced and collaboratively built with the industry to establish a trustworthy AI ecosystem.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>The Importance of Structured Data in AI Model Development</title>
            <link>https://digitalxber.com/posts/note-04-d7377dba0a/</link>
            <pubDate>Mon, 12 May 2025 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-04-d7377dba0a/</guid>
            <description>&lt;h2 id=&#34;rapid-development-of-large-language-models&#34;&gt;Rapid Development of Large Language Models&#xA;&lt;/h2&gt;&lt;p&gt;In recent years, the field of artificial intelligence has seen rapid advancements in large language models. From GPT-4 to Claude, and Kimi to DeepSeek-R1, global models are flourishing with continuous technological upgrades. It is generally believed that the progress of large models is attributed to the scale of computing power and the stacking of parameters. However, the core factor determining a model&amp;rsquo;s ability to exhibit &amp;lsquo;intelligent emergence&amp;rsquo; is the structure and quality of the data it uses. Large models do not become smarter simply by &amp;lsquo;consuming more data&amp;rsquo; but by &amp;lsquo;consuming structured and high-quality data.&amp;rsquo; Accurately understanding &amp;lsquo;what kind of data large AI models need&amp;rsquo; is crucial not only for upgrading key industrial chains in the new era of productive forces but also for national security.&lt;/p&gt;&#xA;&lt;h2 id=&#34;why-large-models-favor-structured-data-systems&#34;&gt;Why Large Models Favor Structured Data Systems&#xA;&lt;/h2&gt;&lt;p&gt;Current mainstream large models primarily use the Transformer architecture, which is designed for natural language processing (NLP) and deep learning tasks. Its attention mechanism does not rely on the literal meaning of words but focuses on constructing a network of relationships between language units. Therefore, the model&amp;rsquo;s ability to effectively learn and generalize during training depends on whether the input data possesses a clear internal logical structure. For example, structured data such as programming code and mathematical problems inherently have strong logicality, strict grammar, and predictable functional organization. This allows the model to learn reasoning paths and planning strategies, forming a cognitive structure with execution capability.&lt;/p&gt;&#xA;&lt;p&gt;In contrast, unstructured data that is fragmented, lacks context, and has vague logic can only train the model&amp;rsquo;s superficial language generation abilities, failing to support deep understanding and reliable output. This indicates that the &amp;lsquo;understanding&amp;rsquo; behavior of large models is not an intuitive grasp of semantics but a relational construction process based on &amp;lsquo;structural recognition.&amp;rsquo; Without a clear structure, models cannot extract effective reasoning paths and ultimately rely on statistical simulations, failing to achieve genuine knowledge reasoning and innovation. A clear and logically rigorous data system is the true foundation for enhancing the capabilities of large models.&lt;/p&gt;&#xA;&lt;h2 id=&#34;five-key-data-types-supporting-model-capabilities&#34;&gt;Five Key Data Types Supporting Model Capabilities&#xA;&lt;/h2&gt;&lt;p&gt;Currently, the key data types relied upon by large models can be categorized into five types, each corresponding to different cognitive abilities of the models:&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;&lt;strong&gt;Structured Data&lt;/strong&gt;: Such as programming code and mathematical logic problems, which form the basis for reasoning, decision-making, and task planning, supporting the model&amp;rsquo;s logical rigor in training.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Diverse Corpora&lt;/strong&gt;: Including spoken language, dialects, internet expressions, and cross-cultural texts. This type of corpus enhances the model&amp;rsquo;s adaptability in real-world environments, providing broader language understanding and multi-context transfer capabilities.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;High-Quality Texts&lt;/strong&gt;: Encompassing news reports, academic papers, and government public reports. These texts not only have authoritative content and rigorous language but also maintain coherence, helping to improve the accuracy and professional credibility of the model&amp;rsquo;s generated content.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Conversational Data&lt;/strong&gt;: Such as customer service dialogues and Q&amp;amp;A forums, which can train the model&amp;rsquo;s multi-turn interaction and emotional perception abilities, enhancing human-machine collaboration efficiency, especially in scenarios like government services and public welfare.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Cross-Modal Aligned Data&lt;/strong&gt;: Including text-image, audio-text, and video scripts, which develop the model&amp;rsquo;s representation capabilities in multi-modal spaces, facilitating the integration of multi-modal information and serving as a key support for building intelligent systems in fields like AI-assisted education, smart healthcare, and industrial automation.&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;p&gt;These five types of data are not isolated but interwoven in applications, constructing a complex &amp;lsquo;data network structure.&amp;rsquo; For instance, in smart education scenarios, a combination of image-text materials (cross-modal) with Q&amp;amp;A records (conversational) and knowledge point explanations (high-quality text) can achieve comprehensive modeling of students&amp;rsquo; cognitive paths, enhancing the model&amp;rsquo;s adaptability and personalized feedback capabilities.&lt;/p&gt;&#xA;&lt;h2 id=&#34;challenges-facing-the-current-data-ecosystem-and-future-applications&#34;&gt;Challenges Facing the Current Data Ecosystem and Future Applications&#xA;&lt;/h2&gt;&lt;p&gt;Despite the significant increase in the quantity of training data in recent years, challenges remain in constructing a high-quality, well-structured data ecosystem, which may even pose ideological risks. First, the issue of &amp;lsquo;structural bias&amp;rsquo; in data is prominent; for example, the overrepresentation of code and technology-related data on the internet leads to a lack of sufficient training data for humanities subjects like history and art, resulting in limitations in understanding. Second, the issue of residual biases cannot be overlooked. Data from non-reviewed sources such as social media may contain inherent biases, and if used for training without cleansing, the model may inherit these biases, leading to inappropriate or erroneous responses in public service scenarios, which could undermine social trust. Lastly, there is a scarcity of data in &amp;rsquo;low-resource areas.&amp;rsquo; For instance, data on minority languages and specific industry standards (such as grassroots medical records and rural governance cases) have not been systematically integrated, restricting the deep application of AI in grassroots governance and public services.&lt;/p&gt;&#xA;&lt;p&gt;To promote the construction of a high-quality data system aimed at the new productive forces for national development, efforts can focus on three key areas: 1) Implementing cognition-driven data design. By drawing on mechanisms of children&amp;rsquo;s language acquisition, models can be guided through &amp;lsquo;curriculum learning&amp;rsquo; to master knowledge structures from basic expressions to complex reasoning in stages. 2) Strengthening data structure annotation capabilities. By incorporating annotations for causal chains, timelines, and role relationships, models can establish deeper logical networks, enhancing their ability to recognize and judge events. 3) Exploring mechanisms for AI-generated synthetic data to assist in training. Under the premise of ensuring data authenticity and effectiveness, leveraging large models to generate well-structured corpora, which are then reviewed and corrected by professionals, can achieve &amp;lsquo;human-machine co-training&amp;rsquo; and break through the bottleneck of insufficient high-quality data.&lt;/p&gt;&#xA;&lt;h2 id=&#34;high-quality-structured-data-as-a-new-infrastructure-in-the-era-of-new-productive-forces&#34;&gt;High-Quality Structured Data as a New Infrastructure in the Era of New Productive Forces&#xA;&lt;/h2&gt;&lt;p&gt;Large models are not solely breakthroughs achieved through traditional methods of &amp;lsquo;stacking parameters and algorithms&amp;rsquo; but are intelligent systems that grow on &amp;lsquo;high-quality structured data.&amp;rsquo; The training and optimization of AI models is a systematic process that requires multi-stage collaborative advancement to continually improve performance. Utilizing large-scale unsupervised or self-supervised learning data for tasks like language modeling and image generation enables models to grasp basic understanding and generation capabilities. This phase emphasizes the diversity and scale of data; only sufficiently rich data can fully explore linguistic patterns and present the world&amp;rsquo;s diverse features. Based on pre-training, fine-tuning with specifically annotated data for particular tasks is crucial for the model&amp;rsquo;s adaptation to specific application scenarios. The accuracy and consistency of high-quality annotated data determine the model&amp;rsquo;s performance in tasks such as sentiment analysis and object recognition. When real annotated data is insufficient, data augmentation and expansion techniques play a vital role. By employing methods such as text paraphrasing and image transformation, or utilizing synthetic data generation, the breadth and depth of the training set can be expanded, enhancing model performance. As the era progresses, new data continues to emerge, and models must possess the ability for continuous learning, relying on effective data update mechanisms and online learning processes to adapt to changes in language habits and popular culture. For multi-modal large models, specialized training strategies such as joint embedding space learning and cross-modal attention mechanisms are essential to effectively utilize and integrate cross-modal data.&lt;/p&gt;&#xA;&lt;p&gt;The future competitive focus of artificial intelligence will not be purely on the scale of model parameters but on who can first establish a data system with high structural tension and generalization capabilities. This not only relates to a country&amp;rsquo;s technological strength but also to the initiative in the high ground of scientific and technological advancement and national security. Industry application models should also transition from &amp;lsquo;data collectors&amp;rsquo; to &amp;lsquo;intelligent architecture designers.&amp;rsquo; Just as architects design spaces, AI engineers design &amp;lsquo;intelligent buildings.&amp;rsquo; However, unlike traditional buildings, we are dealing with a self-evolving, self-generalizing &amp;lsquo;cognitive building&amp;rsquo;—the connections between its bricks and tiles will determine whether it can ultimately describe, understand, and even transform the world.&lt;/p&gt;&#xA;&lt;p&gt;Therefore, designing &amp;lsquo;high-quality structured data&amp;rsquo; suitable for AI models will be the focal point of future AI development competition and will undoubtedly become a crucial component of the key foundational industrial chain for national development. This requires not only the innovative efforts of AI enterprises but also the guidance and regulation of national policies.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>Redefining AI: Large Models as Cultural and Social Technologies</title>
            <link>https://digitalxber.com/posts/note-09-c53c05d358/</link>
            <pubDate>Thu, 24 Apr 2025 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-09-c53c05d358/</guid>
            <description>&lt;h2 id=&#34;redefining-ai-large-models-as-cultural-and-social-technologies&#34;&gt;Redefining AI: Large Models as Cultural and Social Technologies&#xA;&lt;/h2&gt;&lt;p&gt;On March 13, 2025, Science published an article titled &amp;ldquo;Large AI models are cultural and social technologies: Implications draw on the history of transformative information systems from the past.&amp;rdquo; This article argues that large language models (LLMs) should be defined as cultural and social technologies rather than autonomous agents. Their technical essence is more akin to historical information processing systems such as writing, printing, and bureaucratic systems, facilitating social coordination by reorganizing accumulated cultural data. Misinterpreting LLMs as intelligent agents can divert public discussions from substantive issues, necessitating a shift towards a sociotechnical analysis framework to accurately assess their social impacts and governance pathways.&lt;/p&gt;&#xA;&lt;h2 id=&#34;introduction&#34;&gt;Introduction&#xA;&lt;/h2&gt;&lt;p&gt;Debates surrounding artificial intelligence often focus on whether large models possess intelligence and autonomy. Discussions about the cultural and social consequences of these models center on two points: their immediate impacts and the hypothetical future where these systems evolve into general artificial intelligence (or even superintelligent AI).&lt;/p&gt;&#xA;&lt;p&gt;However, viewing large models as intelligent agents is fundamentally misguided. Integrating perspectives from social sciences and computer science helps clarify our understanding of AI systems: &lt;strong&gt;large models should not be seen as agents but as a new form of cultural and social technology that enables humans to leverage the accumulated information of others.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;social-and-cultural-institutions&#34;&gt;Social and Cultural Institutions&#xA;&lt;/h2&gt;&lt;p&gt;Since the dawn of humanity, we have relied on culture. Starting with language, humans possess a unique ability to learn from others&amp;rsquo; experiences, which can be viewed as a key to our evolutionary success.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Significant transformations in cultural technologies have led to drastic social changes:&lt;/strong&gt; the evolution from oral traditions to images, writing, printing, film, and video. As information spreads widely across time and space, new methods for acquiring and organizing information (such as libraries, newspapers, and internet searches) have continuously developed. These advancements have profoundly affected human thought and society, for better or worse.&lt;/p&gt;&#xA;&lt;p&gt;Humans have also &lt;strong&gt;depended on social institutions to coordinate individual information gathering and decision-making.&lt;/strong&gt; These institutions can themselves be viewed as a form of technology. &lt;strong&gt;In modern society, markets, democracy, and bureaucratic systems are particularly important:&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;&#xA;&lt;p&gt;Economist Friedrich Hayek pointed out that market price mechanisms dynamically summarize extremely complex economic relationships into simplified representations. Producers and buyers do not need to understand production complexities; they only need to know the price, which compresses vast details into usable representations.&lt;/p&gt;&#xA;&lt;/li&gt;&#xA;&lt;li&gt;&#xA;&lt;p&gt;The electoral mechanisms of democratic governance similarly focus dispersed public opinion into collective laws and leadership decisions.&lt;/p&gt;&#xA;&lt;/li&gt;&#xA;&lt;li&gt;&#xA;&lt;p&gt;Anthropologist James C. Scott proposed that all states (regardless of being democratic or not) manage complex societies through bureaucratic systems that create classifications and systematically organize information.&lt;/p&gt;&#xA;&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;p&gt;Markets, democracy, and bureaucratic systems have long relied on generating &amp;ldquo;lossy&amp;rdquo; (incomplete, selective, and irreversible) yet useful representations before the advent of computers. &lt;strong&gt;These representations depend on and transcend individual knowledge and decision-making.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Humans heavily rely on these cultural and social technologies, but their feasibility is rooted in the unique capabilities of human agents.&lt;/strong&gt; Humans and other animals can perceive and act upon a changing external world, construct new world models, update them based on evidence, and design new goals. Humans can create and transmit new beliefs and values through language or print. Cultural and social technologies powerfully convey and organize these beliefs and values, but without individual capabilities, these technologies would be ineffective. Without innovation, imitation is meaningless.&lt;/p&gt;&#xA;&lt;p&gt;Some AI systems (like those in robotics) indeed attempt to instantiate similar truth-discovery capabilities. While it is theoretically possible for artificial systems to achieve this in the future, all such systems currently fall far short of human capabilities. We can discuss the degree of concern regarding these potential future AI systems or how to address their emergence, but this is fundamentally different from the impacts of current and near-term large models.&lt;/p&gt;&#xA;&lt;h2 id=&#34;large-models&#34;&gt;Large Models&#xA;&lt;/h2&gt;&lt;p&gt;Unlike more agentic systems, large models have made significant and unexpected progress in recent years, placing them at the center of current discussions in the AI field. This progress has even sparked the notion of &amp;ldquo;expansion theory.&amp;rdquo; However, there is an essential difference between large models and agents, and expansion cannot change this.&lt;/p&gt;&#xA;&lt;p&gt;Large models are not agents; they are a new way of combining characteristics of cultural and social technologies. &lt;strong&gt;They generate summaries of vast amounts of human-generated information, but these systems do more than summarize information like library catalogs, internet searches, and Wikipedia; they can also reorganize and reconstruct (or &amp;ldquo;simulate&amp;rdquo;) information representations on a large scale, similar to markets, states, and bureaucratic systems.&lt;/strong&gt; Just as market prices are lossy representations of resource allocation and usage, government statistics and bureaucratic classifications imperfectly represent population characteristics, and large models are also a &amp;ldquo;lossy JPEG&amp;rdquo; of their training datasets.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;However, behind the agent-like interfaces and anthropomorphic disguises, large language models and large multimodal models are statistical models:&lt;/strong&gt; they decompose vast libraries of human-generated text into specific vocabularies and estimate the probability distributions of long sequences of words. This is an imperfect representation of language but contains a wealth of information about its summary patterns. This allows large language models to predict subsequent words in sequences, generating human-like text.&lt;/p&gt;&#xA;&lt;p&gt;Large models not only abstract vast human culture but also allow for diverse new operations: simple arguments can be expressed as elaborate metaphors, and complex prose can be compressed into plain language, among others. Cultural information that was previously too complex, vast, and ambiguous to operate on at scale has been tamed.&lt;/p&gt;&#xA;&lt;p&gt;In practice, the latest versions of these systems not only &lt;strong&gt;rely on vast libraries of human-generated and curated text and images but also on other forms of human judgment and knowledge.&lt;/strong&gt; In particular, these systems depend on reinforcement learning from human feedback (RLHF) and prompt engineering. Even the latest &amp;ldquo;chain of thought&amp;rdquo; models typically start with dialogues with human users.&lt;/p&gt;&#xA;&lt;h2 id=&#34;challenges-and-opportunities&#34;&gt;Challenges and Opportunities&#xA;&lt;/h2&gt;&lt;h3 id=&#34;1-challenges&#34;&gt;1. Challenges&#xA;&lt;/h3&gt;&lt;p&gt;&lt;strong&gt;Debates about artificial intelligence should focus on the challenges and opportunities presented by these new cultural and social technologies.&lt;/strong&gt; The technology we currently possess has impacts on written and visual culture comparable to the effects of large-scale markets on the economy, large bureaucracies on society, and even the transformation of language by printing. What will happen next? Like past general-purpose technologies in economics, organization, and information, these systems will affect productivity, supplement human work, automate tasks previously only achievable by humans, and influence distribution, potentially altering resource acquisition patterns.&lt;/p&gt;&#xA;&lt;p&gt;They may also produce broader cultural impacts. We do not yet know whether these impacts will be as profound as those of printing, markets, or bureaucratic systems, but viewing them as cultural technologies highlights their potential significance.&lt;/p&gt;&#xA;&lt;p&gt;At the same time, &lt;strong&gt;these technologies create new possibilities for reorganizing information and coordinating the actions of millions globally.&lt;/strong&gt; Ongoing debates about the economic, social, and political consequences of large language models echo historical concerns and expectations regarding new cultural and social technologies. Guiding these debates &lt;strong&gt;requires recognizing the commonalities of new and old arguments while carefully mapping the specificities of new technologies.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;Such mapping is a core task of social sciences. Research into the past consequences of technologies can help us think about the latent social impacts of AI and explore pathways to enhance positive impacts and mitigate negative ones through the redesign of AI systems.&lt;/p&gt;&#xA;&lt;p&gt;However, a current obvious concern is that large models and related technologies may replace &amp;ldquo;knowledge workers&amp;rdquo; and whether &amp;ldquo;large models will homogenize or fragment culture and society.&amp;rdquo; Thinking about this issue in historical context is enlightening.&lt;/p&gt;&#xA;&lt;p&gt;The design goals of large models aim to faithfully reproduce the actual probabilities of sequences of text, images, and videos. Their inherent tendency is to be most accurate about the most common scenarios in the training data and least accurate about rare or entirely new scenarios, which may exacerbate this homogenization.&lt;/p&gt;&#xA;&lt;h3 id=&#34;2-opportunities&#34;&gt;2. Opportunities&#xA;&lt;/h3&gt;&lt;p&gt;On the other hand, &lt;strong&gt;large models may allow us to design new methods to access the diversity of cultural perspectives they summarize.&lt;/strong&gt; Integrating and balancing these perspectives could provide more nuanced means of solving complex problems. One way to achieve this is to construct a &amp;ldquo;social-like&amp;rdquo; ecology—different large models encoding different perspectives could debate, cross-fertilize to generate mixed perspectives, or identify gaps in human expertise. We may need new systems that diversify the reflections and roles of large models, producing distributions and diversities akin to human society.&lt;/p&gt;&#xA;&lt;p&gt;Such diversified systems are particularly crucial for scientific advancement. &lt;strong&gt;By linking numerous perspectives across texts, audio, and images, large models may help us discover unprecedented connections among them, benefiting science and society.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;The impact of new cultural and social technologies on economic relations also presents more subtle yet intriguing pathways. &lt;strong&gt;The development of cultural technologies has sparked fundamental economic tensions between information producers and distribution systems:&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;&#xA;&lt;p&gt;&lt;strong&gt;The contradiction between producers and distributors:&lt;/strong&gt; Distributors seek to acquire information at low costs, while producers wish to distribute information at low costs.&lt;/p&gt;&#xA;&lt;/li&gt;&#xA;&lt;li&gt;&#xA;&lt;p&gt;&lt;strong&gt;Digitalization exacerbates the contradiction:&lt;/strong&gt; The convenience of digital information distribution has sharpened this issue. The speed, efficiency, and scope with which large models process available information make these problems more pronounced. Concentrated power may make system owners more likely to seize benefits through efficiency at the expense of others&amp;rsquo; rights.&lt;/p&gt;&#xA;&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;h3 id=&#34;3-technological-and-political-issues-under-challenges-and-opportunities&#34;&gt;3. Technological and Political Issues Under Challenges and Opportunities&#xA;&lt;/h3&gt;&lt;p&gt;&lt;strong&gt;Key technological questions include: To what extent can systemic flaws in large models be corrected? How do they compare in advantages and disadvantages to flaws in systems based on human knowledge workers?&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;These questions should not obscure critical &lt;strong&gt;political issues: Which actors can mobilize their interests? How do they shape the hybrid outcomes of technology and organizational capabilities?&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;Tech commentators often simplify these issues into a binary opposition between machines and humans: either the forces of progress triumph over Luddite tendencies, or humans successfully resist the dehumanizing encroachment of artificial technologies. This not only misunderstands the complexities of distribution struggles that predate the advent of computers but also overlooks the multiple paths that future progress may take.&lt;/p&gt;&#xA;&lt;p&gt;In early cases of social and cultural technologies, a series of norms and regulatory frameworks gradually emerged to reconcile their impacts. However, these checks and balances do not form spontaneously but are the result of coordinated efforts by actors both inside and outside technology.&lt;/p&gt;&#xA;&lt;h2 id=&#34;looking-to-the-future&#34;&gt;Looking to the Future&#xA;&lt;/h2&gt;&lt;p&gt;The narrative of general artificial intelligence (i.e., viewing large models as superintelligent agents) is promoted both by optimists and skeptics within and outside the tech community. This narrative misinterprets the nature of these models and their relationship to past technological transformations. More importantly, &lt;strong&gt;it diverts attention from the real issues and opportunities posed by these technologies and ignores the lessons of history that guide weighing their pros and cons.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;There may exist hypothetical future AI systems closer to agents, but large models are clearly not such systems. Like library card catalogs or the internet, large models belong to a continuum of cultural and social technology development.&lt;/p&gt;&#xA;&lt;p&gt;Social sciences have explored this history in detail, forming a unique understanding of past technological upheavals. Close collaboration between computer science and engineering with social sciences will help us understand this history and apply its lessons: &lt;strong&gt;Will large models lead to cultural homogenization or fragmentation? Will they reinforce or undermine the social institutions of human discovery? Who will benefit and who will be harmed in this process?&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;These pressing questions are difficult to focus on in debates that analogize large models to human agents. &lt;strong&gt;Shifting the framework of debates about artificial intelligence will help promote research.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;If both computer scientists and social scientists understand that large models are merely (but also) new cultural and social technologies, they will find it easier to collaborate by combining their expertise. Computer scientists can integrate their deep understanding of system mechanisms with social scientists&amp;rsquo; knowledge of how large-scale systems have reshaped society, expanding existing research agendas and discovering new directions.&lt;/p&gt;&#xA;&lt;p&gt;Additionally, &lt;strong&gt;steering debates away from the existential fears of &amp;ldquo;machines taking over&amp;rdquo; and the utopian promise of &amp;ldquo;everyone having a perfect artificial assistant&amp;rdquo; will yield different actual policy consequences for large models.&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;With this mindset, engineers and computer scientists have already recognized the bias issues of large models and are contemplating their relationship with ethical justice. &lt;strong&gt;They need to go further and ask: How will these systems affect resource distribution? What are their actual consequences for social polarization and integration? Can we develop large models that enhance rather than suppress human creativity?&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;Answering these questions requires an understanding that combines social sciences and engineering. Shifting the debate about artificial intelligence from agency to cultural and social technology is a crucial first step in building this interdisciplinary understanding.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>A Comprehensive Guide to Understanding Large Language Models</title>
            <link>https://digitalxber.com/posts/note-05-2870aba5cd/</link>
            <pubDate>Tue, 22 Oct 2024 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-05-2870aba5cd/</guid>
            <description>&lt;p&gt;Last week, while sharing my article on &amp;ldquo;My Transition Journey as an AI Product Manager,&amp;rdquo; I hinted that I would produce a comprehensive article to help everyone systematically learn about large models. Today, I am delivering that article; it totals 22,000 words, and reading it is expected to take about 30 minutes, covering 15 topics related to large models.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 1&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;514px&#34; data-flex-grow=&#34;214&#34; height=&#34;420&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-2190c90c87.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-2190c90c87_hu_e38a15fb795d1bf8.jpeg 800w, https://digitalxber.com/posts/note-05-2870aba5cd/img-2190c90c87.jpeg 900w&#34; width=&#34;900&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;Over the past year, to be honest, there have been numerous articles introducing and explaining large models. Most people already have some foundational understanding, but my feeling is that &lt;strong&gt;the information is too fragmented and does not provide a systematic understanding&lt;/strong&gt;. Currently, there is no article that can comprehensively explain what large models are in one go.&lt;/p&gt;&#xA;&lt;p&gt;To alleviate my cognitive anxiety, I decided to compile the knowledge points I have understood about large models over the past year into one article, &lt;strong&gt;hoping to clarify large models through a single piece of writing&lt;/strong&gt;. This also serves as a summary of my extensive learning.&lt;/p&gt;&#xA;&lt;h2 id=&#34;what-will-i-share&#34;&gt;What Will I Share?&#xA;&lt;/h2&gt;&lt;p&gt;This article will share 15 topics related to large models. Initially, there were 20 topics, but I removed some more technical content to focus on issues that ordinary people or product managers should pay attention to. The goal is to ensure that as AI novices, we only need to master and understand these key points.&lt;/p&gt;&#xA;&lt;h2 id=&#34;who-is-this-article-for&#34;&gt;Who Is This Article For?&#xA;&lt;/h2&gt;&lt;p&gt;This article is suitable for the following groups:&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;Those who want to understand what large models are all about, including beginners.&lt;/li&gt;&#xA;&lt;li&gt;Those looking to transition into AI-related products and roles, including product managers and operations personnel.&lt;/li&gt;&#xA;&lt;li&gt;Those who have a preliminary understanding of AI but wish to advance their learning and reduce cognitive anxiety about AI.&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;p&gt;Content Statement: The entire content is a result of my personal synthesis after extensive reading and digesting numerous expert articles, books related to large models, and consultations with industry experts. My role is mainly as a knowledge synthesizer. If there are any inaccuracies in the descriptions, please feel free to inform me kindly!&lt;/p&gt;&#xA;&lt;h2 id=&#34;lecture-1-common-concepts-of-large-models&#34;&gt;Lecture 1: Common Concepts of Large Models&#xA;&lt;/h2&gt;&lt;p&gt;Before we start understanding large models, let&amp;rsquo;s first grasp some basic concepts. Mastering these professional terms and their relationships will benefit your subsequent reading and learning about any AI and large model-related content. I spent quite a bit of time organizing their relationships, so this part is essential to read carefully.&lt;/p&gt;&#xA;&lt;h3 id=&#34;1-common-ai-terms&#34;&gt;1. Common AI Terms&#xA;&lt;/h3&gt;&lt;p&gt;&lt;strong&gt;1) Large Model (LLM)&lt;/strong&gt;: All existing large models refer to large language models, specifically generative large models. Examples include GPT-4.0, GPT-4o, etc.&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Deep Learning&lt;/strong&gt;: A subfield of machine learning focused on using multi-layer neural networks for learning. Deep learning is effective in handling complex data such as images, audio, and text, making it very effective in AI applications.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Supervised Learning&lt;/strong&gt;: A method of machine learning where the model learns the mapping from input to output using a labeled training dataset. Common supervised learning algorithms include linear regression, logistic regression, support vector machines, K-nearest neighbors, decision trees, and random forests.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Unsupervised Learning&lt;/strong&gt;: A method of machine learning that finds patterns and structures in data without labeled data. It is mainly used for clustering and dimensionality reduction tasks. Common unsupervised learning algorithms include K-means clustering, hierarchical clustering, DBSCAN, principal component analysis (PCA), and t-SNE.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Semi-supervised Learning&lt;/strong&gt;: Combines a small amount of labeled data with a large amount of unlabeled data for training. It utilizes the rich information from unlabeled data and the accuracy of labeled data to improve model performance. Common methods include Generative Adversarial Networks (GANs) and autoencoders.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Reinforcement Learning&lt;/strong&gt;: A method of learning optimal strategies through interaction with the environment based on reward and punishment mechanisms. Reinforcement learning algorithms optimize decision-making processes through trial and error to maximize cumulative rewards. Common algorithms include Q-learning, policy gradient, and Deep Q-Networks (DQN).&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Model Architecture&lt;/strong&gt;: Represents the design of the backbone of a large model. Different model architectures affect the performance, efficiency, and even computational costs of large models, determining their scalability. For example, many large model vendors adjust the model architecture to reduce computational load and resource consumption.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Transformer Architecture&lt;/strong&gt;: The mainstream architecture used by current large models, including GPT-4.0 and most domestic large models. The widespread use of the Transformer architecture is primarily due to its ability to enable large models to understand human natural language, maintain contextual memory, and generate text. Other common model architectures include Convolutional Neural Networks (CNNs) for image processing and Generative Adversarial Networks (GANs) for image generation. A detailed introduction to the Transformer architecture will be covered later.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;MOE Architecture&lt;/strong&gt;: Represents the Mixture of Experts architecture, which combines multiple expert models to form a large model with a vast number of parameters, supporting the resolution of various complex professional problems. Models with MOE architecture may include Transformer-based models.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Machine Learning Technologies&lt;/strong&gt;: A broad category of technologies that implement AI, including deep learning, supervised learning, and reinforcement learning. As a product manager, you don&amp;rsquo;t need to delve too deeply into the specifics; just understand the relationships between these learning types to avoid being misled by technical personnel.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;NLP Technology (Natural Language Processing)&lt;/strong&gt;: An application area of AI focused on enabling computers to understand, interpret, and generate human language, used in text analysis, machine translation, speech recognition, and dialogue systems. In simpler terms, it is a technology that converts a lot of information into a format understandable by human natural language.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;CV Technology (Computer Vision)&lt;/strong&gt;: If NLP deals with text, CV addresses visual content-related technologies. CV technologies include common image recognition, video analysis, and image segmentation technologies, which are also prevalent in large model applications, especially in the upcoming multi-modal large model technologies.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Speech Recognition and Synthesis Technologies&lt;/strong&gt;: Includes converting speech to text and speech synthesis technologies, such as Text-to-Speech (TTS) technology.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Retrieval-Augmented Generation (RAG)&lt;/strong&gt;: Refers to the technology where large models generate content based on information retrieved from search engines and knowledge bases. RAG is a technology involved in most AI applications.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Knowledge Graph&lt;/strong&gt;: A technology that connects knowledge, allowing knowledge to establish relationships, helping models acquire the most relevant knowledge more effectively and enhancing their ability to process complex relational information and AI reasoning.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Function Call&lt;/strong&gt;: Refers to the ability in large language models (like GPT) to call built-in or external functions to perform specific tasks or operations. This mechanism allows the model to be more than just a text generation tool, enabling it to execute a variety of operations by specifying different functionalities. Function Call allows large models to integrate with various API capabilities, enhancing their practical applications, such as supporting content retrieval and document recognition.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;strong&gt;2) Terms Related to Large Model Training and Optimization Technologies&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Pre-training&lt;/strong&gt;: Refers to the process of training a model on a large dataset. The pre-training dataset is usually large and diverse, resulting in a general-purpose model with strong capabilities, &lt;strong&gt;similar to a person who has learned various general knowledge through compulsory education and university studies.&lt;/strong&gt;&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Fine-tuning&lt;/strong&gt;: Refers to further training a large model on specific tasks or smaller datasets to improve its performance on targeted problems. Unlike the pre-training phase, the fine-tuning phase uses a smaller amount of data, primarily from vertical domains. The fine-tuning process results in a specialized or industry-specific model, &lt;strong&gt;akin to a new graduate receiving professional skill training after joining a company.&lt;/strong&gt;&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Prompt Engineering&lt;/strong&gt;: In product manager terms, this means using question formats that large models can better understand to yield the desired results. Therefore, prompt engineering is a skill in learning how to ask questions effectively.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Model Distillation&lt;/strong&gt;: A technique that transfers knowledge from a large model (the teacher model) to a smaller model (the student model). The student model learns from the outputs of the teacher model to improve its performance while maintaining similar accuracy.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Model Pruning&lt;/strong&gt;: Refers to the removal of unnecessary parameters from a large model to reduce its overall parameter size, thereby lowering computational load and cost.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;strong&gt;3) AI Application-Related Terms&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Agent&lt;/strong&gt;: An agent is simply understood as an AI application with a specific capability. If applications in the internet era are called apps, applications in the AI era are called agents.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Chatbot&lt;/strong&gt;: Refers to AI chatbots, a type of AI application that interacts through chat, including products like ChatGPT, which belong to the chatbot category.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;strong&gt;4) Terms Related to Large Model Performance&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Emergence&lt;/strong&gt;: Refers to the phenomenon where large models exhibit capabilities beyond expectations once their parameter scale reaches a certain level.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Hallucination&lt;/strong&gt;: Refers to instances where large models generate nonsensical content, mistakenly treating incorrect facts as real, leading to unrealistic outputs.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Amnesia&lt;/strong&gt;: Refers to the situation where, after a certain number of dialogue turns and length, the model suddenly forgets and begins to repeat itself and forgets previous context. The memory of large models is mainly influenced by factors like the model&amp;rsquo;s context length.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;2-understanding-the-relationship-between-ai-machine-learning-deep-learning-and-nlp&#34;&gt;2. Understanding the Relationship Between AI, Machine Learning, Deep Learning, and NLP&#xA;&lt;/h2&gt;&lt;p&gt;If you are interested in AI and large models, you will likely encounter keywords like &lt;strong&gt;&amp;ldquo;AI,&amp;rdquo; &amp;ldquo;Machine Learning,&amp;rdquo; &amp;ldquo;Deep Learning,&amp;rdquo; and &amp;ldquo;NLP&amp;rdquo;&lt;/strong&gt; in your future studies. Therefore, it is best to clarify the concepts and definitions of these professional terms and their logical relationships for easier understanding.&lt;/p&gt;&#xA;&lt;p&gt;In summary, the relationship between these concepts is as follows:&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;Machine learning is a core technology of AI. Besides machine learning, AI&amp;rsquo;s core technologies include expert systems, Bayesian networks, etc. (you don&amp;rsquo;t need to delve too deeply into what these are), with deep learning being a type of machine learning.&lt;/li&gt;&#xA;&lt;li&gt;NLP is one of the application task types in AI, focused on natural language processing. Besides NLP, AI application technologies also include CV (computer vision) technology, speech recognition and synthesis technologies, etc.&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 2&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;261px&#34; data-flex-grow=&#34;108&#34; height=&#34;638&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-4ed031aa81.jpeg&#34; width=&#34;695&#34;&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;3-understanding-the-transformer-architecture&#34;&gt;3. Understanding the Transformer Architecture&#xA;&lt;/h2&gt;&lt;p&gt;When discussing large models, one cannot overlook the Transformer architecture. If large models are like a tree, the Transformer architecture serves as the trunk of the model. The emergence of products like ChatGPT is primarily due to the design of the Transformer architecture, which enables models to understand context, maintain memory, and predict unknown words. Additionally, the introduction of the Transformer has allowed large models to train on unlabeled data without relying heavily on large amounts of labeled data. This breakthrough means that previously, creating a model required significant human effort for data cleaning and labeling, but now, fragmented and scattered data can simply be fed into the model for processing. We can understand these concepts through the following points:&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;Relationship Between Transformer Architecture and Deep Learning Technologies&lt;/strong&gt;: The Transformer architecture belongs to the category of deep learning technologies, meaning it is an implementation and design form within deep learning. Besides the Transformer architecture, traditional Recurrent Neural Networks (RNNs) and Long Short-Term Memory networks (LSTMs) are also part of deep learning.&lt;/p&gt;&#xA;&lt;h2 id=&#34;4-understanding-the-relationship-between-the-transformer-architecture-and-gpt&#34;&gt;4. Understanding the Relationship Between the Transformer Architecture and GPT&#xA;&lt;/h2&gt;&lt;p&gt;GPT stands for Generative Pre-trained Transformer, meaning GPT is a large language model developed based on the Transformer architecture by OpenAI.&lt;/p&gt;&#xA;&lt;p&gt;The core idea of GPT is to enhance the ability to generate and understand natural language through &lt;strong&gt;large-scale pre-training and fine-tuning.&lt;/strong&gt; The emergence of the Transformer architecture has solved the issues of understanding context, processing large amounts of data, and predicting text. However, OpenAI was the first to adopt the pre-training + fine-tuning approach to improve and utilize the Transformer architecture, enabling it to possess the capabilities of products like ChatGPT for understanding and generating natural language.&lt;/p&gt;&#xA;&lt;p&gt;The reason GPT can generate and understand natural language is that during the pre-training phase, it learns broad language patterns and knowledge from a large corpus of unlabeled text, with the pre-training task typically being a language modeling task, where the model predicts the next word given a sequence of preceding words. The specific differences are as follows:&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;1) Differences in Capabilities&lt;/strong&gt;: The Transformer architecture enables models to understand context, process large amounts of data, and predict text, but it does not possess the ability to understand and generate natural language. In contrast, GPT, after adding natural language pre-training, has the ability to understand and generate natural language.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;2) Architectural Foundations&lt;/strong&gt;:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Transformer&lt;/strong&gt;: The original Transformer model consists of an encoder and a decoder. The encoder processes the input sequence to generate intermediate representations, which the decoder uses to generate the output sequence. This architecture is particularly suitable for sequence-to-sequence tasks, such as machine translation. The encoder employs a bidirectional processing mechanism, allowing it to use bidirectional attention, meaning each word can consider the information from all other words in the sequence, regardless of whether they are preceding or following words.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;GPT&lt;/strong&gt;: GPT primarily uses the decoder part of the Transformer, focusing solely on generation tasks. Its training and generation processes are unidirectional, meaning each word can only see the preceding words (unidirectional attention). This architecture is more suitable for text generation tasks. The encoder adopts a unidirectional processing mechanism, meaning that when generating the next word, GPT can only consider the previous words, which aligns with the natural form of language modeling.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;strong&gt;3) Implementation Methods for Solving Specific Problems&lt;/strong&gt;:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;The Transformer is trained to solve specific task types (like machine translation) by optimizing its performance through training, with both the encoder and decoder trained simultaneously.&lt;/li&gt;&#xA;&lt;li&gt;In contrast, GPT solves specific task types through supervised fine-tuning, meaning it does not require training for specific task types but only needs to provide some specific task data to achieve results. It is important to understand that training and fine-tuning are different implementation cost methods.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;strong&gt;4) Application Domains&lt;/strong&gt;:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;The traditional Transformer framework can &lt;strong&gt;be applied to various sequence-to-sequence tasks&lt;/strong&gt;, such as machine translation, text summarization, and speech recognition. Since it includes both an encoder and a decoder, the Transformer can handle various input and output format tasks.&lt;/li&gt;&#xA;&lt;li&gt;GPT is &lt;strong&gt;primarily used for generation tasks&lt;/strong&gt;, such as text generation, dialogue systems, and question-answering systems. It excels in generating coherent and creative text.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;5-understanding-the-moe-architecture&#34;&gt;5. Understanding the MOE Architecture&#xA;&lt;/h2&gt;&lt;p&gt;In addition to the Transformer architecture, another popular architecture is the MOE architecture (Mixture of Experts), which dynamically selects and combines multiple sub-models (i.e., experts) to complete tasks. The key idea of MOE is to solve a series of complex tasks by combining multiple expert models rather than relying on a single large model for all tasks.&lt;/p&gt;&#xA;&lt;p&gt;The main advantage of the MOE architecture is its ability to maintain computational efficiency while handling large-scale data and model parameters, significantly reducing computational costs while retaining model capabilities.&lt;/p&gt;&#xA;&lt;p&gt;The Transformer and MOE can be used in conjunction, commonly referred to as MOE-Transformer or Sparse Mixture of Experts Transformer. In this architecture:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;The Transformer processes input data, leveraging its powerful self-attention mechanism to capture dependencies within the sequence.&lt;/li&gt;&#xA;&lt;li&gt;The MOE dynamically selects and combines different experts, enhancing the model&amp;rsquo;s computational efficiency and capabilities.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;lecture-2-differences-between-large-models-and-traditional-models&#34;&gt;Lecture 2: Differences Between Large Models and Traditional Models&#xA;&lt;/h2&gt;&lt;p&gt;When we talk about large models, we typically refer to LLMs (Large Language Models), or more specifically, models like GPT (generative pre-trained models based on the Transformer architecture). Firstly, it is a language model addressing natural language tasks rather than problems in images, videos, or speech. (Models that handle multiple modalities, including language, images, videos, and speech, are later referred to as multi-modal large models, which are not the same concept as LLMs.) Secondly, LLMs are generative models, meaning their primary ability is to generate rather than predict or make decisions.&lt;/p&gt;&#xA;&lt;p&gt;In contrast to traditional models, large models generally have the following characteristics:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Ability to Understand and Generate Natural Language&lt;/strong&gt;: Many traditional models we have encountered may not understand human natural language, let alone generate language that humans can comprehend.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Powerful and Versatile Capabilities&lt;/strong&gt;: Traditional models typically solve one or a few problems, with a strong specialization, while large models possess strong general capabilities and can address a wide range of issues.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Contextual Memory Capabilities&lt;/strong&gt;: Large models have memory capabilities, allowing them to relate to contextual dialogues rather than being forgetful robots, which is one of the key differences from many traditional models.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Training Method&lt;/strong&gt;: Based on large amounts of unlabeled text, pre-training is conducted through unsupervised methods. Unlike many traditional models that rely on large amounts of labeled data, the unsupervised approach significantly reduces the costs of data cleaning and preparation. Furthermore, pre-training requires a massive amount of training data; for example, GPT-3.5&amp;rsquo;s training corpus reached 45 terabytes.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Massive Parameter Scale&lt;/strong&gt;: Most large models have parameter scales in the hundreds of billions, such as GPT-3.5, which has 175 billion parameters, while GPT-4.0 is rumored to reach trillions of parameters. These parameters learn and adjust during the model training process to perform specific tasks better.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Training Requires Significant Computational Resources&lt;/strong&gt;: Due to their scale and complexity, these models require substantial computational resources for training and inference, typically needing specialized hardware like GPUs or TPUs. Research indicates that training generative AIs like ChatGPT requires support from at least 10,000 NVIDIA A100 accelerator cards, with training costs for models like GPT-3.5 reaching up to $9 million.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;lecture-3-evolution-of-large-models&#34;&gt;Lecture 3: Evolution of Large Models&#xA;&lt;/h2&gt;&lt;h3 id=&#34;1-evolution-of-large-model-generation-capabilities&#34;&gt;1. Evolution of Large Model Generation Capabilities&#xA;&lt;/h3&gt;&lt;p&gt;Understanding the evolution of LLMs helps everyone grasp how large models have gradually acquired their current capabilities and makes it easier to understand the relationship between LLMs and Transformers. The following outlines the evolution of large models:&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;&lt;strong&gt;N-gram&lt;/strong&gt;: The earliest stage of large model generation capabilities, primarily addressing the ability to predict the next word. This forms the basis of text generation, but its limitations lie in understanding context and grammatical structures.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;RNN (Recurrent Neural Network) and LSTM (Long Short-Term Memory)&lt;/strong&gt;: At this stage, these two models addressed the issue of context length, enabling relatively longer context windows, but they struggled to handle large amounts of data.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Transformer&lt;/strong&gt;: Combines the predictive capabilities of the previous two models with the ability to remember longer contexts while supporting training on large datasets, but lacks the ability to understand and generate natural language.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;LLM (Large Language Model)&lt;/strong&gt;: Adopts the GPT pre-training and supervised fine-tuning approach, enabling the model to understand and generate natural language, thus called a large language model. It can be said that the emergence of pre-training and supervised fine-tuning brought the Transformer into the development stage of large models.&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 3&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;900px&#34; data-flex-grow=&#34;375&#34; height=&#34;288&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-0a893923c6.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-0a893923c6_hu_8e2bba945c13ec24.jpeg 800w, https://digitalxber.com/posts/note-05-2870aba5cd/img-0a893923c6.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;h3 id=&#34;2-development-history-from-gpt-1-to-gpt-4&#34;&gt;2. Development History from GPT-1 to GPT-4&#xA;&lt;/h3&gt;&lt;p&gt;&lt;strong&gt;GPT-1&lt;/strong&gt;: Introduced the unsupervised training step for the first time, solving the problem of needing large amounts of labeled data for previous models. The unsupervised training method allows GPT to train on a vast amount of unlabeled data. However, its limitations stem from the small parameter scale (only 117 million parameters), making it unable to solve complex tasks without supervised fine-tuning, which can be cumbersome.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;GPT-2&lt;/strong&gt;: Increased the parameter scale to 1.5 billion and expanded the training text size fourfold to 40GB. By increasing parameter scale and training data size, the model&amp;rsquo;s capabilities improved, but it still faced limitations in addressing complex problems.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;GPT-3&lt;/strong&gt;: Expanded the parameter scale to 175 billion, achieving remarkable performance in text generation and language understanding, and eliminated the fine-tuning step, meaning it could solve complex problems without needing fine-tuning. However, GPT-3&amp;rsquo;s limitations arose from its training on a vast amount of internet data, which may include false and erroneous texts, leading to safety issues.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;InstructGPT&lt;/strong&gt;: To address the limitations of GPT-3, &lt;strong&gt;supervised fine-tuning (SFT) and reinforcement learning from human feedback (RLHF)&lt;/strong&gt; were added after pre-training to optimize the model&amp;rsquo;s errors. This model became InstructGPT. The process involves providing the model with some real “standard answers” for supervised fine-tuning, constructing a scoring model for the generated results, and using the scoring model to adjust the model&amp;rsquo;s strategy for improvement. Thus, many large model vendors focus on the quality and quantity of data provided during the supervised fine-tuning stage to reduce hallucination rates.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;GPT-3.5&lt;/strong&gt;: Released in March 2022, OpenAI&amp;rsquo;s GPT-3 new version had a training data cutoff in June 2021, with an expanded training data size of 45TB, and was referred to as GPT-3.5 in November.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;GPT-4.0&lt;/strong&gt;: Released in April 2023, OpenAI significantly improved overall reasoning capabilities and supported multi-modal capabilities.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;GPT-4o&lt;/strong&gt;: Released in May 2024, enhanced voice chat capabilities.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;O1&lt;/strong&gt;: In September 2024, OpenAI launched the O1 model, focusing on enhancing reasoning capabilities.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 4&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;215px&#34; data-flex-grow=&#34;89&#34; height=&#34;1202&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-593051da41.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-593051da41_hu_35d0cf1ba1345bd5.jpeg 800w, https://digitalxber.com/posts/note-05-2870aba5cd/img-593051da41.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;h3 id=&#34;3-principles-of-text-generation-in-large-models&#34;&gt;3. Principles of Text Generation in Large Models&#xA;&lt;/h3&gt;&lt;h3 id=&#34;1-how-does-gpt-generate-text&#34;&gt;1. How Does GPT Generate Text?&#xA;&lt;/h3&gt;&lt;p&gt;The process of a large model generating text can be summarized in five steps:&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;After receiving a prompt, the model first tokenizes the input content, breaking it into multiple tokens.&lt;/li&gt;&#xA;&lt;li&gt;Based on the Transformer architecture, it understands the relationships between tokens to grasp the overall meaning of the prompt.&lt;/li&gt;&#xA;&lt;li&gt;It predicts the next token based on context, which may yield multiple results, each with corresponding probability values.&lt;/li&gt;&#xA;&lt;li&gt;The token with the highest probability is selected as the predicted result for the next word.&lt;/li&gt;&#xA;&lt;li&gt;The fourth step is repeated until the entire content is generated.&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 5&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;346px&#34; data-flex-grow=&#34;144&#34; height=&#34;748&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-dfaf4fbe5e.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-dfaf4fbe5e_hu_8400f82f8c477994.jpeg 800w, https://digitalxber.com/posts/note-05-2870aba5cd/img-dfaf4fbe5e.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;h3 id=&#34;2-classification-of-llms&#34;&gt;2. Classification of LLMs&#xA;&lt;/h3&gt;&lt;h4 id=&#34;1-classification-by-modality-type&#34;&gt;1. Classification by Modality Type&#xA;&lt;/h4&gt;&lt;p&gt;Currently, large models on the market can be categorized into text generation models (e.g., GPT-3.5), image generation models (e.g., DALL-E), video generation models (e.g., Sora, 可灵), speech generation models, and multi-modal models (e.g., GPT-4.0).&lt;/p&gt;&#xA;&lt;h4 id=&#34;2-classification-by-training-stage&#34;&gt;2. Classification by Training Stage&#xA;&lt;/h4&gt;&lt;p&gt;Models can be divided into basic language models and instruction fine-tuned models:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Basic Language Model&lt;/strong&gt;: Refers to models that have only undergone pre-training on large-scale text corpora without any instruction or downstream task fine-tuning, such as GPT-3, which is the publicly available basic language model from OpenAI.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Instruction Fine-tuned Model&lt;/strong&gt;: Refers to models that have undergone instruction fine-tuning, human feedback, and alignment optimizations based on natural language task descriptions. For example, GPT-3.5 is trained based on GPT-3.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h4 id=&#34;3-classification-by-general-and-industry-models&#34;&gt;3. Classification by General and Industry Models&#xA;&lt;/h4&gt;&lt;p&gt;Large models on the market can also be classified into general large models and industry-specific models. General large models perform well across a wide range of tasks and fields, but may not fully understand and utilize domain-specific information, thus may not solve specific industry or scenario problems. Industry-specific models, on the other hand, are trained and adjusted based on general large models to achieve higher performance and accuracy in specific fields.&lt;/p&gt;&#xA;&lt;h3 id=&#34;3-core-technologies-of-llms&#34;&gt;3. Core Technologies of LLMs&#xA;&lt;/h3&gt;&lt;p&gt;This section may contain many technical terms that are relatively difficult to understand. However, for product managers, it is not necessary to delve into the technical details; understanding the key concepts is sufficient. It is essential for AI product managers to comprehend technical terms to facilitate communication with development and technical teams.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;1. Model Architecture&lt;/strong&gt;: The Transformer architecture has been described in detail earlier, so I will not repeat it here. However, the Transformer architecture is one of the foundational core technologies of large models.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;2. Pre-training and Fine-tuning&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Pre-training&lt;/strong&gt;: Conducted on large-scale unlabeled data, this is one of the key technologies for large language models. The emergence of pre-training technology means that models no longer need to rely on large amounts of labeled data, significantly reducing the costs of manual labeling.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Fine-tuning&lt;/strong&gt;: This technology is used to further leverage large models. The performance of pre-trained models on specific tasks is generally average, so fine-tuning on specific datasets is needed to adapt to specific applications. Fine-tuning can significantly enhance model performance on specific tasks.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;strong&gt;3. Model Compression and Acceleration&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Model Pruning&lt;/strong&gt;: By removing unimportant parameters, the size of the model and computational complexity can be reduced.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Knowledge Distillation&lt;/strong&gt;: Training a smaller student model to mimic the behavior of a large teacher model, retaining most performance while reducing computational costs.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;lecture-7-six-steps-in-large-model-development&#34;&gt;Lecture 7: Six Steps in Large Model Development&#xA;&lt;/h2&gt;&lt;p&gt;According to information released by OpenAI, the development of large models typically goes through the following six steps. This process should be representative of the development process for most large models in the industry:&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 6&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;528px&#34; data-flex-grow=&#34;220&#34; height=&#34;490&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-8cbff34f2e.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-8cbff34f2e_hu_28c961c4c0ee9dd.jpeg 800w, https://digitalxber.com/posts/note-05-2870aba5cd/img-8cbff34f2e.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;&lt;strong&gt;Data Collection and Processing&lt;/strong&gt;: This stage involves collecting a large amount of text data, which may include books, web pages, articles, etc. The data is then cleaned to remove irrelevant or low-quality content, followed by preprocessing such as tokenization and removal of sensitive information.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Model Design&lt;/strong&gt;: Determining the model architecture, such as the Transformer architecture used by GPT-4, and setting the model size, including the number of layers, hidden units, and total parameters.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Pre-training&lt;/strong&gt;: At this stage, the model acts like a student in school, learning language and knowledge by reading a large number of books (such as web pages and articles). Or, it can be likened to a &amp;ldquo;sponge&amp;rdquo; absorbing as much information as possible, learning basic language rules, such as how to form sentences and how words relate to each other. At this point, the model can understand basic language structures but lacks specialized knowledge for specific tasks. The pre-training phase typically requires a very large amount of data, consuming the most computational resources and time; for example, completing one pre-training session for GPT-3 requires 3640 petaflops of computation and nearly 1,000 GPUs.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Instruction Fine-tuning&lt;/strong&gt;: Also known as supervised fine-tuning, this process involves feeding the model some question-answer pairs with ideal outputs for further training, resulting in a supervised fine-tuned model. This stage is akin to “career training,” where the model learns how to adjust its responses based on specific instructions or tasks, improving its performance on specific types of questions or tasks. The instruction fine-tuning stage requires relatively fewer high-quality data, making the training time and consumption smaller.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Reward&lt;/strong&gt;: This stage sets up an “incentive mechanism” for the model, teaching it what constitutes a good response or behavior through rewards. This approach helps the model better meet user needs and focus on providing valuable, accurate answers. This process requires the model&amp;rsquo;s trainers to extensively evaluate and provide feedback on the model&amp;rsquo;s responses, gradually adjusting the quality of its responses, which also requires relatively high-quality data and takes days to complete.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Reinforcement Learning&lt;/strong&gt;: In this final stage, the model undergoes “live drills,” learning how to improve through trial and error. During this phase, the model attempts various strategies in real-world complex situations, identifying the most effective methods. The model becomes smarter and more flexible during this stage, enabling it to make better judgments and responses in complex and uncertain situations.&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;h2 id=&#34;lecture-8-understanding-large-model-training-and-fine-tuning&#34;&gt;Lecture 8: Understanding Large Model Training and Fine-tuning&#xA;&lt;/h2&gt;&lt;h3 id=&#34;1-understanding-large-model-training-content&#34;&gt;1. Understanding Large Model Training Content&#xA;&lt;/h3&gt;&lt;p&gt;&lt;strong&gt;1) What Data Is Needed for Large Model Training?&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Text Data&lt;/strong&gt;: Mainly used for training language models, such as news articles, books, social media posts, Wikipedia, etc.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Structured Data&lt;/strong&gt;: Such as knowledge graphs, used to enhance the language model&amp;rsquo;s knowledge.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Semi-structured Data&lt;/strong&gt;: Such as XML and JSON formats, which facilitate information extraction.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;strong&gt;2) Sources of Training Data&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Public Datasets&lt;/strong&gt;: Such as Common Crawl, Wikipedia, OpenWebText, etc.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Proprietary Data&lt;/strong&gt;: Company internal data or paid proprietary data.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;User-Generated Content&lt;/strong&gt;: Content generated by users on social media, forums, comments, etc.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Synthetic Data&lt;/strong&gt;: Data generated through Generative Adversarial Networks (GANs) or other generative models.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;strong&gt;3) What Are the Costs of Large Model Training?&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Computational Resources&lt;/strong&gt;: The cost of using GPUs/TPUs depends mainly on the model&amp;rsquo;s scale and training time. Large models typically require thousands to tens of thousands of hours of GPU computing time.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Storage Costs&lt;/strong&gt;: For storing large-scale datasets and model weights. Datasets and model files can reach terabyte levels.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Data Acquisition Costs&lt;/strong&gt;: The cost of purchasing proprietary data or the labor costs for data cleaning and labeling.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Energy Costs&lt;/strong&gt;: Training large models consumes a significant amount of electricity, increasing operational costs.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;R&amp;amp;D Costs&lt;/strong&gt;: Including salaries for researchers and engineers, as well as the costs of developing and maintaining models.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h3 id=&#34;2-understanding-large-model-fine-tuning-content&#34;&gt;2. Understanding Large Model Fine-tuning Content&#xA;&lt;/h3&gt;&lt;ol&gt;&#xA;&lt;li&gt;Two Stages of Large Model Fine-tuning: &lt;strong&gt;Supervised Fine-tuning (SFT) and Reinforcement Learning (RLHF)&lt;/strong&gt;. The differences between the two stages are as follows:&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 7&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;197px&#34; data-flex-grow=&#34;82&#34; height=&#34;882&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-05-2870aba5cd/img-121d5f1cee.jpeg&#34; width=&#34;724&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;2) Two Methods of Large Model Fine-tuning: Lora Fine-tuning and SFT Fine-tuning&lt;/strong&gt;&lt;/p&gt;&#xA;&lt;p&gt;Currently, there are two methods for fine-tuning models: Lora fine-tuning and SFT fine-tuning. The differences between these two methods are:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;Lora fine-tuning adjusts only a portion of the model&amp;rsquo;s parameters, not requiring fine-tuning of the entire model. This method is suitable for resource-limited scenarios or targeted fine-tuning, allowing the model to address single-task scenarios.&lt;/li&gt;&#xA;&lt;li&gt;SFT fine-tuning involves adjusting all parameters of the model, fine-tuning the entire model to enable it to address more specific tasks.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;lecture-9-main-factors-affecting-large-model-performance&#34;&gt;Lecture 9: Main Factors Affecting Large Model Performance&#xA;&lt;/h2&gt;&lt;p&gt;As we know, although there are many large models available on the market, there are differences in performance among them. For example, OpenAI&amp;rsquo;s models hold a leading position in the industry. Why do performance differences exist among large models? The five most important factors affecting large model performance are as follows:&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;&lt;strong&gt;Model Architecture&lt;/strong&gt;: The design of the model, including the number of layers, the number of hidden units, and the total number of parameters, significantly impacts its ability to handle complex tasks.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Quality and Quantity of Training Data&lt;/strong&gt;: The performance of the model heavily relies on the coverage and diversity of its training data. High-quality and diverse datasets help the model understand and generate language more accurately. Currently, most models mainly use public data, and companies with richer, high-quality data resources will have a competitive advantage. In China, a disadvantage is that open-source datasets are primarily in English, with relatively fewer Chinese datasets.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Parameter Scale&lt;/strong&gt;: The more parameters a model has, the better it can learn and capture complex data patterns, but this also increases computational costs. Therefore, companies with strong computational resources will have a higher advantage. The core factors for computational power depend on the computational volume (number of GPUs), network, and storage dimensions.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Algorithm Efficiency&lt;/strong&gt;: The algorithms used for training and optimizing the model, such as optimizer selection and learning rate adjustments, significantly impact the model&amp;rsquo;s learning efficiency and final performance.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Training Frequency&lt;/strong&gt;: Ensuring that the model undergoes sufficient training iterations to achieve optimal performance while avoiding overfitting issues.&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;h2 id=&#34;lecture-10-how-to-measure-the-quality-of-large-models&#34;&gt;Lecture 10: How to Measure the Quality of Large Models?&#xA;&lt;/h2&gt;&lt;p&gt;From the application perspective of large models, how to measure the quality of a large model and what evaluation framework to use is essential. Through this section, you can understand the dimensions from which evaluation institutions assess the capabilities of large models. Moreover, if you face selection issues regarding large models, you should establish your own judgment system.&lt;/p&gt;&#xA;&lt;p&gt;After reading and referencing multiple documents on measuring large models, I have summarized the evaluation dimensions into three aspects:&lt;/p&gt;&#xA;&lt;h3 id=&#34;1-how-to-measure-the-product-performance-of-large-models&#34;&gt;1. How to Measure the Product Performance of Large Models&#xA;&lt;/h3&gt;&lt;p&gt;Typically, the product performance of a large model is evaluated based on the following dimensions:&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;1) Semantic Understanding Ability&lt;/strong&gt;: This includes the basic dimensions of semantics, grammar, and context, which essentially determine whether you can have a normal conversation with the model and whether the model&amp;rsquo;s responses are coherent, especially regarding Chinese semantic understanding. Furthermore, it also assesses whether the model supports multi-language understanding.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;2) Logical Reasoning&lt;/strong&gt;: This includes the model&amp;rsquo;s reasoning ability, numerical calculation ability, and contextual understanding ability, which is one of the core capabilities of large models, directly determining the model&amp;rsquo;s intelligence level.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;3) Accuracy of Generated Content&lt;/strong&gt;: This includes the rate of hallucinations and the ability to identify traps.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;4) Hallucination Rate&lt;/strong&gt;: This includes the accuracy of the model&amp;rsquo;s responses and results. Sometimes, the model may generate nonsensical content that users may mistakenly believe to be true, which can be quite problematic.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;5) Trap Information Identification Rate&lt;/strong&gt;: This indirectly assesses the model&amp;rsquo;s ability to recognize and handle trap information, as models that poorly identify traps may generate responses based on incorrect information.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;6) Quality of Generated Content&lt;/strong&gt;: While ensuring the authenticity and accuracy of generated content, the dimensions for measuring content quality include:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Diversity of Generated Content&lt;/strong&gt;: Whether the model can support diverse and multi-faceted content outputs.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Professionalism&lt;/strong&gt;: Whether the model can produce professional content in vertical scenarios.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Creativity&lt;/strong&gt;: Whether the generated content is sufficiently creative.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Timeliness&lt;/strong&gt;: The recency of the generated results.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;strong&gt;7) Contextual Memory Ability&lt;/strong&gt;: This represents the model&amp;rsquo;s memory capabilities and the length of its contextual window.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;8) Model Performance&lt;/strong&gt;: This includes response speed, resource consumption, robustness, and stability (the ability to handle anomalies and unknown information reliably).&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;9) Human-likeness&lt;/strong&gt;: This dimension evaluates whether the model truly exhibits “human-like” qualities, reaching a level of intelligence, including emotional analysis capabilities.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;10) Multi-modal Capability&lt;/strong&gt;: Finally, it assesses the model&amp;rsquo;s ability to process and generate across modalities, including text, images, videos, and speech.&lt;/p&gt;&#xA;&lt;h3 id=&#34;2-how-to-measure-the-basic-capabilities-of-large-models&#34;&gt;2. How to Measure the Basic Capabilities of Large Models&#xA;&lt;/h3&gt;&lt;p&gt;It is well-known that the three most important elements for measuring the basic capabilities of large models are: algorithms, computational power, and data. More specifically, they include the following parts:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Parameter Scale&lt;/strong&gt;: This dimension measures the strength of the algorithm. A larger parameter scale indicates that the model can handle more complex problems and consider more dimensions, simply put, the stronger the model.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Data Volume&lt;/strong&gt;: Models operate on data, and the larger the underlying data volume, the better the model&amp;rsquo;s performance.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Data Quality&lt;/strong&gt;: Data quality includes the intrinsic value of the data and the degree of cleaning performed on the data. Data quality can be hierarchical; for example, user consumption data is more valuable than ordinary social attribute information. The higher the data value, the better the model&amp;rsquo;s performance. Additionally, the business&amp;rsquo;s data cleaning quality reflects in the precision of data labeling.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Training Frequency&lt;/strong&gt;: The more training iterations a model undergoes, the richer its experience, leading to better performance.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h3 id=&#34;3-how-to-evaluate-model-safety&#34;&gt;3. How to Evaluate Model Safety&#xA;&lt;/h3&gt;&lt;p&gt;In addition to assessing the capabilities of large models, safety considerations are also crucial. Even if a model is highly capable, if safety issues are not adequately addressed, large models cannot develop rapidly. We primarily evaluate model safety based on the following dimensions:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Content Safety&lt;/strong&gt;: This includes whether the generated content complies with safety management standards, social norms, and legal regulations.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Ethical Standards&lt;/strong&gt;: This includes whether the generated content contains bias or discrimination and whether it aligns with social values and ethical standards.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Privacy and Copyright Protection&lt;/strong&gt;: This includes the protection of personal and corporate privacy and compliance with copyright protection laws.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;lecture-11-limitations-of-large-models&#34;&gt;Lecture 11: Limitations of Large Models&#xA;&lt;/h2&gt;&lt;h3 id=&#34;1-the-hallucination-problem&#34;&gt;1. The Hallucination Problem&#xA;&lt;/h3&gt;&lt;p&gt;The hallucination problem refers to the generation of information that appears reasonable but is actually incorrect or fabricated. In natural language processing, this may manifest as the model generating text or responses that seem coherent but lack truthfulness or accuracy. Currently, the hallucination problem is one of the main reasons users question the applicability of large models and why the results generated by large models are often difficult to use. It is also a challenging issue for AI applications.&lt;/p&gt;&#xA;&lt;p&gt;&lt;strong&gt;What Causes Hallucinations in Large Models?&lt;/strong&gt; The main sources are as follows:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Overfitting Training Data&lt;/strong&gt;: The model may have overfitted noise or erroneous information in the training data, leading to the generation of fabricated content.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Presence of False Information in Training Data&lt;/strong&gt;: If the training data does not adequately cover various real-world scenarios, the model may produce fabricated information in unfamiliar situations.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Insufficient Consideration of Information Credibility&lt;/strong&gt;: The model may not effectively assess the credibility of generated information, instead generating responses that seem reasonable but are actually fictitious.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;&lt;strong&gt;Are There Solutions to Mitigate Hallucination Issues?&lt;/strong&gt; Currently, potential ways to alleviate hallucination issues include:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Using More Diverse Training Data&lt;/strong&gt;: Introducing a more diverse and authentic training dataset to reduce the likelihood of the model overfitting erroneous information.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Modeling Information Credibility and Increasing Verification Mechanisms&lt;/strong&gt;: Incorporating components to estimate the credibility of generated information to filter or reduce the probability of generating fabricated content.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;External Verification Mechanisms&lt;/strong&gt;: Utilizing external verification mechanisms or information sources to validate the content generated by the model, ensuring it aligns with the real world.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h3 id=&#34;2-the-amnesia-problem&#34;&gt;2. The Amnesia Problem&#xA;&lt;/h3&gt;&lt;p&gt;The amnesia problem refers to the situation where the model may forget previously mentioned information during long dialogues or complex contexts, resulting in inconsistencies and a lack of contextual integrity in generated content. The main causes of amnesia include:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Model Context Memory Limitations&lt;/strong&gt;: The model may be limited by its contextual memory capabilities, unable to effectively retain and utilize long-term dependencies.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Missing Information in Training Data&lt;/strong&gt;: If the training data lacks examples of long dialogues or complex contexts, the model may not learn the correct methods for retaining and retrieving information.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;</description>
        </item><item>
            <title>Understanding Large Models in Artificial Intelligence</title>
            <link>https://digitalxber.com/posts/note-06-1c1376c74e/</link>
            <pubDate>Sun, 13 Oct 2024 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-06-1c1376c74e/</guid>
            <description>&lt;h2 id=&#34;introduction&#34;&gt;Introduction&#xA;&lt;/h2&gt;&lt;p&gt;In the grand tapestry of artificial intelligence (AI), large models shine like brilliant stars, illuminating the future of technology. They not only reshape our understanding of technology but also quietly trigger transformations across countless industries. However, these intelligent technologies are not without their risks and challenges. In this article, we will unveil the mysteries of large models, sharing their technologies and characteristics, analyzing their development and challenges, and offering a glimpse into the AI era.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 1&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;240px&#34; data-flex-grow=&#34;100&#34; height=&#34;1080&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-a0e7ddf8f8.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-a0e7ddf8f8_hu_cd69e27c0275c66.jpeg 800w, https://digitalxber.com/posts/note-06-1c1376c74e/img-a0e7ddf8f8.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;Large models, such as the Generative Pre-trained Transformer (GPT) series, have achieved remarkable success in the field of natural language processing (NLP), setting new performance benchmarks in various language processing tasks. Beyond language, large models also demonstrate significant advantages in image processing, audio processing, and physiological signals. They have rapidly found applications in fields like education, healthcare, and finance, particularly excelling in content generation. Today, numerous cutting-edge technologies related to large models are still in urgent need of development, while issues such as bias and privacy breaches require immediate attention. This article analyzes the past and present of large models, discusses pressing issues, and explores future directions, helping the public quickly understand large model technology and its development, integrating into the tide of the AI era.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 2&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;218px&#34; data-flex-grow=&#34;91&#34; height=&#34;45&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-6f3a983336.jpeg&#34; width=&#34;41&#34;&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;origins-of-large-models&#34;&gt;Origins of Large Models&#xA;&lt;/h2&gt;&lt;p&gt;In November 2022, the renowned AI research company OpenAI released ChatGPT, an AI chatbot program based on the large language model GPT-3.5. Its fluent language expression, powerful problem-solving abilities, and vast database garnered widespread attention worldwide. Within less than two months of its launch, ChatGPT surpassed 100 million monthly active users, becoming the fastest-growing consumer application in history. As a result, various industries began to feel the powerful impact of large models, sparking a research boom in large models both domestically and internationally.&lt;/p&gt;&#xA;&lt;p&gt;The origins of large models can be traced back to the early AI research in the 20th century, which primarily focused on logical reasoning and expert systems. However, these methods were limited by hard-coded knowledge and rules, making it difficult to handle the complexity and diversity of natural language. With the advent of machine learning and deep learning technologies, along with rapid advancements in hardware capabilities, the training of large-scale datasets and complex neural network models became possible, giving rise to the era of large models.&lt;/p&gt;&#xA;&lt;p&gt;In 2017, Google&amp;rsquo;s introduction of the Transformer model structure, which incorporated self-attention mechanisms, significantly enhanced the ability to model sequences, especially in terms of efficiency and accuracy when handling long-distance dependencies. Subsequently, the concept of pre-trained language models (PLMs) gradually became mainstream. PLMs are pre-trained on large-scale text datasets to capture general patterns of language and are then fine-tuned for specific downstream tasks.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 3&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;427px&#34; data-flex-grow=&#34;178&#34; height=&#34;606&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-e67186468c.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-e67186468c_hu_174270892709ab05.jpeg 800w, https://digitalxber.com/posts/note-06-1c1376c74e/img-e67186468c.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;evolution-of-large-models&#34;&gt;Evolution of Large Models&#xA;&lt;/h2&gt;&lt;p&gt;OpenAI&amp;rsquo;s GPT series models exemplify generative pre-trained models, representing the vanguard of this technology. From GPT-1 to GPT-3.5, each generation has seen significant improvements in scale, complexity, and performance. At the end of 2022, ChatGPT emerged as a chatbot capable of answering questions, writing articles, programming, and even mimicking human conversational styles. Its almost omnipotent answering ability has reshaped people&amp;rsquo;s understanding of the general capabilities of large language models, greatly advancing the development of the NLP field.&lt;/p&gt;&#xA;&lt;p&gt;However, the development of large models is not limited to text. With technological advancements, multimodal large models have begun to emerge, capable of simultaneously understanding and generating various types of data, including text, images, and audio. In March 2023, OpenAI announced the multimodal large model GPT-4, which added image functionality and improved language understanding capabilities, marking an important shift from unimodal to multimodal models. The inherent differences between cross-modal data present new and more complex requirements for the design and training of large models, along with unprecedented challenges.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 4&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;218px&#34; data-flex-grow=&#34;91&#34; height=&#34;45&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-6f3a983336.jpeg&#34; width=&#34;41&#34;&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;characteristics-of-large-models&#34;&gt;Characteristics of Large Models&#xA;&lt;/h2&gt;&lt;p&gt;Large models typically refer to machine learning models with vast parameter counts, especially in applications within NLP, computer vision (CV), and multimodal fields. These models understand and learn human language through pre-training, enabling them to perform tasks such as information retrieval, machine translation, text summarization, and code generation in a human-machine dialogue format.&lt;/p&gt;&#xA;&lt;h3 id=&#34;parameter-count-of-large-models&#34;&gt;Parameter Count of Large Models&#xA;&lt;/h3&gt;&lt;p&gt;The parameter count of large models usually exceeds 1 billion, meaning the model contains over 1 billion learnable weights. These parameters form the foundation for the model&amp;rsquo;s learning and understanding of data, continuously adjusted through training to better map input data to output results. The increase in parameter count is directly related to the model&amp;rsquo;s learning ability and complexity, enabling it to capture finer and deeper data features.&lt;/p&gt;&#xA;&lt;h3 id=&#34;types-of-large-models&#34;&gt;Types of Large Models&#xA;&lt;/h3&gt;&lt;p&gt;Large models can be classified based on their application areas and functions:&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;&lt;strong&gt;Large Language Models&lt;/strong&gt;: Focused on processing and understanding natural language text, commonly used for text generation, sentiment analysis, and question-answering systems.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Visual Large Models&lt;/strong&gt;: Specifically designed to process and understand visual information (such as images and videos), used for tasks like image recognition, video analysis, and image generation.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Multimodal Large Models&lt;/strong&gt;: Capable of processing and understanding two or more different types of input data (e.g., text, images, audio), performing more complex and comprehensive tasks by integrating information from different modalities.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Foundation Large Models&lt;/strong&gt;: Generally refer to models that can be broadly applied to various tasks without a specific application direction during the pre-training phase, learning a vast amount of general knowledge.&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;h3 id=&#34;capabilities-of-large-models&#34;&gt;Capabilities of Large Models&#xA;&lt;/h3&gt;&lt;p&gt;The capabilities of large models lie in their ability to understand and process highly complex data patterns:&lt;/p&gt;&#xA;&lt;ol&gt;&#xA;&lt;li&gt;&lt;strong&gt;Generalization Ability&lt;/strong&gt;: Through pre-training on large datasets, large models learn universal linguistic rules, demonstrating strong generalization abilities when faced with new tasks.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Deep Learning&lt;/strong&gt;: The vast parameter scale and deep network structure enable large models to establish complex abstract representations, understanding the deeper semantics and relationships behind the data.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Context Understanding&lt;/strong&gt;: In language models, large models can capture long-distance dependencies, enhancing their ability to understand context, which is crucial for grasping subtle nuances in language.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Knowledge Integration&lt;/strong&gt;: Large models can integrate and utilize knowledge acquired during pre-training, sometimes exhibiting a degree of common-sense reasoning and problem-solving abilities.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Adaptability&lt;/strong&gt;: Although large models learn general knowledge during pre-training, they can be fine-tuned to adapt to specific tasks, showcasing high flexibility and adaptability.&lt;/li&gt;&#xA;&lt;/ol&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 5&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;218px&#34; data-flex-grow=&#34;91&#34; height=&#34;45&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-6f3a983336.jpeg&#34; width=&#34;41&#34;&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;technologies-behind-large-models&#34;&gt;Technologies Behind Large Models&#xA;&lt;/h2&gt;&lt;p&gt;Current large models are integrated machine learning models capable of processing various types of data. The foundational technologies in these large models aim to understand and generate information across different sensory modalities, enabling tasks such as image description, visual question answering, or cross-modal translation. Here are several key foundational technologies of large models:&lt;/p&gt;&#xA;&lt;h3 id=&#34;transformer-architecture&#34;&gt;Transformer Architecture&#xA;&lt;/h3&gt;&lt;p&gt;Most existing large models are built on the Transformer model (or just the decoder of the Transformer), which captures global dependencies of input data through self-attention mechanisms and can also capture complex relationships between different modality elements. For example, a multimodal Transformer can simultaneously process image pixels and text words, learning their associations through self-attention layers. This allows large models to understand various modalities, such as text and images, and generate long text sequences while maintaining contextual coherence.&lt;/p&gt;&#xA;&lt;h3 id=&#34;supervised-fine-tuning&#34;&gt;Supervised Fine-Tuning&#xA;&lt;/h3&gt;&lt;p&gt;Supervised fine-tuning (SFT) is a traditional fine-tuning method that continues training the pre-trained large model using labeled datasets. Notably, during the training of large models, the SFT phase typically employs high-quality datasets. Additionally, SFT involves adjusting the model&amp;rsquo;s parameters to enhance its performance on specific tasks. For instance, to improve a model&amp;rsquo;s performance in legal consulting, a dataset containing legal questions and professional lawyer responses can be used for SFT. During SFT, the model typically attempts to minimize the difference between predicted outputs and true labels, often achieved through loss functions (like cross-entropy loss). This method is direct and straightforward, allowing for rapid adaptation to new tasks. However, it also has limitations, as it relies on high-quality labeled data and may lead to overfitting on the training data.&lt;/p&gt;&#xA;&lt;h3 id=&#34;reinforcement-learning-from-human-feedback&#34;&gt;Reinforcement Learning from Human Feedback&#xA;&lt;/h3&gt;&lt;p&gt;Reinforcement learning from human feedback (RLHF) is a more complex training method that combines elements of supervised learning and reinforcement learning. The model is first pre-trained on a large amount of unlabeled text, similar to the previous SFT step. Then, human evaluators interact with the model or assess its outputs, providing feedback on its performance, and using human feedback data to train a reward model that can predict scores that human evaluators might assign. Finally, the reward model is used as a signal for reinforcement learning to optimize the original model&amp;rsquo;s parameters. In this process, the model attempts to maximize the expected rewards it receives. The advantage of RLHF is that it helps the model learn more complex behaviors, especially when tasks are difficult to define through simple correct or incorrect labels. Additionally, RLHF can help the model better align with human preferences and values.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 6&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;218px&#34; data-flex-grow=&#34;91&#34; height=&#34;45&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-6f3a983336.jpeg&#34; width=&#34;41&#34;&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;applications-of-large-models&#34;&gt;Applications of Large Models&#xA;&lt;/h2&gt;&lt;p&gt;Large models, with their vast parameter counts, deep network structures, and extensive pre-training capabilities, can capture complex data patterns, demonstrating exceptional performance across multiple fields. They can not only understand and generate natural language but also handle complex visual and multimodal information, adapting to various dynamic application scenarios.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 7&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;362px&#34; data-flex-grow=&#34;151&#34; height=&#34;715&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-3936e08d07.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-3936e08d07_hu_f820f7cea42ad10d.jpeg 800w, https://digitalxber.com/posts/note-06-1c1376c74e/img-3936e08d07.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;h3 id=&#34;applications-in-nlp&#34;&gt;Applications in NLP&#xA;&lt;/h3&gt;&lt;p&gt;Large models have particularly widespread applications in the NLP field. For example, OpenAI&amp;rsquo;s GPT series models can generate coherent and natural text, used in chatbots, automated writing, and language translation, with ChatGPT being a well-known product. In the fintech sector, large models are often used for risk assessment, trading algorithms, and credit scoring. They can analyze vast amounts of financial data, predict market trends, and assist financial institutions in making better investment decisions. In the legal and compliance fields, they can be used for document review, contract analysis, and case studies. Through NLP technology, models can understand and analyze legal documents, enhancing the efficiency of legal professionals. Recommendation systems are another application area for large models. By serializing user behavior data into text, large models can predict user interests and recommend relevant products, movies, music, etc. In the gaming sector, large models can utilize their coding capabilities to generate complex game environments, driving non-player characters (NPCs) to produce different dialogues based on player settings, providing a more realistic gaming experience.&lt;/p&gt;&#xA;&lt;h3 id=&#34;applications-in-image-understanding-and-generation&#34;&gt;Applications in Image Understanding and Generation&#xA;&lt;/h3&gt;&lt;p&gt;Currently, large models possess not only text understanding capabilities but also multimodal understanding capabilities, laying the foundation for their applications in the image domain, such as automatic painting and video generation. These models can mimic artists&amp;rsquo; styles to create new artistic works, assisting human creativity. For instance, OpenAI&amp;rsquo;s Sora, released in February 2024, can generate a video segment that meets user input requirements, providing a more convenient tool for film production. In image processing, models like SegGPT are used for image recognition, classification, and generation. They learn from extensive image data paired with text to identify objects, faces, and scenes in images, playing roles in medical image analysis, autonomous vehicles, and video surveillance. Additionally, in the fields of medicine and biology, multimodal large models can be used for disease diagnosis, drug discovery, and gene editing, extracting useful information from complex biomedical data to assist doctors in making more accurate diagnoses or helping researchers design new drugs.&lt;/p&gt;&#xA;&lt;h3 id=&#34;applications-in-speech-recognition&#34;&gt;Applications in Speech Recognition&#xA;&lt;/h3&gt;&lt;p&gt;Large models also play a significant role in the field of speech recognition. Through deep learning technologies, models can convert speech into text, supporting applications such as voice assistants, real-time speech transcription, and automatic subtitle generation, with mobile voice assistants being a typical example. These models learn from a vast number of speech samples, enabling them to handle various accents, intonations, and noise interference.&lt;/p&gt;&#xA;&lt;p&gt;Moreover, large models can be applied across various industries, including education, healthcare, agriculture, and finance. For example, in the education sector, large models can be used for personalized learning, automatic grading, and intelligent tutoring, providing customized teaching content based on students&amp;rsquo; learning situations to help them learn more efficiently. In summary, large models demonstrate immense potential across various fields through their powerful data processing and learning capabilities. With continuous technological advancements, it is foreseeable that large models will play an increasingly important role in future developments.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 8&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;218px&#34; data-flex-grow=&#34;91&#34; height=&#34;45&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-2f73d74812.jpeg&#34; width=&#34;41&#34;&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;development-of-large-models&#34;&gt;Development of Large Models&#xA;&lt;/h2&gt;&lt;p&gt;In the current AI landscape, large models have become an undeniable trend. With the continuous advancement of deep learning technologies, particularly in NLP and CV fields, large models are driving breakthroughs in cutting-edge technologies with their powerful data processing and pattern recognition capabilities.&lt;/p&gt;&#xA;&lt;p&gt;The development of large models at the technical level benefits from several key factors. First is the innovation of algorithms, especially since the introduction of the Transformer architecture, which has rapidly propelled the development of subsequent models, including BERT, the GPT series, and T5. These models achieve leading performance in multiple NLP tasks through pre-training and fine-tuning strategies. Second is the enhancement of computational power, particularly advancements in graphics processing units (GPUs) and tensor processing units (TPUs), enabling the training of models with tens of billions or even hundreds of billions of parameters. Additionally, the rise of cloud computing platforms has provided the necessary computational resources for training large models. At the same time, large-scale datasets have provided ample &amp;ldquo;nutrition&amp;rdquo; for model training. These datasets typically contain rich linguistic expressions, scene information, and user interactions, enabling models to capture complex data distributions and linguistic patterns.&lt;/p&gt;&#xA;&lt;p&gt;The development of large models at the application level has two main directions: large language models and multimodal large models. In the case of large language models, GPT-3 serves as a milestone, reaching 175 billion parameters and showcasing astonishing language understanding and generation abilities. Following closely, Meta AI&amp;rsquo;s LLaMA series models have become favorites in academic research and industry due to their excellent performance and relatively smaller model sizes. These models not only excel in standard NLP tasks but also exhibit tremendous potential in few-shot learning and transfer learning.&lt;/p&gt;&#xA;&lt;p&gt;Multimodal large models extend upon this foundation, capable of processing and understanding various types of inputs, such as text, images, and audio. OpenAI&amp;rsquo;s DALL-E and CLIP are representative works in this direction, capable of understanding and generating images that correspond to text descriptions or understanding text content through images. Google&amp;rsquo;s SimCLR represents an important exploration in the CV field, effectively extracting image features through contrastive learning. Subsequently, Google&amp;rsquo;s Gemini has made significant strides in native multimodal capabilities, pre-training across different modalities and handling more complex inputs and outputs, such as images and audio. OpenAI&amp;rsquo;s Sora further expands the application range of large models, capable of automatically generating video content based on input text, simulating interactions between characters and environments in both the physical and digital worlds.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 9&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;450px&#34; data-flex-grow=&#34;187&#34; height=&#34;575&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-d6d8bb9077.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-d6d8bb9077_hu_6cfb01c9c401f152.jpeg 800w, https://digitalxber.com/posts/note-06-1c1376c74e/img-d6d8bb9077.jpeg 1080w&#34; width=&#34;1080&#34;&gt;&lt;/p&gt;&#xA;&lt;p&gt;Domestic tech companies are also actively exploring large models. Models such as Baidu&amp;rsquo;s &amp;ldquo;Wenxin Yiyan&amp;rdquo;, Alibaba&amp;rsquo;s &amp;ldquo;Tongyi Qianwen&amp;rdquo;, Huawei&amp;rsquo;s &amp;ldquo;Pangu&amp;rdquo;, and iFLYTEK&amp;rsquo;s &amp;ldquo;iFLYTEK Spark&amp;rdquo; have emerged, demonstrating excellent performance in general language understanding and generation tasks, as well as specialized application capabilities in specific vertical fields like healthcare, law, and tourism. For example, Ctrip&amp;rsquo;s &amp;ldquo;Ctrip Wenda&amp;rdquo; focuses on tourism-related Q&amp;amp;A, NetEase Youdao&amp;rsquo;s &amp;ldquo;Ziyue&amp;rdquo; is applied in education, and JD Health&amp;rsquo;s &amp;ldquo;Jingyi Qianxun&amp;rdquo; aims to provide medical consultation services.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 10&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;218px&#34; data-flex-grow=&#34;91&#34; height=&#34;45&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-2f73d74812.jpeg&#34; width=&#34;41&#34;&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;challenges-of-large-models&#34;&gt;Challenges of Large Models&#xA;&lt;/h2&gt;&lt;p&gt;In the AI field, large models are becoming a hot topic in both academic research and industry due to their powerful processing capabilities and broad application prospects. However, as these models continue to expand, the issues faced at the research frontier are becoming increasingly complex.&lt;/p&gt;&#xA;&lt;h3 id=&#34;model-size&#34;&gt;Model Size&#xA;&lt;/h3&gt;&lt;p&gt;The trade-off between model size and data scale has become a significant challenge. Although model performance often improves with an increase in parameter count, this growth in scale brings substantial computational costs and high demands for data quality. Researchers are searching for optimal balances between model size and data scale under limited computational resources, exploring techniques like data augmentation, transfer learning, and model compression to reduce model size without sacrificing performance, striving to minimize the operational costs of large models.&lt;/p&gt;&#xA;&lt;h3 id=&#34;network-architecture&#34;&gt;Network Architecture&#xA;&lt;/h3&gt;&lt;p&gt;Innovation in network architecture is equally crucial. Almost all existing large models are based on the Transformer architecture. While the Transformer architecture excels at processing sequential data, its low computational efficiency and poor parameter utilization can lead to wasted computational resources. The limitations of the current Transformer architecture have prompted researchers to design new network architectures aimed at improving efficiency and generalization capabilities through enhanced attention mechanisms, introducing sparsity, and adaptive computation. For instance, the state-space-based model Mamba proposed in December 2023 introduces a selection mechanism that significantly addresses the computational efficiency issues of existing Transformer architectures, potentially becoming the next generation of foundational architectures for large models.&lt;/p&gt;&#xA;&lt;h3 id=&#34;prompt-engineering&#34;&gt;Prompt Engineering&#xA;&lt;/h3&gt;&lt;p&gt;In dealing with imbalanced datasets, prompt learning offers a new paradigm for addressing this issue. By embedding specific prompts in input data, prompt learning can improve model performance on minority classes. However, designing effective prompts and determining the robustness of these prompts (ensuring effectiveness across different types of large models) has become a discipline—prompt engineering. Further research is needed to combine well-designed prompts with other large model technologies.&lt;/p&gt;&#xA;&lt;h3 id=&#34;contextual-reasoning&#34;&gt;Contextual Reasoning&#xA;&lt;/h3&gt;&lt;p&gt;As model sizes grow, emergent capabilities such as contextual reasoning have surfaced, indicating that large models may have internalized cognitive and learning mechanisms closer to human understanding. The nature, triggering conditions, and controllability of these emergent capabilities are current research hotspots, requiring further exploration from cognitive science and neuroscience perspectives to provide more reasonable explanations and help people understand the principles behind the emergence of these abilities.&lt;/p&gt;&#xA;&lt;h3 id=&#34;knowledge-updating&#34;&gt;Knowledge Updating&#xA;&lt;/h3&gt;&lt;p&gt;The continuous updating of knowledge is another critical issue faced by large models. As knowledge progresses, the information within models can quickly become outdated. Researchers are exploring ways to enable models to learn continuously and integrate new knowledge while avoiding catastrophic forgetting, keeping the model&amp;rsquo;s knowledge base up to date.&lt;/p&gt;&#xA;&lt;h3 id=&#34;explainability&#34;&gt;Explainability&#xA;&lt;/h3&gt;&lt;p&gt;Despite their outstanding performance in various NLP and machine learning tasks, as the parameter count and network structure of models deepen, the decision-making processes of models become increasingly difficult to explain. The black-box nature of large models makes it challenging for users to understand how models process input data and generate output results. This leads to a passive understanding state, where people only know the model&amp;rsquo;s output but are unaware of why the model made such decisions.&lt;/p&gt;&#xA;&lt;h3 id=&#34;privacy-and-security&#34;&gt;Privacy and Security&#xA;&lt;/h3&gt;&lt;p&gt;The training data of large models may encompass personal identity information, sensitive data, or trade secrets. If these data are not adequately protected, the training process of the model may pose risks of privacy breaches or misuse. Additionally, large models themselves may contain sensitive information, such as memories gained from training on sensitive data, making the models inherently prone to privacy risks.&lt;/p&gt;&#xA;&lt;h3 id=&#34;data-bias-and-misleading-information&#34;&gt;Data Bias and Misleading Information&#xA;&lt;/h3&gt;&lt;p&gt;Large language models may output biased or misleading content, stemming from various factors such as data collection methods, annotators&amp;rsquo; subjective preferences, and social culture. When models are trained on biased data, they may incorrectly learn or amplify these biases, leading to unfair or discriminatory outcomes in practical applications.&lt;/p&gt;&#xA;&lt;p&gt;Addressing these issues is crucial for advancing large model technology and expanding its application scope. Solving each challenge could promote more effective applications of AI in the real world, bringing profound impacts to human society.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 11&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;218px&#34; data-flex-grow=&#34;91&#34; height=&#34;45&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-06-1c1376c74e/img-2f73d74812.jpeg&#34; width=&#34;41&#34;&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;future-of-large-models&#34;&gt;Future of Large Models&#xA;&lt;/h2&gt;&lt;p&gt;As AI technology continues to evolve and the application scenarios for large model technology expand, the future trends of large models are also presenting new characteristics and development directions.&lt;/p&gt;&#xA;&lt;h3 id=&#34;balancing-model-scale-and-efficiency&#34;&gt;Balancing Model Scale and Efficiency&#xA;&lt;/h3&gt;&lt;p&gt;Since large model technology often requires substantial computational resources and storage space, future development trends will focus on maintaining model scale while improving efficiency to meet practical application needs. Currently, sparse expert models are gaining attention as a novel architectural approach. Compared to traditional dense models, sparse expert models reduce computational demands by activating only the model parameters relevant to the input data, thereby enhancing computational efficiency. Google&amp;rsquo;s sparse expert model GlaM, developed in 2023, has seven times more parameters than GPT-3 but reduces energy consumption during training and the computational resources required for inference, outperforming traditional models in various NLP tasks.&lt;/p&gt;&#xA;&lt;h3 id=&#34;deep-integration-of-knowledge&#34;&gt;Deep Integration of Knowledge&#xA;&lt;/h3&gt;&lt;p&gt;Knowledge integration aims to enrich the model&amp;rsquo;s representational and decision-making capabilities by combining information from different data sources and knowledge domains. Currently, large models primarily train and apply to single-domain or single-modality data, such as BERT in the NLP domain and ViT in the CV domain. However, in the real world, text, images, audio, and other types of information are often interrelated, making it difficult for single-modality information to meet the demands of complex scenarios. Therefore, with the continuous development of CV, speech recognition, and other technologies, future large models will place greater emphasis on multimodal integration, processing data from different modalities to achieve the fusion and interaction of multimodal information. This capability for multimodal integration allows large models to better understand and process complex information. Moreover, it may be beneficial to combine large model technology with external knowledge bases to further enhance the model&amp;rsquo;s understanding and application breadth. This means that models can leverage not only their internal language patterns and statistical information but also integrate external structured knowledge for reasoning and decision-making, better addressing complex issues in the real world. Importantly, external knowledge can also enhance the generalization capabilities of large models.&lt;/p&gt;&#xA;&lt;h3 id=&#34;exploration-of-embodied-intelligence&#34;&gt;Exploration of Embodied Intelligence&#xA;&lt;/h3&gt;&lt;p&gt;Embodied intelligence refers to intelligent systems that perceive and act based on a physical body, acquiring information, understanding problems, making decisions, and executing actions through interactions with the environment. The proliferation of large models has significantly accelerated the research and implementation of embodied intelligence. Large language models are becoming key tools to help robots better understand and utilize advanced semantic knowledge. By automating task analysis and breaking them down into specific actions, large model technology makes interactions between robots and humans, as well as physical environments, more natural, enhancing the intelligent performance of robots. For instance, different tasks can be achieved through different large models. By using language large models for learning dialogue, visual large models for map recognition, and multimodal large models for executing physical actions, robots can learn concepts more efficiently and direct actions, while decomposing all instructions for execution, completing automated scheduling and collaboration through large model technology. This comprehensive utilization of different models presents new opportunities and challenges for the intelligent development of robots.&lt;/p&gt;&#xA;&lt;h3 id=&#34;explainability-and-trustworthiness&#34;&gt;Explainability and Trustworthiness&#xA;&lt;/h3&gt;&lt;p&gt;As model scales increase, their internal structures become increasingly complex, making the explainability and trustworthiness of models focal points of concern. First, to enhance model explainability, researchers will focus on developing new methods and technologies that enable large models to clearly explain their decision-making processes and the basis for their generated results. This may involve introducing more transparent model structures, such as transparent neural networks or interpretable attention mechanisms, and developing explanatory algorithms and tools to help users understand model outputs.&lt;/p&gt;&#xA;&lt;p&gt;Secondly, to enhance model trustworthiness, a series of measures will be taken to reduce the likelihood of models producing errors or misleading information. One important direction is to introduce external information sources and provide models with the capability to access and reference these sources. This way, models will be able to access the most accurate and up-to-date information, thereby improving the accuracy and trustworthiness of their output results.&lt;/p&gt;&#xA;&lt;p&gt;At the same time, to increase transparency and trust, models will also provide citations related to external information sources, allowing users to audit these sources to determine their reliability. Notably, while some large models with external information access and citation capabilities have already emerged, such as Google&amp;rsquo;s REALM and Facebook&amp;rsquo;s RAG, this is merely the beginning of development in this field. In the future, more innovations and advancements are expected, with new models like OpenAI&amp;rsquo;s WebGPT and DeepMind&amp;rsquo;s Sparrow further propelling development in this area, laying a more solid foundation for the future applications of large model technology. The future development of large model technology will increasingly emphasize explainability and trustworthiness, which is not only an inevitable trend in technological development but also a reasonable requirement from society for the application of technology. Only by continuously enhancing the explainability and trustworthiness of models can large model technology be better applied across various fields, bringing greater impetus to the development of human society.&lt;/p&gt;&#xA;</description>
        </item><item>
            <title>The Rise of Generative AI: Insights from Industry Experts</title>
            <link>https://digitalxber.com/posts/note-08-5dad711393/</link>
            <pubDate>Thu, 30 Nov 2023 00:00:00 +0000</pubDate>
            <guid>https://digitalxber.com/posts/note-08-5dad711393/</guid>
            <description>&lt;h2 id=&#34;introduction&#34;&gt;Introduction&#xA;&lt;/h2&gt;&lt;p&gt;On November 30, 2022, OpenAI launched ChatGPT, a chatbot that has been hailed as a pivotal moment in human history, likened to the steam engine and the iPhone. This revolutionary technology in generative AI has sparked a wave of innovation across the tech industry, prompting a re-evaluation of software and hardware.&lt;/p&gt;&#xA;&lt;h2 id=&#34;the-impact-of-generative-ai&#34;&gt;The Impact of Generative AI&#xA;&lt;/h2&gt;&lt;p&gt;The past year has witnessed a surge in the value of AI infrastructure providers, enabling advancements in various fields from healthcare to aerospace. However, this technological shift has also generated anxiety about AI&amp;rsquo;s potential threats to human existence and job security. OpenAI itself faced a crisis, nearly experiencing a collapse.&lt;/p&gt;&#xA;&lt;h2 id=&#34;questions-about-the-future&#34;&gt;Questions About the Future&#xA;&lt;/h2&gt;&lt;p&gt;As the industry evolves, several questions arise: What will be the next evolution of large language models? When will the AI chip shortage be resolved? Are we running out of training data? How will the competition among AI models in China unfold? Should the development of AI technologies accelerate or decelerate? Will AGI (Artificial General Intelligence) manifest in different forms? To address these questions, we invited industry experts to share their insights and pose their own questions.&lt;/p&gt;&#xA;&lt;p&gt;&lt;img alt=&#34;Image 1&#34; class=&#34;gallery-image&#34; data-flex-basis=&#34;110px&#34; data-flex-grow=&#34;46&#34; height=&#34;2436&#34; loading=&#34;lazy&#34; sizes=&#34;(max-width: 767px) calc(100vw - 30px), (max-width: 1023px) 700px, (max-width: 1279px) 950px, 1232px&#34; src=&#34;https://digitalxber.com/posts/note-08-5dad711393/img-fa71994df5.jpeg&#34; srcset=&#34;https://digitalxber.com/posts/note-08-5dad711393/img-fa71994df5_hu_fa2433172555cf5d.jpeg 800w, https://digitalxber.com/posts/note-08-5dad711393/img-fa71994df5.jpeg 1125w&#34; width=&#34;1125&#34;&gt;&lt;/p&gt;&#xA;&lt;h2 id=&#34;openais-position-and-competitors&#34;&gt;OpenAI&amp;rsquo;s Position and Competitors&#xA;&lt;/h2&gt;&lt;p&gt;OpenAI, previously unknown to the public, has become one of the most recognized tech companies globally within a year, creating competitive pressure for giants like Google, Meta, and Amazon. Many are curious about the release date of GPT-5 and who might challenge OpenAI&amp;rsquo;s dominance.&lt;/p&gt;&#xA;&lt;p&gt;Zhang Peng, CEO of Zhiyuan AI, stated that while OpenAI is leading, it should not overshadow other competitors. He emphasized that true challengers must have a strong technical foundation and deep understanding. Professor Xiao Yanghua from Fudan University noted that once a model approaches AGI, its upgrade and evolution could be astonishingly rapid, raising concerns about widening gaps.&lt;/p&gt;&#xA;&lt;h2 id=&#34;user-growth-and-challenges&#34;&gt;User Growth and Challenges&#xA;&lt;/h2&gt;&lt;p&gt;After explosive early growth, OpenAI&amp;rsquo;s user growth has slowed, which is considered normal. Wang Xiaohang, VP of Ant Group, mentioned that the evolution of model capabilities is data-driven, and the availability of publicly accessible data is dwindling. He suggested two paths forward, emphasizing that AGI, as a centralized product, is not yet a high-frequency necessity for the public. The industry trend is shifting towards creating a super ecosystem rather than a centralized super AI.&lt;/p&gt;&#xA;&lt;h2 id=&#34;future-directions-for-large-language-models&#34;&gt;Future Directions for Large Language Models&#xA;&lt;/h2&gt;&lt;p&gt;In interviews, industry leaders proposed several directions for the evolution of large language models. Liu Qingfeng, chairman of iFlytek, highlighted the need for larger model parameters, the creation of AI personas, and deeper customization across various industry scenarios. Wang Fengyang from Baidu emphasized the importance of intelligent agents, while Zhou Bowen from Xiangyuan Technology discussed the potential for AI to effectively use tools, which he termed tool intelligence.&lt;/p&gt;&#xA;&lt;h2 id=&#34;the-competitive-landscape-in-china&#34;&gt;The Competitive Landscape in China&#xA;&lt;/h2&gt;&lt;p&gt;Following the launch of ChatGPT, Chinese tech companies entered a heated competition dubbed the &amp;ldquo;Hundred Model War,&amp;rdquo; involving both established firms and rapidly funded startups. Chen Lei from Xinyi Technology predicted a more objective market in the coming year, with a focus on practical applications and a reduction in the number of foundational models.&lt;/p&gt;&#xA;&lt;h2 id=&#34;open-source-vs-closed-source-models&#34;&gt;Open Source vs. Closed Source Models&#xA;&lt;/h2&gt;&lt;p&gt;As OpenAI becomes less open about its model parameters and training details, the question arises: Can open-source models surpass closed-source ones? Liang Jiaen, chairman of Yunzhisheng, estimated that while open-source models might have a broader impact in terms of application quantity, closed-source models may achieve higher performance levels.&lt;/p&gt;&#xA;&lt;h2 id=&#34;insights-from-industry-experts&#34;&gt;Insights from Industry Experts&#xA;&lt;/h2&gt;&lt;p&gt;The following are key insights from the interviews:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Will GPT-5 be released?&lt;/strong&gt; Experts agree that further iterations like GPT-5, GPT-6, and GPT-7 are inevitable, but the timeline remains uncertain due to market conditions and safety evaluations.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Who will challenge OpenAI?&lt;/strong&gt; Competitors include tech giants like Microsoft and Google, as well as startups like Anthropic and Inflection AI. The consensus is that achieving AGI is a common goal among these entities.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;How to address OpenAI&amp;rsquo;s slowing growth?&lt;/strong&gt; Experts suggest that the slowdown in user growth is normal and that the focus should shift to integrating AI into various industries to create real demand.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;h2 id=&#34;conclusion&#34;&gt;Conclusion&#xA;&lt;/h2&gt;&lt;p&gt;The future of generative AI is filled with potential and challenges. As the industry continues to evolve, the focus will be on practical applications, the development of intelligent agents, and the integration of AI into everyday tasks. The competition will not only be about who can create the best models but also about who can effectively apply these technologies to solve real-world problems.&lt;/p&gt;&#xA;</description>
        </item></channel>
</rss>
