• 大模型在航天嵌入式软件测试中的应用研究

    Research on the application of large models in embedded software testing for spacecraft

    • 大语言模型(Large Language Model, LLM),尤其是一些预训练的深度神经网络,拥有强大的表征学习能力,不仅能学习典型的软件缺陷,还能够深度分析软件隐含的缺陷,给出风险提示和改进建议。大模型的应用,使得测试方法和流程更加智能化,且能够发现之前难以捕捉的复杂缺陷。航天器软件以其独特性、专业性、复杂性、高可靠性、高安全性的特点,需要通过专业的方法训练专业的大模型来满足航天嵌入式软件测试的需要。研究表明,大模型结合有效的测试方法形成的完备思维链(Chain of Thought, Cot)可以有效提高软件缺陷检出率。将测试专家的思维和大模型相结合,用专家思维链引导大模型查找软件缺陷,在提示词中增加测试专家的思维链可以有效提升软件缺陷检出效率,保证软件质量,显著提升测试团队的整体专业水平。

       

      Abstract: Large Language Model (LLM), especially some pre-trained deep neural networks, have strong representation learning capabilities and can automatically learn abstract features and potential error patterns from software code. Large models, with their strong natural language processing capabilities, can not only learn typical software defects but also deeply analyze hidden defects in software and provide risk warnings and improvement suggestions. The application of large models makes testing methods and processes more intelligent and can discover complex defects that are difficult to capture before. Spacecraft software has unique, specialized, complex, high reliability, and high security features, and requires specialized methods to train specialized large models to meet the needs of space embedded software testing. Research shows that a test defect set refined from a large amount of test data, combined with an effective testing method, can train a dedicated large model. Experimental results show that combining the expert's thinking with the large model and guiding the large model to find software defects using the expert's thinking chain can effectively improve code quality and defect detection efficiency, significantly improve the overall professional level of the testing team.

       

    /

    返回文章
    返回