ACTA AERONAUTICAET ASTRONAUTICA SINICA >
Avionics embedded system benchmark test method
Received date: 2015-05-28
Revised date: 2015-07-20
Online published: 2015-07-30
Supported by
National Science and Technology Major Project Foundation of China (2012ZX01041-006);Aeronautic Science Foundation of China (2012ZC31006,2013ZC31005);MIIT Specific Research Projects Foundation for Civil Aircraft (MJ-S-2012-05)
Performance evaluation is an important method to find the performance bottleneck of embedded computer systems, guide equipment selection, balance related components, optimize the system structure and improve the system's actual performance. After the introduction of indicators and benchmarks of embedded computer systems performance evaluation, we analyze the characteristics of benchmark test methods and test procedures, build different test environments, as well as the comparative testing principle of benchmark method, and the main testing process. Then, select the SPEC CPU 2000 benchmarks set, cut and transplant them under the airborne embedded environment for comparison test on a domestic airborne embedded real-time operating system using different embedded operating systems, embedded file systems and compiler options as variable parameters. Conclusions show that the performance of a domestic real-time operating system for airborne embedded computer is roughly equal to similar foreign commercial operating system, file system has less effect on the benchmark results and compiler optimization can significantly improve the performance of embedded systems.
HAO Yukai , CUI Xining , LI Leilei , YANG Qiong . Avionics embedded system benchmark test method[J]. ACTA AERONAUTICAET ASTRONAUTICA SINICA, 2016 , 37(4) : 1327 -1335 . DOI: 10.7527/S1000-6893.2015.0204
[1] 罗志强. 航空电子综合化系统[M]. 北京:北京航空航天大学出版社, 1990:26-31. LUO Z Q. Integrated avionics systems[M]. Beijing:Beijing University of Aeronautics and Astronautics Press, 1990:26-31(in Chinese).
[2] Aeronautical Radio, Inc. ARINC specification 653:Avionics application software standard interface part 1 required services[S]. Annapolis:Aeronautical Radio, Inc., 2006.
[3] Aeronautical Radio, Inc. ARINC speecification 653:Avionics application software standard interface part 2 extended services[S]. Annapolis:Aeronautical Radio, Inc., 2007.
[4] 高晓光, 薛亚勇, 温增葵. IMA双层调度算法中的任务可调度性分析方法[J]. 航空学报, 2015, 36(2):585-595. GAO X G, XUE Y Y, WEN Z K. Task schedulability analyzing method of two-level hierarchical scheduling algorithm in integrated modular avionics[J]. Acta Aeronautica et Astronautica Sinica, 2015, 36(2):585-595(in Chinese).
[5] 张海涛, 龚龙庆. 用于嵌入式计算机性能评测技术及其方法研究[J]. 现代电子技术, 2008, 31(14):49-54. ZHANG H T, GONG L Q. Research of embedded computer performance evaluation technology[J]. Modern Electronics Technique, 2008, 31(14):49-54(in Chinese).
[6] 李向东, 张晓红, 戴小氐, 等. 机载嵌入式系统中多种机内自测试的软件设计[J]. 计算机测量与控制, 2014, 22(4):1310-1313. LI X D, ZHANG X H, DAI X D, et al. Design of build-in test in embedded systems onboard[J]. Computer Measurement & Control, 2014, 22(4):1310-1313(in Chinese).
[7] SIEH V, TSCHACHE O, BALBACH F. VERIFY:Evaluation of reliability using VHDL models with embedded fault descriptions[C]//Proceedings of the 27th International Symposium on Fault-Tolerant Computing. Piscataway, NJ:IEEE Press, 1997:32-36.
[8] YIN Y F, LIU B, NI H Y. Real-time embedded software testing method based on extended finite state machine[J]. Journal of Systems Engineering and Electronics, 2012, 23(2):276-285.
[9] ARLAT J. Conceptual framework, deliverable CF1, state of the art:IST 2000-25425[R]. 2001.
[10] HENNESSY J L, PATTERSON D A. Computer architecture:A quantitative approach[M]. 5th ed. San Francisco:Morgan Kaufmann, 2011:125-135.
[11] WEISS A R, CLUCAS R. The standardization of embedded benchmarking:The pitfalls and the opportunities[C]//Proceedings of Embedded Systems Conference. Piscataway, NJ:IEEE Press, 1999:492-508.
[12] HALANG W A, GUMZEJ R, COLNARIC M, et al. Measuring the performance of real-time systems[J]. The International Journal of Time-Critical Computing System, 2000, 18(1):59-68.
[13] WANG A J A. Using Java Hartstone benchmark in a realtime systems course[C]//Proceedings of 201232nd Annual Frontiers in Education. Piscataway, NJ:IEEE Press, 2002:S2F-25-S2F-30.
[14] KRISHNASWAMY U, SCHERSON I D. A framework for computer performance evaluation using benchmark sets[J]. IEEE Transactions on Computers, 2000, 49(12):1325-1338.
[15] 徐钦桂, 李勇, 张杰明, 等. 基于比较策略的嵌入式系统性能基准测试研究[J]. 计算机工程与应用, 2005, 41(19):93-96. XU Q G, LI Y, ZHANG J M, et al. Research of performance benchmark for embedded systems based on comparison strategy[J]. Computer Engineering and Applications, 2005, 41(19):93-96(in Chinese).
[16] Standard Performance Evaluation Corporation. SPEC CPU 2000 V1.3[EB/OL]. (2007-6-7)[2015-1-15]. http://www.spec.org/cpu2000/index.html.
[17] KLEINOSOWSKI A, LILJA D J. MinneSPEC:A new SPEC benchmark workload for simulation-based computer architecture research[R]. Minnesota:University of Minnesota ARCTiC Labs, 2002.
[18] 葛仁北. SPEC CPU 2000基准程序运行路径分析[J]. 计算机工程, 2007, 33(7):38-40. GE R B. Path analysis of SPEC CPU 2000 benchmark[J]. Computer Engineering, 2007, 33(7):38-40(in Chinese).
[19] HENNING J L. SPEC CPU 2000:Measuring CPU performance in the new millennium[J]. Computing Practices, 2000, 33(7):25-26.
[20] Wind River Systems Inc. Tornado user's guide (Windows Version 1.0)[M]. Alameda, CA:Wind River Systems Inc., 1996:34-40.
[21] Wind River Systems Inc. VxWorks programmer's guide 5.5[M]. Alameda, CA:Wind River Systems Inc., 1996:62-65.
/
〈 | 〉 |