Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Astron. Space Sci., 19 September 2025

Sec. Astronomical Instrumentation

Volume 12 - 2025 | https://doi.org/10.3389/fspas.2025.1657487

A novel space intelligent computing and data processing architecture--the spacecraft payload health management unit (SPHMU)

Wenjie Zhao,Wenjie Zhao1,2Jianing Rao,
Jianing Rao1,2*Miao MaMiao Ma1Jianzhe ZhangJianzhe Zhang1Shuanglong LiShuanglong Li1Wenyuan JiaWenyuan Jia1Zhenxing Dong,Zhenxing Dong1,2Baocheng HouBaocheng Hou1Guofeng XueGuofeng Xue1Guodong Yin,Guodong Yin1,2Yan Zhu,Yan Zhu1,2Junshe An,Junshe An1,2
  • 1National Space Science Center, The Chinese Academy of Sciences, Beijing, China
  • 2University of Chinese Academy of Sciences, Beijing, China

This paper primarily introduces a novel on-board intelligent computing and intelligent data processing architecture. This architecture leverages the collaborative work of high reliability computers (HRC) and high performance computers (HPC) to achieve highly reliable payload management and data interaction, as well as high-performance intelligent computing and data processing objectives in space application environments. This paper implements the novel on-board intelligent computing and intelligent data processing architecture using a combination of HRC and HPC. Specifically, HRC is responsible for data interaction with the spacecraft bus, including telemetry, telecommand, and control (TT&C) information of various payloads, as well as time and space information. HPC, on the other hand, primarily executes application software (APP) to process the received payload data. This includes computationally intensive tasks such as attitude and orbit determination. Communication between HRC and HPC is facilitated through a high reliability field programmable gate array (FPGA). This paper verifies the stability of this architecture through rigorous testing at the aerospace level. The novel on-board intelligent computing and intelligent data processing architecture possesses excellent computing capabilities, spatial adaptability, reliability, and a flexible structure, laying a foundation for high-performance computing in space.

1 Introduction

The rapid advancement of China’s aerospace technology has driven the increasing prioritization of on-board computational capabilities and data processing efficiency across diverse space missions. These missions, spanning from space station operations to near-Earth surveillance, medium/high-orbital exploration, and deep-space investigations, now mandate stringent requirements for equipment miniaturization, weight reduction, and power consumption optimization.

Consequently, highly integrated intelligent load management systems are becoming critical components in modern aerospace platforms. Such devices exhibit compact physical dimensions, ultra-lightweight construction, multi-core heterogeneous computing architectures, and radiation-hardened reliability characteristics essential for next-generation mission-critical applications.

A Payload Management Unit (PMU) serves as a critical component in various systems, ensuring the effective operation and integration of payloads. These systems range from spacecraft and satellites to unmanned aerial vehicles (UAVs) and industrial equipment. The PMU’s functions include power distribution, data handling, thermal management, and communication interface, all tailored to the specific requirements of the payload and the overall system (Del Castillo et al., 2025; Zhou and An, 2013).

In spacecraft, a PMU is essential for managing the power, data, and control signals between the spacecraft and its scientific instruments or other payloads. These systems highlight the importance of efficient payload management in enhancing mission readiness and mitigating risks (Del Castillo et al., 2025). Traditional PMUs inherently lack high-performance computational capabilities, as illustrated in Figure 1, necessitating continuous advancements in on-board high-performance computing (OHPC) to meet the evolving demands of space missions.

Figure 1
Diagram illustrating a satellite data system. Multiple computing devices connect to various payloads, which are linked to a PMU containing HRC, a low-speed bus, and FPGA. The PMU interfaces with a satellite data platform via a 1553B CAN connection.

Figure 1. Traditional payload management unit.

In response to these evolving requirements, we have designed a highly innovative Space Payload Health Management Unit (SPHMU) for a specific spacecraft program, realizing a high-integration, high-reliability device for spaceborne intelligent computing and intelligent data processing. Compared to traditional payload management units, this SPHMU retains a High-Reliability Computer (HRC) while integrating a High-Performance Computer (HPC) dedicated to high-performance space computing and data processing; thus, it not only maintains the fundamental payload management and data processing functions of traditional units but also provides significantly enhanced computational capabilities for tasks such as attitude and orbital determination, orbital maneuver calculation, and autonomous operations, establishing a robust high-performance software and hardware platform that offers excellent adaptability and scalability for supporting diverse scientific mission objectives (Abdu et al., 2023; Ramírez et al., 2022; Ben Yahia et al., 2025). The dimensional and mass characteristics of the SPHMU are presented in Table 1.

Table 1
www.frontiersin.org

Table 1. The dimensional and mass characteristics of the SPHMU.

As illustrated in Figure 1, the traditional payload management unit solely facilitates high-speed data transmission and data processing between various payloads, as well as high-speed data transfer and telemetry and telecommand (TT&C) operations with the Satellite data platform; however, this necessitates distributing the data returned from the spacecraft bus back to individual payloads, leading to distributed computations and data processing among the payloads themselves which impose significant resource overhead in terms of weight and power consumption, while simultaneously the transmission of this data substantially consumes bus bandwidth resources (Bhandari et al., 2024).

The operational architecture of the SPHMU (Space Payload Health Management Unit) designed for a specific mission model is shown in Figure 2, embodying a highly integrated and reliable platform for spaceborne intelligent computing and intelligent data processing. Compared to traditional payload management units, this SPHMU retains a High-Reliability Computer (HRC) while integrating a High-Performance Computer (HPC) dedicated to spaceborne high-performance computing and data processing. Consequently, it not only maintains fundamental payload management and data processing capabilities inherent to conventional units but also delivers advanced computational functionalities, including real-time precision orbit determination software, autonomous orbital maneuver software for elliptical trajectories, and intelligent on-orbit data analysis with anomaly alert systems. These functionalities are deployed as software applications (Apps) via a software-defined hardware platform, enhancing operational flexibility and mission adaptability (Hui et al., 2025; Morton et al., 2021).

Figure 2
Diagram illustrating a satellite data management system. Multiple payloads connect to an SPHMU, which contains HRC, FPGA, and HPC components. Data flows through low-speed and high-speed buses. The system connects to a satellite data platform via 1553B CAN.

Figure 2. SPHMU work mode.

2 System design of SPHMU

2.1 Overall hardware design

In order to adapt to different data interfaces with different loads, the designed SPHMU hardware includes RS422 data interfaces, LVDS data interfaces, AD data acquisition interfaces, OC output data interfaces, Ethernet interfaces, USB interfaces, 1553B bus interfaces, CAN bus interfaces, Mass storage interfaces and program data storage interfaces. These interfaces can connect with actual payload electronics and the satellite data platform, providing sufficient data processing capabilities and a sufficient number of data interfaces, making it practically applicable to various payload electronics through different configurations.

The SPHMU comprises five integrated components as illustrated in Figure 3: a Power Supply Unit, an HRC (High-Reliability Computer) Unit, an HPI (High-Performance Interface) Unit, and two HPC (High-Performance Computer) Units (HPC-A and HPC-B). The Power Supply Unit converts bus power into secondary power outputs (e.g., 5V, 3.3V) required by the system (Martins et al., 2024).

Figure 3
Five color-coded rectangular blocks labeled

Figure 3. The SPHMU design block diagram.

The HRC Unit, responsible for external data interfaces, data storage, and control functions, collects payload status from the HPI Unit and exchanges data with the spacecraft management system via a configurable MIL-STD-1553B or CAN bus. It further monitors and manages the current, voltage, and operational status of the HPC Units. Architecturally, the HRC integrates a radiation-hardened CPU, SDRAM, NOR Flash, a watchdog timer, a refresh controller, transformers, and MIL-STD-1553B transceivers. Its core employs the Loongson XX300 processor from Loongson Technology, designed for space applications with a total ionizing dose (TID) tolerance ≥300 krad(Si) and a single-event latch-up (SEL) threshold ≥75 MeV·cm2/mg. To mitigate single-event upsets (SEUs), the CPU implements error-correcting code (ECC) for SDRAM; additionally, it supports dual program memory sectors for booting—selectively activated during power-on reset or watchdog reset—thus avoiding system failure due to corruption in a single memory sector and enhancing overall reliability.

The High-Performance Interface (HPI) Unit primarily facilitates external communication via RS422 interfaces, LVDS interfaces, analog signal acquisition, switch command (OC) control, and memory management, utilizing a hardware architecture comprising an FPGA, analog multiplexers, amplifiers, ADCs, OC circuits, and NAND Flash. Specifically, the FPGA—implemented using a SRAM-based device from Xilinx—operates as a critical bridge between the HRC and HPC units: it exchanges scientific data with the HPC units via a high-speed internal bus while simultaneously transmitting status information to the HRC unit through a low-speed internal bus, thus serving as the core interconnect mechanism for bidirectional data and control pathways (Zhao et al., 2022).

The High-Performance Computing (HPC) Unit primarily comprises a high-performance CPU, peripheral memory modules, and interface circuitry, collectively enabling the bootloading of the Linux operating system, execution of user-installed applications, and processing of diverse data transmitted from the High-Reliability Monitoring Module, as well as performing critical functions such as precision orbit determination, elliptical orbit maneuver calculations, and intelligent data analysis with anomaly alerting. Designed as a standardized board-level solution, this unit integrates a Loongson XX2000 CPU (manufactured by Loongson Technology), DDR4 SDRAM, SPI Flash, eMMC storage, watchdog circuitry, Ethernet ports, USB interfaces, HDMI output, DC/DC converters, and debugging circuits. The XX2000 CPU—fabricated using a 28 nm FD-SOI process and meeting industrial-grade (IND) quality standards—exhibits a total ionizing dose (TID) tolerance exceeding 100 krad(Si) and complete immunity to single-event latch-up (SEL). Architecturally, the CPU incorporates dual LA364 cores operating at 0.8–1.2 GHz, a 72-bit DDR3/4 memory controller, and comprehensive system peripheral I/Os, thereby consolidating computational robustness with radiation-hardened reliability for spaceborne applications.

2.2 1553B&CAN bus design

The SPHMU employs a configurable dual-bus architecture integrating both MIL-STD-1553B and CAN bus standards as its critical communication backbone (Figures 4, 5), with meticulous design considerations paramount due to their fundamental role in mediating payload status monitoring and spacecraft command execution; specifically, this architecture implements cold redundancy with duplicated A/B channels for both bus types, significantly enhancing system fault tolerance through hardware-based redundancy while strictly adhering to spacecraft bus design specifications, thereby ensuring robust signal integrity through controlled impedance routing, electromagnetic interference (EMI) shielding, and validated signal termination protocols across all mission operating conditions.

Figure 4
Diagram of two 1553B bus configurations labeled SPHMU-A and SPHMU-B. Each setup shows a 1553B controller connected via isolation transformers and couplers, leading to both A-Bus and B-Bus, with impedance values marked as seventy to eighty-five ohms. Resistors of 0.75 ohms are integrated in the couplers for both configurations.

Figure 4. The complex 1553B Bus construction standards.

Figure 5
Diagram of a CAN bus circuit with components labeled. Two sections, SPHMU-A and SPHMU-B, have identical configurations. Each section contains two TJA1040T transceivers connected to two SJA1000T controllers, which in turn connect to a CPU. The CAN Bus IN and OUT lines are depicted, with resistors labeled 120 ohms between the TJA1040Ts. Dashed lines represent the bus connections across the diagram.

Figure 5. The complex CAN Bus construction standards.

RS422 and LVDS are both signal transmission technologies widely used in industrial, communication, and embedded systems, characterized by their high efficiency, stability, and strong anti-interference capabilities.

Taking the 1553B bus as an example, each word (command word/data word/status word) is fixed at 20 bits (3 synchronization heads+16 data bits+1 odd parity bit), with a transmission rate of 1 Mbps. The transmission time of a single word is Equation 1:

Tword=201Mbps=20μs(1)

Considering the message interval (standard requirement ≤14 μs) and protocol overhead (command word, status word), the actual effective bandwidth is significantly reduced. For example, the actual efficiency of transmitting 32 data words is approximately (Equation 2):

ηactual=32×1634×20+1468%(2)

To reduce reflection and noise, the terminal resistance needs to match the cable characteristic impedance (usually 78 Ω) and be verified by the following Equation 3:

Zmatch=Zcable2(3)

The characteristic impedance of twisted pair cable is about 78 Ω.When indirectly coupled, the terminal resistor is directly integrated onto the coupler.

Taking CAN bus as an example, the explicit level requirement is Equation 4:

Vdiff_dom=VCAN_HVCANL1.5V(4)

Implicit level requirement (Equation 5):

Vdiff_dom=VCAN_HVCANL0.5V(5)

Common mode voltage tolerance (Equation 6):

2VVCM7V(6)

Terminal resistors need to be deployed at both ends of the bus to suppress signal reflection (Equation 7):

Rterm=Z02=120Ω±5%(7)

The formula for signal quality assessment is (Equation 8):

Stotal=Wedge×Sedge+Wamp×Samp+Wref×SrefWtotal(8)

Typical settings (Wedge, Wamp, Wref) = (50%, 25%, 25%)

Sedge scores the edge rate (0%–100%), and scores 100% when the edge time tedge is less than 10% and the bit time is less than 10%

Samp is a stable amplitude rating.

Samp = 100%. When Udistortion = Vdiff_dom ≥2.2V, Udisturb is the minimum difference between explicit/implicit levels.

Sref: Reflection Distortion Score. Sref = 100% when Upp = Udisturb (no overshoot/undershoot).

The transmission speed of CAN bus is relatively high, so twisted pair cables are used in PCB design, with a line length matching error of ≤5 mil to avoid branching (stub length <3 cm).

2.3 HPC hardware design

The HPC unit employs Loongson Zhongke’s XX2000 processor chip, interfacing with a 32 GB eMMC module via the eMMC interface to store the operating system and application data, where the eMMC chip is specifically selected for single-event latch-up (SEL) immunity to mitigate radiation-induced failures in space environments. To enhance system reliability, triple-redundant kernel images are stored in partitioned sectors of the eMMC, and during boot initialization, a two-out-of-three voting mechanism is executed to validate kernel integrity, thereby preventing boot failures caused by anomalies in any single storage region. The SPI Flash is dedicated to hosting the CPU’s bootloader firmware, while the eMMC exclusively maintains the full operating system and application binaries.

For computational operations, the HPC unit receives multi-channel telemetry data through inter-layer connectors from the HPI unit for on-orbit real-time processing, concurrently feeding its own operational telemetry back to the high-reliability monitoring unit. Comprehensive debugging and expansion interfaces—including serial port (UART), USB, Gigabit Ethernet, and HDMI—are integrated into the high-performance CPU to facilitate hardware/software diagnostics and peripheral connectivity.

2.4 Space reliability design of SPHMU

As illustrated in Figure 2, scientific data from external payloads are transmitted via multiple bus interfaces to the FPGA within the SPHMU for on board storage and preliminary processing. The integrated data processing unit employs a dual-layer redundancy architecture comprising primary redundancy and cross-redundancy mechanisms to ensure operational reliability in space environments.

As shown in Figure 6, the design block diagram of the data processing unit (DPU) for SPHMU is presented. Due to the connection of different payloads, redundant system backup design and hardware interface cross-backup design are implemented. Radiation-hardened high-reliability components are selected to avoid single-point failures, thereby ensuring equipment reliability.

Figure 6
Diagram showing a dual data processing unit (DPU) architecture with two identical sections labeled DPU A and DPU B. Each section contains an FPGA connected to NAND flash memory and power modules. Both sections interface with payloads and include components labeled SG2003, OP27+TLV2548+AD854, and RS422/LVDS/TLK2711. The units are connected via an inner connector.

Figure 6. The DPU of SPHMU.

The external data bus of the data processing unit integrates LVDS, RS422, and 2,711 bus architectures, achieving a maximum operational bandwidth of 2.4 Gb/s. The internal bus system comprises HPI (HRC), PCIe 2.0 (HPC), and LIO (HPC) interconnects—three CPU-native communication protocols with peak data transfer rates up to 1 GB/s. FPGA and HPC achieve slow and fast data interaction through the CPU’s built-in LIO bus (1 Mbps) and PCIe bus (1 GB/s) respectively, as shown in Figure 7.

Figure 7
Diagram illustrating data flow between an HPC and an FPGA. Low-speed data signals are shown with names like LIO_ADDR, LIO_DATA, LIO_CS, and others, moving between the two components. A PCIe bus at the bottom is labeled as high-speed data.

Figure 7. Data transmisson aarchitecture of FPGA and HPC.

For space environmental reliability, the majority of components employ radiation-hardened devices with a total ionizing dose (TID) tolerance ≥300 krad(Si) and single-event latch-up (SEL) threshold ≥75 MeV·cm2/mg. For components not meeting single-event effect (SEE) requirements, mitigation measures including tantalum shielding implementation, ECC-based software redundancy, and FPGA periodic refresh mechanisms are integrated. Additionally, a homogeneous primary/backup architecture is adopted for critical systems, with redundant interfaces employing dual-channel failover protocols to ensure operational continuity under extreme space radiation conditions.

2.5 Functional description of SPHMU

The SPHMU module supports a comprehensive set of interface standards with specific directionality and quantity configurations as enumerated in Table 2, including LVDS (Low-Voltage Differential Signaling), RS422 (differential serial communication), 1553B (MIL-STD-1553 avionics bus), CAN (Controller Area Network), OC, AD (Analog-to-Digital), mass storage interfaces, USB 3.0 (Universal Serial Bus 3.0), HDMI 2.1 (High-Definition Multimedia Interface 2.1), and Gigabit Ethernet.

Table 2
www.frontiersin.org

Table 2. Interfaces Configuration by SPHMU.

The test setup between the universal ground test (UGT) equipment and the SPHMU is shown in Figure 8.

Figure 8
Two black electronic devices on a green table are interconnected with numerous white and blue cables. The devices feature several connectors and mounted components on top.

Figure 8. The UGT with SPHMU.

3 Software design of SPHMU

In the previous section, We have proposed a detailed hardware design for the SPHMU. This design is compatible with most general interfaces, different construction standards, and communication protocols, and it provides an ample number of interfaces. This section primarily focuses on the software design of the SPHMU. Given the foundational nature of the software design implemented in both the HRC (High-Reliability Computer) unit and the HPI (High-Performance Interface) unit, this work concentrates its primary focus on detailing the software architecture and computational workflows deployed within the HPC (High-Performance Computing) unit, which encapsulates the system’s advanced algorithmic processing, autonomous decision-making capabilities, and mission-specific application frameworks essential for next-generation spaceborne intelligence (Bhargav Reddy and Vijayreddy, 2025; Bhandari et al., 2024).

3.1 HRC unit software design

The HRC (High-Reliability Computer) unit management software is architecturally partitioned into four functional modules:

MIL-STD-1553B bus communication management, RS422 payload communication management, engineering telemetry management, and health monitoring with operational maintenance.

The 1553B communication module facilitates deterministic data exchange with the spacecraft management computer, encompassing reception of command directives and broadcast time-code messages alongside transmission of device engineering parameters.

Engineering telemetry management systematically aggregates operational status metrics from individual subsystems into consolidated telemetry parameters, periodically transmitting these datasets to the spacecraft management computer via the 1553B bus while concurrently packaging the Payload Health Manager’s engineering parameters into standardized CCSDS-compliant telemetry source packets.

Health and maintenance operations integrate system preservation capabilities, encompassing initialization routines, time synchronization protocols, Error Detection and Correction (EDAC) processing, 1553B link integrity monitoring, and on-orbit software reconfiguration functionalities essential for sustained mission longevity.

3.2 HPC unit software design

Utilizing the Loongson-embedded operating system, the HPC unit delivers core functionalities including communication protocols for external devices and FPGA interfacing mechanisms, thereby establishing a robust foundation-layer infrastructure for upper-level application software.

The application support middleware orchestrates dynamic startup and termination of all software applications executing on the XX2000 processor, wherein it ingests inbound data packets and command instructions through standardized interfaces, executes protocol-compliant unpacking and parsing operations, and rigorously validates data integrity and command legality—automatically discarding invalid or noncompliant transmissions. Following command header interpretation, it autonomously executes locally managed directives while concurrently routing application-tier commands to designated targets. Beyond reception duties, the system aggregates internal state diagnostics and application runtime telemetry into unified CCSDS-compliant downlink packets.

Additionally, it administers essential maintenance operations encompassing system initialization, precision time synchronization, in-flight software updates, and cross-platform configuration management to sustain deterministic operations in space environments.

The HPC unit application software integrates four core functional modules: command and telemetry management, health monitoring with operational maintenance, and application scheduling and oversight.

The command management module facilitates deterministic data exchange via RS422 interface with the high-reliability monitoring CPU, executing reception of spacecraft management computer directives and broadcast time-code synchronization signals. Telemetry management systematically aggregates operational status parameters through periodic sampling of internal system diagnostics, structures these metrics into standardized engineering datasets for transmission to the high-reliability monitoring CPU over dedicated internal buses, and further orchestrates autonomous packaging of its operational parameters into consolidated CCSDS-compliant telemetry source packets.

Health and maintenance operations implement essential system preservation capabilities, including initialization sequences, precision time protocol synchronization, cross-bus communication integrity monitoring, and on-orbit firmware reconfiguration protocols.

The application scheduling and oversight module dynamically instantiates payload applications upon ground-commanded directives by spawning child processes with unique process identifiers, continuously monitors application runtime status and memory utilization through cyclic diagnostic checks, embeds resultant telemetry into engineering status packets, and enforces strict access constraints against frequent eMMC storage operations to mitigate radiation-induced wear and single-event functional interrupts. Figure 9 is the Linux program running interface of SPHMU.

Figure 9
Desktop screen with an image of a modern oval glass building and a classical building reflected in water. The sky is blue with scattered clouds. The Loongnix logo is in the bottom right corner.

Figure 9. Linux program running interface of SPHMU.

4 Experiment of SPHMU

As a standardized aerospace device, the SPHMU has undergone comprehensive acceptance-level environmental testing including Environmental Stress Screening (ESS) thermal cycling, functional thermal cycling, mechanical assessments comprising random and sinusoidal vibration profiles, thermal vacuum validation, electromagnetic compatibility (EMC) verification, residual magnetic field measurement, and electrostatic discharge (ESD) immunity trials, demonstrating robust performance stability across all parameters, memory playback data integrity, and consistent software execution throughout the validation campaign. EMC testing equivalently complies with IEC 61000–4 compliance requirements. The thermal vacuum test standards are shown in Table 3.

Table 3
www.frontiersin.org

Table 3. Conditions of thermal vacuum test.

During environmental testing, the operational temperature range of the SPHMU, module spans from −15 °C to +80 °C, compliant with the thermal control requirements specified for satellite systems.

During the environmental testing phase of the SPHMU, the operational execution of the High-Performance Computing (HPC) unit’s software is illustrated in Figure 10, which is orbital data reception and computational processing from Channel 0 of the MIL-STD-1553B bus, demonstrating sustained deterministic functionality and fault-free performance under simulated mission profiles, with zero deviations in computational throughput, memory integrity, or task completion deadlines observed throughout thermal, mechanical, and electromagnetic stress profiles equivalent to worst-case orbital conditions. Sphmu in thermal vacuum test is shown in Figure 11.

Figure 10
Terminal window displaying repeated commands. Messages such as

Figure 10. Operational execution of HPC Unit’s sftware.

Figure 11
Two black electronic components with multiple cables connected to them, including yellow, gray, and purple wires. The components have several screws and connectors visible, and there are red and white wires on the surface nearby.

Figure 11. Sphmu in thermal vacuum test.

We conducted a detailed comparison of OBC performance both domestically and internationally, and the comparison results are shown in Table 4. Based on the comparison results in the manual, we consider using LOONGSON XX2000 CPU as our HPC main chip. Because it can be seen from Table 4 that XX2000 has a higher frequency and can perform faster operations, the results of which are shown in Table 5. XX2000 also has a larger DDR4 SDRAM runtime capacity, which can handle more tasks.

Table 4
www.frontiersin.org

Table 4. Comparison results of OBCs (AAC Clyde Space, 2025).

Table 5
www.frontiersin.org

Table 5. The actual high-performance computing test of SPHMU.

5 Conclusion

This paper introduces a novel architecture for spaceborne intelligent computing and intelligent data processing, leveraging synergistic operation between a High-Reliability Computer (HRC) and a High-Performance Computer (HPC) to simultaneously achieve robust payload management with deterministic data interchange and advanced computational capabilities for on-orbit intelligence in space applications. Within this framework, the HRC exclusively manages all spacecraft bus interactions, including telecommand/telemetry operations for diverse payloads and spatiotemporal information synchronization, while the HPC executes specialized application software (Apps) for processing payload-derived datasets—particularly computation-intensive tasks such as real-time attitude determination, orbital trajectory propagation, and navigation solutions demanding significant processing throughput.

The Space Payload Health Management Unit (SPHMU) designed for specified mission profiles embodies this architecture through a highly integrated and radiation-hardened platform, integrating an HPC alongside a retained HRC to enable space-grade high-performance computing and data processing; consequently, it preserves fundamental payload management and data handling functionalities inherent in conventional units while augmenting capabilities with advanced computational modules, including real-time precision orbit determination algorithms, autonomous maneuver software for elliptical trajectories, and intelligent on-orbit analytics with anomaly detection systems—all deployable as containerized applications via dedicated hardware supporting software-defined installation paradigms.

This innovative architecture for spaceborne high-performance computing and intelligent data processing establishes a technological foundation extensible to future space stations, low-Earth-orbit observation missions, medium/high-Earth-orbit exploration platforms, and deep-space probes; furthermore, its Linux-based desktop-like operational interface significantly enhances astronaut usability during scientific operations, offering intuitive control for complex computational tasks while exhibiting exceptional adaptability across diverse space mission profiles through its open architecture and containerized software ecosystem.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

WZ: Project administration, Visualization, Formal Analysis, Writing – original draft, Data curation, Conceptualization, Investigation, Methodology. JR: Visualization, Data curation, Validation, Writing – review and editing, Conceptualization. MM: Software, Writing – original draft. JZ: Supervision, Methodology, Writing – review and editing. SL: Supervision, Writing – original draft. WJ: Writing – original draft, Validation. ZD: Investigation, Conceptualization, Writing – original draft. BH: Investigation, Writing – original draft, Methodology, Data curation. GX: Investigation, Writing – review and editing, Conceptualization, Formal Analysis, Project administration. GY: Supervision, Visualization, Writing – original draft. YZ: Investigation, Writing – review and editing. JA: Conceptualization, Investigation, Writing – review and editing.

Funding

The author(s) declare that no financial support was received for the research and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

AAC Clyde Space (2025). Command and Data Handling. Uppsala, Sweden: AAC Clyde Space. Available online at: https://www.aac-clyde.space/what-we-do/space-products-components/command-data-handling (Accessed on August 2025).

Google Scholar

Abdu, T. S., Kisseleff, S., Lagunas, E., Grotz, J., Chatzinotas, S., and Ottersten, B. (2023). Demand-aware onboard payload processor management for high-throughput NGSO satellite systems. IEEE Trans. Aerosp. Electron. Syst. 59 (5), 1–18. doi:10.1109/TAES.2023.3245044

CrossRef Full Text | Google Scholar

Ben Yahia, O., Garroussi, Z., Bélanger, O., Sansò, B., Frigon, J.-F., Martel, S., et al. (2025). Evolution of high-throughput satellite systems: a vision of programmable regenerative payload. IEEE Commun. Surv. and Tutorials 27 (3), 1565–1597. doi:10.1109/comst.2024.3450292

CrossRef Full Text | Google Scholar

Bhandari, S., Vu, T. X., and Chatzinotas, S. (2024). User-centric flexible resource management framework for LEO satellites with fully regenerative payload. IEEE J. Sel. Areas Commun. 42 (5), 1246–1261. doi:10.1109/JSAC.2024.3365883

CrossRef Full Text | Google Scholar

Bhargav Reddy, M., and Vijayreddy, D. (2025). The recent applications of remote sensing in agriculture-a review. Journal of Agriculture Biotechnology & Applied Sciences 1 (2), 28–35. doi:10.63143/jabaas.2120233

CrossRef Full Text | Google Scholar

Del Castillo, M. O., McRobbie, J., Woods, A., Mearns, R., Morgan, J., Trenti, M., et al. (2025). The SpIRIT payload management system: a paradigm for smallsats multi payload integration. Acta Astronaut. 236, 359–370. doi:10.1016/j.actaastro.2025.06.050

CrossRef Full Text | Google Scholar

Hui, M., Zhai, S., Wang, D., Hui, T., Wang, W., Du, P., et al. (2025). A review of LEO-satellite communication payloads for integrated communication, navigation, and remote sensing: opportunities, challenges, future directions. IEEE Internet Things J. 12 (12), 18954–18992. doi:10.1109/JIOT.2025.3553942

CrossRef Full Text | Google Scholar

Martins, W. A., Lagunas, E., Skatchkovsky, N., Ortiz, F., Eappen, G., Simeone, O., et al. (2024). Satellite adaptive onboard beamforming using neuromorphic processors. In: 2024 IEEE 35th international symposium on personal, indoor and mobile radio communications (PIMRC); September 2024 02–05; Valencia, Spain: IEEE, 1–6.

CrossRef Full Text | Google Scholar

Morton, Y. J., van Diggelen, F., Spilker, J. J., Parkinson, B. W., Lo, S., and Gao, G. (2021). Position, navigation, and timing technologies in the 21st century: integrated satellite navigation. In: Sensor systems, and civil applications. Hoboken, NJ, USA: Wiley.

Google Scholar

Ramírez, T., Mosquera, C., and Alagha, N. (2022). Flexible user mapping for radio resource assignment in advanced satellite payloads. IEEE Trans. Broadcast. 68 (3), 723–739. doi:10.1109/TBC.2022.3178018

CrossRef Full Text | Google Scholar

Zhao, W., Shi, Y., Zhou, S., Liu, X., Zhu, Y., Wang, W., et al. (2022). A novel space high speed data transmission system of 2.4Gbps. In: 2022 IEEE 5th international conference on electronics technology (ICET); May 2022 13–16; Chengdu, China: IEEE, 1055–1059. doi:10.1109/ICET55676.2022.9824727

CrossRef Full Text | Google Scholar

Zhou, L., and An, J. (2013). Design of a payload data handling system for satellites. In: 2013 third international conference on instrumentation, measurement, computer, communication and control; September 2013 21-23; Shenyang, China: IEEE, 1217–1220.

CrossRef Full Text | Google Scholar

Keywords: intelligent computing, intelligent data processing, attitude and orbit, high-reliability computer, high-performance computer

Citation: Zhao W, Rao J, Ma M, Zhang J, Li S, Jia W, Dong Z, Hou B, Xue G, Yin G, Zhu Y and An J (2025) A novel space intelligent computing and data processing architecture--the spacecraft payload health management unit (SPHMU). Front. Astron. Space Sci. 12:1657487. doi: 10.3389/fspas.2025.1657487

Received: 01 July 2025; Accepted: 29 August 2025;
Published: 19 September 2025.

Edited by:

Anshu Kumari, Physical Research Laboratory, India

Reviewed by:

Harsha Avinash Tanti, Indian Institute of Technology Indore, India
Vaibhav Rathore, Physical Research Laboratory, India

Copyright © 2025 Zhao, Rao, Ma, Zhang, Li, Jia, Dong, Hou, Xue, Yin, Zhu and An. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jianing Rao, cmFvamlhbmluZ0Buc3NjLmFjLmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.