I’m currently a PhD student in the NCSU Systems Lab department of Computer Science at NC State University advised by Prof. Frank Mueller. Prior to that, I earned my Master degree at Wuhan National Laboratory for Optoelectronics (WNLO) at Huazhong University of Science and Technology (HUST) under the supervise of Prof. Ke Zhou. And I receive my Bacholar degree at Hainan University.

My Research interest mainly focused on the System Architecture, High Performance Computer and Storage techniques (e.g. Persistent Memory and CXL memory).

🌎 Motto

  • 读万卷书 行万里路
  • Gain wisdom through reading and broaden horizons through travel.

🔥 News

  • 2025.12:  🏆 Our work FitCache has been accepted to IPDPS 2026. See you in New Orleans!
  • 2025.10:  🎉 Passed the PhD Written Preliminary Exam. Thanks to everyone for the support!
  • 2024.10:  🏆 Poster accepted at SC 2024. Looking forward to presenting in Atlanta.
  • 2023.08:  🎓 Started my PhD journey at NC State University!

📖 Educations

  • 08/2023 – Present, North Carolina State University, Raleigh, USA
    Ph.D. candidate in Computer Science (Advisor: Prof. Frank Mueller)

  • 09/2019 – 06/2023, Huazhong University of Science and Technology, Wuhan, China
    M.Sc. in Computer System Architecture
    • Thesis: Research on Performance Anomaly Detection and Root Cause Analysis for Storage Systems in Data Centers
    • Advisor: Prof. Ke Zhou
    • Focus: Data Center, Storage System, Anomaly Detection, Root Cause Analysis, Machine Learning
  • 09/2015 – 06/2019, Hainan University, Haikou, China
    B.Eng. in Computer Science and Technology

📝 Publications

* These authors contributed equally to this work.

  1. Guangxing Hu, Awais Khan, Christopher zimmer, Michael Brim, Frank Mueller.
    “FitCache: A Transparent Drop-In Framework for Multi-Tier Caching to Accelerate Distributed Deep Learning Workloads.”
    IPDPS 2026, New Orleans, USA.

  2. Guangxing Hu, Awais Khan, Frank Mueller.
    “A Zero-Copy Storage with Metadata-Driven File Management Using Persistent Memory.”
    SC Poster, 2024, Atlanta, USA.

  3. Yu Liu, Yunchuan Guan, Tianming Jiang, Ke Zhou, Hua Wang, Guangxing Hu, Ji Zhang, Wei Fang, Zhuo Cheng, Ping Huang.
    “SPAE: Lifelong Disk Failure Prediction via End-to-End GAN-based Anomaly Detection with Ensemble Update.”
    Future Generation Computer Systems, 2023.

  4. Yanzhao Xie*, Guangxing Hu*, Yu Liu, Zhiqiu Lin, Ke Zhou, Yuhong Zhao.
    “How Visual Chirality Affects the Performance of Image Hashing.”
    Neural Computing and Applications, 2022.

  5. Yangtao Wang, Yanzhao Xie, Lisheng Fan, Guangxing Hu.
    “STMG: Swin transformer for multi-label image recognition with graph convolution network.”
    Neural Computing and Applications, 2021.

🎖 Honors and Awards

  • National Grand Prize of Massive Storage Algorithm Competition 2023
  • Academic Scholarship of Huazhong University of Science and Technology 2019–2022
  • Grand Prize, Group Programming Ladder Tournament, China Collegiate Computing Contest 2018
  • The Principal Scholarships of Hainan University (Top 2%) 2018

💻 Internships

  • 2025.06 - 2025.08, Oak Ridge National Laboratory (ORNL), Oak Ridge, TN, USA.
    Designed FitCache, a multi-tier caching system for accelerating deep learning training. Implemented coordinated caching across memory and storage tiers to reduce data access latency, and evaluated the system at world-leading supercomputer Frontier. This work was presented at IPDPS 2026.

  • 2024.06 - 2024.08, Oak Ridge National Laboratory (ORNL), Oak Ridge, TN, USA.
    Extended the HVAC deep learning caching framework with a Persistent Memory–based cache tier, enabling zero-copy access via DAX. Deployed and debugged the system on Frontier and ARC clusters, addressing MPI and runtime issues across environments. Achieved preliminary I/O performance improvements and presented the work in an SC 2024 poster.

💬 Talks

  • 2025.08, Leveraging Memory Hierarchies for Scalable Caching in Deep Learning I/O on HPC Systems, Oak Ridge, TN, USA.

🔖 Projects

DPUBridge overview

DPUBridge

In progress · Host memory to DPU relay over GVMI and RDMA

A layered bridge that enables a DPU to read local host memory through GVMI aliasing and mirror payloads to a remote DPU using RDMA. The design separates control plane metadata exchange from data plane transfers to keep modules testable and clean.

  • Control plane: Host↔DPU exchange for GVMI access metadata, plus DPU↔DPU exchange for RDMA QP connection setup
  • Data plane: GVMI backed reads from host buffers, then RDMA writes into a remote mirror region
  • Key techniques: GVMI aliasing for host memory access, RDMA verbs for transport, explicit control and data plane separation
  • Design focus: clean boundaries between memory access, networking, and relay scheduling to keep bring up and debugging tractable
  • Extensible: pluggable scheduling policies and offload data to the Parallel File Systems(e.g., BeeGFS)

Status: bringing up end to end path (GVMI read + RDMA write) and stabilizing the control plane protocol.

FitCache I/O flow

FitCache

IPDPS 2026 · Multi tier caching for distributed deep learning I/O

Transparent drop in framework that intercepts POSIX reads and serves training data from the fastest tier among memory, NVMe, and PFS via concurrent fetch.

  • LD_PRELOAD client intercepts open read close and forwards requests to a cache server
  • Hierarchical cache across DRAM or PMem and NVMe above Lustre or BeeGFS
  • Concurrent multi tier fetch returns the earliest responder, tolerating tier variability
  • Scaled evaluation on Frontier up to 2048 GPUs and smaller clusters

Key results: up to 40% training time reduction and up to 71.6% per batch I/O latency reduction (188.3ms to 53.4ms).

Visual chirality effect on hash codes

Visual Chirality in Image Hashing

Neural Computing and Applications 2023 · Data augmentation analysis for deep hashing

This project studies visual chirality, where flipping an image can change its semantics and shift the hash code distribution, which can hurt retrieval when flip augmentation is used.

  • Designed a ResNet 50 based classifier to identify images with chiral cues by comparing predictions of original and flipped images
  • Constructed chiral datasets with controlled proportions of chiral images on VOC2007, MS-COCO, and NUS-WIDE
  • Evaluated three representative hashing pipelines: HashNet, DCH, and BYOL plus DCH under multiple hash code lengths
  • Provided visual explanations via activated region comparisons between original and flipped images

Key results: performance varies strongly with the proportion of chiral data, and the best performance consistently appears when chiral images account for about 15% to 25% or 75% to 85%.