Efficient workload placement in AIX is critical for maximizing performance for user workload on multi-srad or multi-node systems. IBM’s Active System Optimizer (ASO) introduces strategies like cache affinity and memory affinity to improve job locality and system throughput. This blog explores the new feature in these strategies to suit them better for customer workloads.
AIX's Active System Optimizer (ASO) previously treated each shared memory segment (SHM) and each process as separate jobs. This isolated view limited placement strategies from understanding inter-process communication via shared memory.
To solve this, a SHM clustering mechanism has been introduced.
How It Works
Example
Imagine 8 processes interacting across 4 SHM segments. If P2 accesses both SHM1 and SHM2, these jobs are now grouped and placed together to reduce cross-node traffic.
Benefits
Clustering based on shared memory and affinitizing workloads to a common domain significantly enhances data locality, leading to tangible performance improvements. Performance engineering tests have shown:
These gains highlight the effectiveness of ASO’s clustering and affinity strategies in real-world, performance-sensitive environments.
Usage
This optimization is enabled by default. ASO automatically identifies a domain large enough to accommodate all threads within a cluster, ensuring balanced placement without manual intervention.
By design, ASO limits this clustering logic to processes that share memory segments, ensuring it focuses on workloads that benefit most from locality-aware scheduling.
Final Thoughts
This enhancement represents a meaningful advancement in how AIX, through ASO, intelligently places workloads across system resources:
Together, these improvements deliver a more efficient and scalable AIX environment—especially for workloads where memory access patterns directly influence performance.