Phison’s Pascari X-Series lineup is engineered to address the diverse storage demands of enterprise environments, delivering tailored solutions optimized for both read-intensive and write-intensive workloads. At the core of the lineup is the X200P, a high-capacity model that supports storage capacities of up to 30.72TB with a 1 DWPD (Drive Writes Per Day) rating. Leveraging Gen5 PCIe technology and TLC NAND, the X200P is available in U.3, U.2, and E3.S form factors—offering flexibility to integrate seamlessly into various enterprise infrastructure setups.
Designed for versatility, the X200P excels across a wide range of enterprise use cases, including large-scale content delivery networks, AI inference workloads, and cold data archiving—where high capacity and reliable read performance are paramount. Complementing the X200P is Phison’s X200E series, a high-endurance lineup optimized specifically for write-intensive scenarios. Boasting up to 3 DWPD and capacity options spanning from 1.6TB to 25.6TB, the X200E is ideally suited for mission-critical applications such as transactional databases, real-time data analytics, and high-volume log processing.
Focusing on the Phison Pascari X200P for this review, Phison provided the 7.68TB U.2 model for testing. To thoroughly evaluate its performance under real-world enterprise pressure, we subjected the drive to our complete suite of rigorous enterprise benchmarks, assessing key metrics such as throughput, latency, and stability across varied workload profiles.
Phison Pascari X200P Series Specifications
| Specifications Phison Pascari X200P Series | 1.92TB | 3.84TB | 7.68TB | 15.36TB | 30.72TB |
|---|---|---|---|---|---|
| Form Factor | U.2 | ||||
| Interface | PCIe 5.0 x4, 2×2 | ||||
| NVMe | 2.0 | ||||
| NAND Flash | 3D TLC | ||||
| Sequential Read (MB/s) | 14,800 | 14,800 | 14,800 | 14,800 | 14,000 (Est.) |
| Sequential Write (MB/s) | 4,300 | 8,600 | 8,700 | 8,350 | 7,500 (Est.) |
| 4K Random Read (IOPS) | 2,400K | 3,000K | 3,000K | 3,000K | 2,300K (Est.) |
| 4K Random Write (IOPS) | 170K | 380K | 500K | 500K | 283K (Est.) |
| Read Latency (μs) | 60 | ||||
| Write Latency (μs) | 10 | ||||
| Power – Active (W) | <25 | ||||
| Power – Idle (W) | 5 | ||||
| DWPD(7) | 1 | ||||
| UBER | <1 sector per 1018 bits read | ||||
| MTBF (million hours) | 2.5 | ||||
| Limited Warranty (years) | 5 | ||||
| Operating Temp. (°C) | 0 to 70 | ||||
| Non-Operating Temp. (°C) | -40 to 85 | ||||
| Dimensions (mm) | 100.10 (L) x 69.85 (W) x 15.00 (H) | ||||
| Weight (g) | 188 | 199 | 201 | 168 | <250 |
Build and Design: Phison Pascari X200P 7.68TB
Our test unit is the 7.68TB U.2 2.5″ variant of the Phison Pascari X200P, engineered to deliver high-performance storage for enterprise applications. It features a PCIe 5.0 interface, fully compliant with the NVMe 2.0 specification, and is built around high-endurance 3D TLC NAND—with the full X200P lineup supporting capacities up to 30.72TB to accommodate diverse enterprise storage needs.
Physical Design & Form Factor
Physically, the X200P adheres to the standard 2.5″ U.2 form factor, with precise dimensions of 100.10mm (length) × 69.85mm (width) × 15.00mm (height) and a weight of 201 grams. The drive is encased in a sleek black aluminum housing with integrated passive cooling, a design optimized to efficiently manage thermal output during sustained, high-intensity workloads. For added flexibility in dense storage environments, the X200P also supports E3.S configurations, making it adaptable to various enterprise infrastructure setups.
Performance Specifications
From a performance standpoint, the X200P boasts impressive rated metrics: up to 14,800MB/s sequential read, 8,700MB/s sequential write, 3 million IOPS random read, and 500,000 IOPS random write. It also delivers efficient power consumption, with an active power draw below 25W and an idle power consumption of just 5W—making it a cost-effective choice for sustained high-throughput enterprise operations.
The drive carries a 1 DWPD (Drive Writes Per Day) endurance rating, a 2.5 million-hour MTBF (Mean Time Between Failures), and a 5-year limited warranty. Designed for 24/7 enterprise operation, it operates reliably within a temperature range of 0°C to 70°C, ensuring consistency in demanding data center environments.
Enterprise-Grade Features
Phison equips the X200P with a comprehensive suite of enterprise-class data protection and manageability features to safeguard critical data and simplify deployment:
- Power Loss Protection (PLP) to prevent data loss during unexpected power interruptions
- ISE (Instant Secure Erase) and TCG Opal 2.0 support for secure data sanitization
- AES-XTS 256-bit Encryption for end-to-end data security
- End-to-End Data Path Protection and Metadata Protection to ensure data integrity
- SECDED (Single Error Correction Double Error Detection) for enhanced data reliability
- Sanitize operations for compliant data disposal
- NVMe-MI (Management Interface) and SMBus compatibility for streamlined device management
- Support for up to 128 namespaces to optimize storage allocation
Collectively, the Pascari X200P lineup combines robust industrial-grade build quality, cutting-edge performance, and enterprise-grade reliability—positioning it as a strong contender for demanding storage environments such as cloud infrastructure, AI/ML workloads, and virtualized data centers.
Performance Testing
Drive Testing Platform
We conducted all benchmarking for this review using a Dell PowerEdge R760 running Ubuntu 22.04.02 LTS, paired with a Serial Cables Gen5 JBOF (Just a Bunch of Flash) for broad compatibility with U.2, E1.S, E3.S, and M.2 SSDs. The complete system configuration is outlined below:
- 2 x Intel Xeon Gold 6430 (32-Core, 2.1GHz) processors
- 16 x 64GB DDR5-4400 RAM modules
- 480GB Dell BOSS SSD for boot and system operations
- Serial Cables Gen5 JBOF for SSD testing
Drives Compared
To provide a fair and relevant comparison, we tested the Pascari X200P 7.68TB against a group of 7.68TB PCIe Gen5 NVMe SSDs with TLC NAND flash, all targeted at enterprise high-performance environments. The comparison set includes:
- Phison Pascari X200P 7.68TB
- Micron 9550 7.68TB
- SanDisk SN861 7.68TB
- Solidigm PS1010 7.68TB
- Kingston DC3000ME 7.68TB
Testing was conducted using a mix of real-world and synthetic benchmarks—including CDN workload simulations, FIO (Flexible I/O Tester), and GDSIO (GPU Direct Storage I/O)—to evaluate performance across sustained throughput, latency, mixed I/O patterns, and GPU-accelerated workloads. By standardizing capacity, interface, and NAND type, this evaluation delivers a clear comparison of how the Pascari X200P performs against its peers under demanding enterprise conditions.
CDN Performance Testing
To simulate realistic mixed-content CDN (Content Delivery Network) workloads, we subjected each SSD to a multi-phase benchmarking sequence designed to replicate the I/O patterns of content-heavy edge servers. This sequence included a range of block sizes (both large and small), distributed across random and sequential operations, with varying concurrency levels to mimic real-world edge server demands.
Preconditioning & Saturation
Before initiating main performance tests, each SSD underwent a full device fill with a 100% sequential write pass using 1MB blocks, utilizing synchronous I/O and a queue depth of 4 (allowing four simultaneous jobs). This step ensured the drive entered a steady-state condition representative of real-world usage. Following the sequential fill, a three-hour randomized write saturation stage was executed, using a weighted block size distribution heavily favoring 128K transfers (98.51%), with minor contributions from sub-128K blocks down to 8K—emulating the fragmented write patterns common in distributed cache environments.
Main Testing Suite
The main testing focused on scaled random read and write operations to measure each drive’s behavior under variable queue depths and job concurrency. Each test ran for 5 minutes (300 seconds), followed by a 3-minute idle period to allow internal recovery mechanisms to stabilize performance metrics. Two key test profiles were used:
- A fixed block size distribution favoring 128K (98.51%), with the remaining 1.49% composed of smaller transfer sizes (64K to 8K). Tests were run across 1, 2, and 4 concurrent jobs, with queue depths of 1, 2, 4, 8, 16, and 32—profiling throughput scalability and latency under typical edge-write conditions.
- A heavily mixed block size profile mimicking CDN content retrieval, featuring a dominant 128K component (83.21%) and a long tail of over 30 smaller block sizes (4K to 124K), each with fractional frequency representation. This distribution reflects diverse request patterns encountered during video segment fetching, thumbnail access, and metadata lookups, and was tested across the same matrix of job counts and queue depths.
This combination of preconditioning, saturation, and mixed-size randomized access tests reveals how SSDs handle sustained CDN-like environments, emphasizing responsiveness and efficiency in bandwidth-heavy, highly parallelized scenarios.
CDN Workload Results
CDN Workload Read 1 (Single Job)
In this test simulating light content delivery traffic, the Pascari X200P started at the back of the pack at QD1 (765MB/s) and QD2 (1,403MB/s). As queue depth increased, the drive scaled efficiently, moving to the middle of the field through QD8 and QD16. By QD32, it reached 13,516.8MB/s, finishing third overall—behind the Kingston DC3000ME and Micron 9550, but outperforming the SanDisk SN861 and Solidigm PS1010 at the top end.
CDN Workload Read 2 (Two Jobs)
CDN Workload Read 4 (Four Jobs)
With four concurrent jobs, the Pascari X200P showed strong scaling through queue depths. It trailed all drives at QD1 (2,982MB/s) but steadily gained ground through QD2 and QD4. By QD8, it moved to the front of the pack and maintained this lead through QD16 and QD32, finishing first overall at QD32 with 15,257.6MB/s—ahead of the Micron 9550 and Kingston DC3000ME.
CDN Workload Write 1 (Single Job):
In the single-job CDN write test, the Pascari X200P trailed the pack, achieving a maximum speed of 1,885MB/s at QD1 and scaling gradually to 5,913MB/s at QD32—finishing fourth overall. The SanDisk SN861 and Micron 9550 led the group, followed by the Kingston DC3000ME, while the X200P maintained consistent scaling but less aggressive write performance in this low-threaded scenario.
CDN Workload Write 2 (Two Jobs):
With two concurrent jobs, the Pascari X200P finished fourth overall. It achieved 2,762MB/s at QD1, scaled through QD16, but exhibited some performance tapering by QD32 (reaching 4,585MB/s). The Micron 9550 and SanDisk SN861 led, followed by the Kingston DC3000ME, with the X200P maintaining stable performance through mid-queue depths but trailing the leaders.
CDN Workload Write 4 (Four Jobs):
With four concurrent jobs, the Pascari X200P held mid-pack performance through most of the test. It achieved 2,845MB/s at QD1, remained competitive with the Kingston DC3000ME and Solidigm PS1010 through mid-queue depths, but tailed off slightly at QD32 (3,613MB/s), finishing fifth overall. The Micron 9550 and SanDisk SN861 led the field, with the Kingston DC3000ME in third. The X200P delivered consistent write scaling under moderate loads but showed limits at deeper queue depths in this four-threaded workload.
DLIO Checkpointing Benchmark
To evaluate the X200P’s real-world performance in AI training environments, we used the Data and Learning Input/Output (DLIO) benchmark tool—developed by Argonne National Laboratory specifically to test I/O patterns in deep learning workloads. DLIO provides insights into how storage systems handle critical AI tasks such as checkpointing, data ingestion, and model training. We used DLIO benchmark version 2.0 (August 13, 2024, release), with results illustrating how the X200P and competing drives handle 36 checkpoints—essential for saving model states periodically and preventing progress loss during interruptions.
Test Configuration
To reflect real-world AI scenarios, our testing was based on the LLAMA 3.1 405B model architecture. We implemented checkpointing using torch.save() to capture model parameters, optimizer states, and layer states, simulating an eight-GPU system with a hybrid parallelism strategy (4-way tensor parallelism and 2-way pipeline parallel processing). This configuration resulted in checkpoint sizes of 1,636GB—representative of modern large language model (LLM) training requirements.
DLIO Results
The Pascari X200P demonstrated strong initial responsiveness but exhibited increased checkpoint times as the workload intensified. In early checkpoints (1–4), it stayed competitive with the pack, averaging 467 seconds—keeping pace with drives like the Solidigm PS1010 and Micron 9550.
By the midpoint (Checkpoints 5–9), however, the X200P’s performance diverged. Checkpoint times rose sharply, peaking at 689.68 seconds by Checkpoint 12 (the highest in the group). Across the final three checkpoints, it averaged 672 seconds—roughly 19.3% slower than the next-slowest drive (Kingston DC3000ME) and 23% slower than the group average.
When viewed by pass averages, the X200P showed a clear trajectory of performance degradation: it averaged 467.93 seconds in Pass 1 (slightly behind the field), 662.04 seconds in Pass 2 (14.5% slower than the next-slowest drive and 17.4% slower than the group average), and 674.48 seconds in Pass 3 (remaining the slowest drive, 18.9% slower than the average of the other four drives, which was approximately 567 seconds).
FIO Performance Benchmark
To measure storage performance across common industry metrics, we used FIO (Flexible I/O Tester), with a standardized testing process for all drives: two full drive fills with a sequential write workload for preconditioning, followed by steady-state performance measurement. A new preconditioning fill was run for each new transfer size to ensure accurate results. We focused on the following FIO benchmarks: 128K Sequential, 64K Random, 16K Random, 4K Random, and 128K Sequential Precondition.
FIO Test Results
128K Sequential Precondition (IODepth 256 / NumJobs 1):
128K Sequential Precondition Latency (IODepth 256 / NumJobs 1):
128K Sequential Write (IODepth 16 / NumJobs 1):
128K Sequential Write Latency (IODepth 16 / NumJobs 1):
128K Sequential Read (IODepth 64 / NumJobs 1):
128K Sequential Read Latency (IODepth 64 / NumJobs 1):
64K Random Write:
64K Random Write Latency:
64K Random Read:
64K Random Read Latency:
16K Random Write:
16K Random Write Latency:
16K Random Read:
16K Random Read Latency:
4K Random Write:
The X200P delivered steady results starting at 1/1 (91.9K IOPS), generally sitting in the middle to lower end of the pack across most queue depths and thread combinations. Its peak throughput reached 1.64 million IOPS at 32/16—competitive but trailing top results from SanDisk and Micron in some scenarios.
4K Random Write Latency:
4K Random Read:
4K Random Read Latency:
GPU Direct Storage (GDS) Testing
We also conducted Magnum IO GPU Direct Storage (GDS) testing—a feature developed by NVIDIA that allows GPUs to bypass the CPU when accessing data on NVMe drives or other high-speed storage devices. By enabling direct communication between the GPU and storage via the PCIe bus, GDS eliminates CPU bottlenecks, reduces latency, and improves data throughput—critical for data-intensive AI workloads.
How GPU Direct Storage Works
Traditionally, GPU data processing requires data to travel from NVMe drives through the CPU and system memory before reaching the GPU—introducing latency and consuming valuable CPU resources. GDS eliminates this inefficiency by creating a direct path between the GPU and storage, reducing data movement overhead and enabling faster, more efficient transfers.
This is particularly beneficial for AI/ML workloads (e.g., deep learning), which require processing terabytes of data—any transfer delay can lead to underutilized GPUs and longer training times. GDS also excels in streaming large datasets (video processing, NLP, real-time inference) by freeing up CPU resources for other tasks, enhancing overall system performance.
GDSIO Test Results
GDSIO Read Throughput: At 16K block size, throughput started at 0.56 GiB/s (QD1) and increased to 1.80 GiB/s at QD128—modest but steady scaling, reflecting acceptable performance for small transfer sizes. At 128K block size, performance improved more noticeably: 2.39 GiB/s (QD1) to 5.10 GiB/s (QD128), demonstrating better scaling efficiency. At 1M block size, throughput started at 3.63 GiB/s and scaled to 6.15 GiB/s at QD128—delivering the highest absolute read bandwidth, making it well-suited for large sequential transfers.
GDSIO Read Latency: Results highlighted a clear relationship between block size, thread count, and latency. At 16K block size (1 thread), latency was 0.026ms, spiking to 1.076ms at 128 threads. At 128K block size, latency rose from 0.050ms (1 thread) to 3.056ms (128 threads). At 1M block size, latency started at 0.268ms (1 thread) and peaked at 20.324ms under maximum parallelism.
GDSIO Write Throughput (16K Block Size): Throughput started at 0.58 GiB/s (25.17µs latency) at QD1 and rose to 1.22 GiB/s (1.59ms latency) at QD128—modest bandwidth gain but steep latency increase, suggesting early saturation at this small I/O size.
GDSIO Write Throughput (128K Block Size): Performance scaled better, starting at 2.63 GiB/s (45.55µs) and increasing to 4.94 GiB/s (3.16ms) at QD128—healthy throughput gain but sharp latency scaling, indicating growing overhead at high queue depths.
GDSIO Write Throughput (1M Block Size): The drive started strong at 4.52 GiB/s (215µs) and peaked at 5.02 GiB/s (24.9ms) at QD128—minimal throughput gain compared to 128K, with the highest latency of all tests, signaling limited efficiency gains from larger transfers beyond 128K at deep queues.
GDSIO Write Latency: Latency increased consistently with block size and thread count. At 16K block size (1 thread), latency was 0.025ms, climbing to 1.595ms at 128 threads. At 128K block size, latency rose from 0.046ms to 3.159ms (128 threads). At 1M block size, latency started at 0.215ms and reached 24.917ms at maximum thread depth. Despite this expected rise, the X200P led the group at higher block sizes and thread counts, maintaining the lowest latency under heavy parallel write workloads.
Conclusion
The Phison Pascari X200P 7.68TB SSD is an enterprise-grade storage solution featuring TLC NAND and optimized for PCIe Gen5 performance, catering to general-purpose and content-heavy workloads. It is engineered for environments where high throughput, strong scalability, and deployment flexibility take priority over hyperscale-specific tuning. With support for U.2, U.3, and E3.S form factors, plus enterprise-grade features like power-loss protection, AES-XTS 256-bit encryption, and NVMe-MI management, the X200P provides a solid foundation for enterprise storage infrastructure.
In terms of performance, the X200P excels in sequential and read-intensive scenarios—consistently ranking near the top in 128K and 64K tests and scaling effectively under CDN workloads. FIO testing confirms its strength in sequential reads and competitive performance across random read workloads. While it trails top-tier drives (e.g., Micron, SanDisk) in write-intensive and highly concurrent conditions, its predictable, efficient write behavior makes it well-suited for a wide range of mid-tier enterprise deployments.
GDSIO testing further highlights the drive’s strengths in throughput-focused applications: it maintains excellent latency at smaller block sizes and leads under heavy parallel access with large block transfers. Although latency increases at deeper queue depths, Phison’s tuning ensures the drive remains stable and responsive under sustained pressure.
Overall, the Pascari X200P is a well-rounded enterprise SSD with strong performance and a feature set tailored for real-world workloads. It will be interesting to see if Phison can transition from a controller-first company to one offering a deep set of integrated drive solutions—and the X200P appears to be a promising step in that direction.
Designed for versatility, the X200P excels across a wide range of enterprise use cases, including large-scale content delivery networks, AI inference workloads, and cold data archiving—where high capacity and reliable read performance are paramount. Complementing the X200P is Phison’s X200E series, a high-endurance lineup optimized specifically for write-intensive scenarios. Boasting up to 3 DWPD and capacity options spanning from 1.6TB to 25.6TB, the X200E is ideally suited for mission-critical applications such as transactional databases, real-time data analytics, and high-volume log processing.
Focusing on the Phison Pascari X200P for this review, Phison provided the 7.68TB U.2 model for testing. To thoroughly evaluate its performance under real-world enterprise pressure, we subjected the drive to our complete suite of rigorous enterprise benchmarks, assessing key metrics such as throughput, latency, and stability across varied workload profiles.
Beijing Qianxing Jietong Technology Co., Ltd.
Sandy Yang/Global Strategy Director
WhatsApp / WeChat: +86 13426366826
Email: yangyd@qianxingdata.com
Website: www.qianxingdata.com/www.storagesserver.com
Sandy Yang/Global Strategy Director
WhatsApp / WeChat: +86 13426366826
Email: yangyd@qianxingdata.com
Website: www.qianxingdata.com/www.storagesserver.com
Business Focus:
ICT Product Distribution/System Integration & Services/Infrastructure Solutions
With 20+ years of IT distribution experience, we partner with leading global brands to deliver reliable products and professional services.
“Using Technology to Build an Intelligent World”Your Trusted ICT Product Service Provider!



