Any benchmarks for non-pine64 pci to NVMe adapters?
#1
I have two things I would like to find benchmarks/information for. I apologize ahead of time if they are in another thread. I couldn't find them via google or the search function on this forum.

1. Does using a higher quality pci to NVMe adapter improve the performance? Ideally, the PCI lanes should each be able to contribute 400 MB/s for a total of 1.6 GB/s. The only benchmarks i've seen cap out at 540 MB/s write and 1.3 GB/s read. The SSD being tested in those is rated to be able to achieve much higher performance in both areas. I don't know if the difference is due to internal limitations of the processor or the PCI adapter itself. Hence my question.

2. I would also like to know if using a USB type C to ethernet adapter improves the network performance compared to using the built-in ethernet port.
#2
Samsung 970 Pro NVMe M.2 500GB with pine64 PCIe NVMe adapter.


Code:
rock64@rockpro64v2_0:/mnt$ sudo iozone -e -I -a -s 100M -r 4k -r 16k -r 512k -r 1024k -r 16384k -i 0 -i 1 -i 2
    Iozone: Performance Test of File I/O
            Version $Revision: 3.429 $
        Compiled for 64 bit mode.
        Build: linux

    Contributors:William Norcott, Don Capps, Isom Crawford, Kirby Collins
                 Al Slater, Scott Rhine, Mike Wisner, Ken Goss
                 Steve Landherr, Brad Smith, Mark Kelly, Dr. Alain CYR,
                 Randy Dunlap, Mark Montague, Dan Million, Gavin Brebner,
                 Jean-Marc Zucconi, Jeff Blomberg, Benny Halevy, Dave Boone,
                 Erik Habbinga, Kris Strecker, Walter Wong, Joshua Root,
                 Fabrice Bacchella, Zhenghua Xue, Qin Li, Darren Sawyer,
                 Vangel Bojaxhi, Ben England, Vikentsi Lapa.

    Run began: Sat Jul 28 12:08:50 2018

    Include fsync in write timing
    O_DIRECT feature enabled
    Auto Mode
    File size set to 102400 kB
    Record Size 4 kB
    Record Size 16 kB
    Record Size 512 kB
    Record Size 1024 kB
    Record Size 16384 kB
    Command line used: iozone -e -I -a -s 100M -r 4k -r 16k -r 512k -r 1024k -r 16384k -i 0 -i 1 -i 2
    Output is in kBytes/sec
    Time Resolution = 0.000001 seconds.
    Processor cache size set to 1024 kBytes.
    Processor cache line size set to 32 bytes.
    File stride size set to 17 * record size.
                                                              random    random     bkwd    record    stride                                    
              kB  reclen    write  rewrite    read    reread    read     write     read   rewrite      read   fwrite frewrite    fread  freread
          102400       4    83920   146526   171217   172733    56965   145921                                                          
          102400      16   271229   414900   454454   460018   193626   413496                                                          
          102400     512  1021580  1033256  1007794  1057973   990788  1075201                                                          
          102400    1024  1066333  1107758  1038792  1079089  1048932  1116344                                                          
          102400   16384   918513  1418530  1433672  1529740  1523500  1389826                                                          

iozone test complete.


with

Code:
rock64@rockpro64v2_0:/mnt$ uname -a
Linux rockpro64v2_0 4.18.0-rc5-1050-ayufan-ge70bd2ab8802 #1 SMP PREEMPT Thu Jul 26 08:33:14 UTC 2018 aarch64 aarch64 aarch64 GNU/Linux
Sorry for any mistakes. English is not my native language

1. Quartz64 Model B, 4GB RAM

2. Quartz64 Model A, 4GB RAM

3. RockPro64 v2.1

https://linux-nerds.org/
#3
(09-09-2018, 04:19 PM)crhawle Wrote: 1. Does using a higher quality pci to NVMe adapter improve the performance? Ideally, the PCI lanes should each be able to contribute 400 MB/s for a total of 1.6 GB/s. The only benchmarks i've seen cap out at 540 MB/s write and 1.3 GB/s read. The SSD being tested in those is rated to be able to achieve much higher performance in both areas. I don't know if the difference  is due to internal limitations of the processor or the PCI adapter itself. Hence my question.
I have 2 different PCIe/NVMe adapters. Neither noticably has any chip on them, not sure what else would materially affect quality. Both work fine, never felt either of them was faster or slower than anything else.
But for sure NVMe results (as per Bullet64 above) on iozone are hugely variable on ROCKPro64, in particular down to the kernel in use. 4.18 kernel results are significantly better than anything I have seen on 4.4 (as in 2 or 3 times better) which can make comparisons meaningless. Let alone the NVMe devices themselves.
  • ROCKPro64 v2.1 2GB, 16Gb eMMC for rootfs, SX8200Pro 512GB NVMe for /home, HDMI video & sound, Bluetooth keyboard & mouse. Arch (6.12 kernel, Openbox desktop) for general purpose daily PC.
  • PinePhone Pro Explorer Edition, daily driver, rk2aw & U-boot on SPI, Arch/SXMO on eMMC
  • PinePhone BraveHeart now v1.2b 3/32Gb, Tow-boot with pmOS/SXMO on eMMC


Possibly Related Threads…
Thread Author Replies Views Last Post
  ADATA XPG SX8200 Pro 2TB NVMe SSD Perf Atch 0 993 04-01-2023, 02:51 PM
Last Post: Atch
  Full Size Bracket for the NVME PCI-E Card publiclewdness 0 1,633 12-31-2021, 12:53 AM
Last Post: publiclewdness
  PCI-E X4 to M.2/NGFF NVMe SSD - support? AndyOfLinux 22 32,143 12-13-2021, 11:04 AM
Last Post: Cerberus
  zfs on rockpro64 - any dual nvme PCI adapter for software raid ? romadr 0 2,194 04-22-2021, 02:59 PM
Last Post: romadr
  case for RockPro64 + PCI-E M.2 NVMe Card? AndyOfLinux 9 14,856 01-05-2021, 05:05 PM
Last Post: kuleszdl
  12V 3A power supply and NVMe - PCIe kaylios 3 5,705 08-25-2019, 03:31 AM
Last Post: kaylios
  using PINEBOOK Pro M.2/NGFF NVMe SSD Interface Adapter abdel 4 8,936 07-26-2019, 09:48 PM
Last Post: pfeerick
  Pine64 NAS fan problems User 11259 3 6,343 04-13-2019, 05:28 PM
Last Post: g_t_j
  Discovered issues with PCIe NVMe & GPU Load in desktop 4.4.174 causing board lockup axelf 0 2,205 04-08-2019, 03:51 PM
Last Post: axelf
  Fitting nvme card in open enclosure vbb 0 2,177 12-29-2018, 02:54 PM
Last Post: vbb

Forum Jump:


Users browsing this thread: 1 Guest(s)