aboutsummaryrefslogtreecommitdiff

print("Hello, World!")

One program in a potpourri of programming languages.

Benchmark 1: ./testc
  Time (mean ± σ):      16.8 ms ±   3.4 ms    [User: 16.5 ms, System: 0.6 ms]
  Range (min … max):    13.0 ms …  28.4 ms    191 runs
 
Benchmark 2: ./testcpp
  Time (mean ± σ):      17.8 ms ±   4.0 ms    [User: 17.2 ms, System: 0.8 ms]
  Range (min … max):    13.5 ms …  29.1 ms    175 runs
 
Benchmark 3: python test.py
  Time (mean ± σ):     430.7 ms ±  23.4 ms    [User: 426.1 ms, System: 4.0 ms]
  Range (min … max):   388.5 ms … 466.9 ms    10 runs
 
Benchmark 4: java test
  Time (mean ± σ):      62.3 ms ±  11.7 ms    [User: 57.5 ms, System: 8.0 ms]
  Range (min … max):    49.2 ms …  87.5 ms    41 runs
 
Benchmark 5: mono test.exe
  Time (mean ± σ):      32.8 ms ±   6.6 ms    [User: 29.9 ms, System: 3.1 ms]
  Range (min … max):    25.9 ms …  48.5 ms    101 runs
 
Benchmark 6: node test.js
  Time (mean ± σ):      61.7 ms ±   8.4 ms    [User: 57.0 ms, System: 5.5 ms]
  Range (min … max):    50.4 ms …  79.1 ms    53 runs
 
Benchmark 7: perl test.pl
  Time (mean ± σ):     666.4 ms ±  33.8 ms    [User: 663.9 ms, System: 1.5 ms]
  Range (min … max):   619.0 ms … 725.2 ms    10 runs
 
Benchmark 8: ruby test.rb
  Time (mean ± σ):     249.1 ms ±  13.7 ms    [User: 243.7 ms, System: 4.9 ms]
  Range (min … max):   229.8 ms … 271.4 ms    12 runs
 
Benchmark 9: ./testgo
  Time (mean ± σ):      21.4 ms ±   5.3 ms    [User: 21.0 ms, System: 0.8 ms]
  Range (min … max):    16.9 ms …  34.7 ms    153 runs
 
  Warning: Statistical outliers were detected. Consider re-running this benchmark on a quiet PC without any interferences from other programs. It might help to use the '--warmup' or '--prepare' options.
 
Benchmark 10: ./testrs
  Time (mean ± σ):      22.0 ms ±   5.4 ms    [User: 21.6 ms, System: 0.6 ms]
  Range (min … max):    14.5 ms …  31.1 ms    167 runs
 
Benchmark 11: ./tesths
  Time (mean ± σ):      90.7 ms ±   5.8 ms    [User: 89.2 ms, System: 1.5 ms]
  Range (min … max):    82.2 ms … 100.2 ms    33 runs
 
Benchmark 12: julia test.jl
  Time (mean ± σ):     262.2 ms ±  11.4 ms    [User: 213.5 ms, System: 48.3 ms]
  Range (min … max):   250.8 ms … 291.6 ms    11 runs
 
Benchmark 13: Rscript test.r
  Time (mean ± σ):      3.086 s ±  0.084 s    [User: 3.065 s, System: 0.020 s]
  Range (min … max):    2.913 s …  3.235 s    10 runs
 
Benchmark 14: ecl --shell test.lisp
  Time (mean ± σ):     790.7 ms ±  24.2 ms    [User: 1190.7 ms, System: 93.0 ms]
  Range (min … max):   757.7 ms … 831.1 ms    10 runs
 
Benchmark 15: lua test.lua
  Time (mean ± σ):     137.0 ms ±   9.8 ms    [User: 136.2 ms, System: 0.8 ms]
  Range (min … max):   120.4 ms … 158.4 ms    21 runs
 
Benchmark 16: elixir test.exs
  Time (mean ± σ):     298.4 ms ±  11.9 ms    [User: 416.0 ms, System: 50.9 ms]
  Range (min … max):   272.6 ms … 319.2 ms    10 runs
 
Summary
  './testc' ran
    1.06 ± 0.32 times faster than './testcpp'
    1.27 ± 0.41 times faster than './testgo'
    1.31 ± 0.42 times faster than './testrs'
    1.95 ± 0.56 times faster than 'mono test.exe'
    3.67 ± 0.90 times faster than 'node test.js'
    3.70 ± 1.03 times faster than 'java test'
    5.39 ± 1.15 times faster than './tesths'
    8.14 ± 1.76 times faster than 'lua test.lua'
   14.80 ± 3.13 times faster than 'ruby test.rb'
   15.59 ± 3.26 times faster than 'julia test.jl'
   17.74 ± 3.69 times faster than 'elixir test.exs'
   25.60 ± 5.41 times faster than 'python test.py'
   39.61 ± 8.34 times faster than 'perl test.pl'
   47.00 ± 9.71 times faster than 'ecl --shell test.lisp'
  183.44 ± 37.82 times faster than 'Rscript test.r'

¯\_(ツ)_/¯