hi guys, i'm with doubt about the "best" engine (best = no INSERT lag ~0.100s or less, good SELECT speed ~ 1 to 10 seconds is ok) i have two tables, they only "receive" INSERT and SELECT queries (no DELETE/UPDATE/ALTER) they increase 200MB/day and 3.000.000 rows/day my doubt is, how to have small database size and good read rate? it have ~ 104 inserts / second, but each insert is multi values like ... INSERT INTO table VALUES (),(),(),(),(),() .... in other words... today i'm using aria (it's crash safe) and i don't have problems with concurrent insert / table locks etc, my doubt is about table size and read speed with bigger database i'm considering using spider engine or any other shard system with >100GB , or maybe a big raid-6 system, server is a xeon 8core with 16GB and 2hd 500GB RAID1 SAS (>180MB/s) the client side is a C++ program that i don't have access to change and recompile it, in other words i can only change server side and load balance data table is something like: (datetime decimal(22,6),value decimal(22,6)) the SELECT code use most part of recent data ~5 DAYS, in outlier queries something near to ~1 MONTH ago ideas are wellcome -- Roberto Spadim