mus1cb0x: juniper is cheaper. any idea of which is better? junos being fbsd based is a big selling point, but that's only 1 part of the story thanks mercutio i use openbsd and freebsd myself :) how would you use that to replace a 1U top of rack switch? oh i thought you were talking about routers. i'm using cisco switches. ex4200s are rather highly for juniper switches though. highly? not too expensive, and good enough with routers people seem to suggest mx480 normally at the lower end ex3200 is used too but i dunno, what are you looking for in a switch? oh nothing really, just curious how datacenter networking gear has evolved over the past 5 years it hasn't much afaik it's mostly just that prices have come down a bit, power use has come down a bit. is 10 gige to the client with 100gig uplink common in switches now? it's still expensive to do 10 gigabit. or is it still 1gige to client ahh nope. seems like they are holding on to 10gige prices yeah, this new 2.5 gigabit should help things. what's that? 2.5 gigabit is coming to desktops soon no shit? they can do 2.5 gigabit in similar power envelope to current gigabit. using existing cable. over regular cat 5 copper? yeah afaik is cat 6 copper still not worth it? i dunno, i've never experienced any cables that work better with cat6 than cat5e. and most cat6 is thicker than cat5e. but you can get similar awg cat6. do gbics still have any value? the current recommendation is meant to be for cat6a. the ones that went in cisco 2950 and 3550 switches gbics? is that just sfp modules? yea the uplink modules i can't remember what interface they have it's probably just ftp sfp i can't type. yea (i've been drinking) ah hey don't feel down, things will pick up :/ haha :P mostly gigabit is good enough for most people atm. 10 gigabit helps iscsi performance ah so for san environments but fibre channel and sas are fine too. i hear hot swap sas ssd drives are badass if you want to do 300mb/sec sas or fc is fine. if you want to do 80mb/sec gigabit is fine. if you have ssd san and want to do 3000mb/sec then you should just keep it local :) expensive too. i like infiniband myself That's what she said!! this is why i like server based storage instead of san stuff. storage being on the server is nice and fast, and noone expects network to be HD calibre latency i can't notice the difference been local and rmeote at home is infiniband still expensive? i have 16 gigabit infiniband limited to about 13 gigabit because of pci-e x4 limitations on one end. it was $150 US for 4 cards for me. not bad but they were running in pci-e 1 mode. i reflashed them to run in pci-e 2 mode. dodgy as i suppose :/ you using zfs? but i get about 1300mb/sec with nfs from ssd raid. zfs+mdadm root 3x120gb ssd atm. what kinda overhead are you seeing zfs impose? but i'm upgrading it, just waiting on one more ssd. zfs seems to be slower with lots of small files. how much slower tho but with sequential it can still max out stuff. i dunno it's just a feeling, id on't benchmark. well i don't have a good benchmark :) i was playing with tar the other day, and it's faster to extract a file to /tmp then to local storage. err than http://kotaku.com/the-new-zelda-is-open-world-looks-absolutely-incredibl-1588673841 so i had this great idea of trying to figure out how to make tar work with async i/o nothing on linux uses good queue depths for ssd's which is fine for hard-disks.. also xz wasn't multithreaded. new version that just came out is. been hoping arch updates it soon also one of my ssd's seems to like to randomly drop out sometimes. and mdadm needs to sync the whole disk. and i manually need to add it back in. whereas zfs just catches up ssd performance still has quite a lot of potential for improvement. there can be random latency spikes etc. but to get maximum performance out of that kind of technology more and more has to be exposed to the OS instead of acting like a dumb hard-disk nvme may help a little. the new slc write caches can help more desktop-like work quite a lot. but most stuff doesn't even need to be synchronous, it's just there aren't good interfaces. like if i'm installing a bunch of packages, as long as everything i installed gets reversed i'm fine with it chunking into one big transaction zfs seems to love ram and a little extra cpu i have i7-4770 and 16gb of ram on my main linux box at home. upgrading to 32gb of ram. but i have it on one 8gb server with raid, and one 8gb desktop with no raid and it's fine on both. there were some early fragmentation issues ddr4 should make 64gb more affordable though i figure just stick more ram in for the most part. but haven't the most frequently used and most recently used means it behaves pretty well when doing cache thrashing. openbsd is actually adding some kind of simple dual queue system too afaik. well someone was proposing it, i dunno if it's in. but disk caching is an area that quite a lot of research has been done in, but most people are doing simple systems for. one issue, is how do you test performance. like one of the common issues with most recently used is that it unloads common binarys when you do "benchmarking" type applications. and that may make the benchmark work better, especially if the benchmark doesn't used that much disk. but it makes my pet peeves worse: starting vim to edit a file, tab completing a file, and logging in to a machine. good benchmarks are easy for me to write i just take the same group of tasks and repeat the performance of it the thing is, interactive things want lower latency for reads, background things matter less. it's kind of complicated. synchronous things want lower latency too. so my current thinking is that things should just be shifted to be more async. most things shouldn't really be much faster on ssd than hard-disk. because they should be reasonably predictable, and be able to just be read sequentially i like async too damn what happened to coding hah i'm writing a tiny curl implementation why? it works, but i got distracted with trying to use as little cpu as possible it's cos i did ldd on curl :) hahahhaa oops :) and because iw as noticing that curl used heaps of cpu on infiniband but i tested on some other systems. there are quite a few problem cases, and performance can vary. but something like downloading on adsl uses a surprising amount of cpu because packets don't get coalesced that well you get a packet every msec or such. so it has to wake up, process a packet, wake up process a packet etc. but on xeon e3110 (core2 duo 3 ghz basically) there's significant cpu usage maxing out gigabit. and fiddling with the coalescing settings makes a difference (intel e1000) but my initial testing was all infiniband, and then i upgraded my server from i5-2500k to i7-4770 my desktop is i7-3770 nice i7-4770 uses WAY less cpu than i7-3770 at curl type applicaitons. it kind of blew my mind. the i7-3770 even has faster ram. basically cache speed has gone up heaps afaik. but all the benchmarks floating around are about games. where the differences seem much smaller. there's really very little in the way of benchmarks for more server type stuff, other than pointless stuff like requesting static files on a web server. which basiclaly doesn't matter curl http://192.168.50.254:24/200m > /dev/null 0.01s user 0.05s system 98% cpu 0.054 total ./microcurl/microcurl http://192.168.50.254:24/200m > /dev/null 0.00s user 0.03s system 91% cpu 0.033 total files downloading at localhost to /dev/null is way faster with my system :) curl http://202.49.140.24:24/10m > /dev/null 0.04s user 0.07s system 4% cpu 2.408 total ./microcurl/microcurl http://202.49.140.24:24/10m > /dev/null 0.00s user 0.07s system 2% cpu 2.406 total and a little less cpu over vdsl. ok i should stop rambling :) haha it was good stuff