September 2015 - March 2016
Mike Schmidt, Jeff Weber, Steve Emmerson, Tom Yoksas
0.25 degree GFS data was added to the CONDUIT data stream starting with the 12Z run on July, 28 . Monitoring has shown that peak CONDUIT data volumes increased from approx. 8 GB/hr to approx. 21 GB/hr for all forecast hours for the 0.25 degree GFS.
``bqb
Data Volume Summary for emo.unidata.ucar.edu
Maximum hourly volume 38312.123 M bytes/hour
Average hourly volume 19769.266 M bytes/hour
Average products per hour 327728 prods/hour
Feed Average Maximum Products
(M byte/hour) (M byte/hour) number/hour
CONDUIT 7878.758 [ 39.854%] 21708.893 89182.364
NGRID 5206.945 [ 26.339%] 8596.371 37461.750
NEXRAD2 2836.811 [ 14.350%] 4648.726 41083.250
NEXRAD3 1578.781 [ 7.986%] 1886.712 89978.795
FNMOC 1274.134 [ 6.445%] 3163.891 3624.205
HDS 282.288 [ 1.428%] 579.506 19087.614
NOTHER 274.371 [ 1.388%] 783.128 1360.091
NIMAGE 155.168 [ 0.785%] 251.517 181.523
GEM 76.094 [ 0.385%] 474.508 792.295
IDS|DDPLUS 65.822 [ 0.333%] 81.338 44297.227
EXP 47.827 [ 0.242%] 91.171 328.068
FNEXRAD 45.000 [ 0.228%] 83.920 43.614
UNIWISC 44.757 [ 0.226%] 90.713 21.386
LIGHTNING 2.395 [ 0.012%] 6.377 284.977
GPS 0.113 [ 0.001%] 0.997 1.045
``bqe
Recently, top level IDD relays and the sites that they are feeding CONDUIT data to have been experiencing unusually high latencies that correspond with the transmission of the 0.25 degree GFS data. Current testing suggests that a large fraction of the latencies being experienced originate at or near NCEP. Investigations are ongoing.
(.xml for machine access)
http://www.nws.noaa.gov/os/notification/tin14-28hrrr-cca.htm
http://www.nws.noaa.gov/os/notification/tin13-43estofs_noaaport_aaa.htm
Briefly, these additions are comprised of:
The IDD relay cluster, described in the June 2005 CommunitE-letter article Unidata's IDD Cluster, routinely relays data to more than 1250 downstream connections.
Data input to the cluster nodes averages around 20 GB/hr (~0.5 TB/day); average data output from the entire cluster exceeds 2.9 Gbps (~32 TB/day); peak rates routinely exceed 6.4 Gbps (which would be ~70 TB/day if the rate was sustained).
Cluster real server backends and accumulator nodes routinely have instantaneous output volumes that exceed a Gpbs. Bonding of pairs of Ethernet interfaces was needed to be able to support these output data rates. The next generation of cluster machines will need to have 10 Gbps Ethernet capability.
The increase in IDD data volume over the past six months is largely attributable to the addition of 0.25 degree GFS data to CONDUIT, the overall increase in the volume of data being transmitted in NOAAPort (which now routinely exceeds 10 GB/hr), and the increase in dual polarization NEXRAD data. During the end of August/beginning of September GOES-R test period, the NOTHER datastream (which contains simulated GOES-R products among other things) pushed the total volume of data being sent over NOAAPort to peaks in excess of 20 GB/hr.
We support the following goals described in Unidata Strategic Plan:
The IDD-Brasil, the South American peer of the North American IDD operated by the UPC, is helping to extend real-time data delivery outside of the U.S. to countries in South America and Africa. The Universidad de Costa Rica is experimenting with relaying data received in the IDD to Colombia.
Prepared March, 2016