50 million records under one domain using Bind
Bill Larson
wllarso at swcp.com
Tue Dec 30 11:39:32 UTC 2008
On Dec 29, 2008, at 11:35 PM, David Ford wrote:
> I use DLZ w/ postgres. It's been working pretty good for me for a
> while
> now.
Another "just out of curiosity" question. What sort of performance do
you see with BIND/DLZ/Postgres?
The http://bind-dlz.sourceforge.net/ site has some BIND-DLZ
performance test results listed. I don't know what version of BIND-9
they were using and I'm sure it is not current. With straight BIND-9
they were seeing 16,000 QPS, a reasonable number. With the Postgres
DLZ they saw less than 600 QPS. I'm sure that this performance can be
improved with fast hardware and (hopefully) a newer version of BIND.
With 50 million records, it would take about one day to perform a
single query for each of these records with the server doing nothing
else. It doesn't appear to me that you could serve this many records
using BIND-DLZ with Postgres in any environment that actually uses all
50 million RRs. Then again, at 16000 QPS, it would still take about
an hour to perform a single query for each of these 50 million records.
Granted, the startup/reload speed increase using DLZ will be
impressive, what I am questioning is having 50 million DNS resource
records on any DNS system. Is DNS an appropriate "database" for
storing 50 million records?
Bill Larson
> -david
>
> Andrew Ferk wrote:
>>> What are the backend database options available? Is bind-sdb active
>>> developed and is it production ready?
>>>
>>
>> You can use mysql with dlz. I have yet to get it successfully
>> working, but that's another issue.
>>
>> One of the reasons I wanted to use a database was for the speed
>> increase. I would probably look into using dlz.
>
> _______________________________________________
> bind-users mailing list
> bind-users at lists.isc.org
> https://lists.isc.org/mailman/listinfo/bind-users
More information about the bind-users
mailing list