H
I have developed an ASP application that can send as much as say 150MB of XML data to the client, if all records (approx 6,000) in the database are returned. Normally, users would choose filtering options and at most about 500 records would be returned at any time. It is rare for this volume of data to be download, but it can happen
My application runs on a corporate intranet. A rival contactor's 'expert' believes that it is intolerable for my application to generate this much data under any circumstances as it would unduly load the network
Ofcourse it is a simple matter for me to elimate large volume transactions by limiting any client request to a lesser number of records. If they really wanted all records several transactions would be necessary. That is, a paging mechanism could be provided
Given that there are many very large documents, pictures, and other media held on this intranet server, which users can view at any time I don't perceive this as a problem. Its pretty much a double standard. Also, the server script timeout prevents the server from being overloaded.
What do people think
Tim