By using this site, you agree to our updated Privacy Policy and our Terms of Use. Manage your Cookies Settings.
446,187 Members | 1,038 Online
Bytes IT Community
+ Ask a Question
Need help? Post your question and get tips & solutions from a community of 446,187 IT Pros & Developers. It's quick & easy.

.NET Client - Server application gets frozen, need ideas to fix!

P: 2
You cannot vote on your own post
Helloes good people of,
I have a .NET client-server application running with a few hundred of clients. The project was migrated from VB6 to .NET about a year ago and it's a platform for card/board games. Although I'll be trying to give as much detail as I can below, the problem is getting a channel frozen when there are 40-70 players inside.


1. Server (.NET 4.0)

* Divided into three projects: ServerNET, Listener, Channel
* Listener acts like a login server where clients connect first. It is responsible for checking stuff like version and account info. Also lets the client choose which channel to connect. It's basicly a TCPListener in a do-while, listening anyone trying to connect forever. It is not the reason why both sides get frozen.
* Channel represents a single port, clients get connected to Channels after they are done with Listener. Much like a space shuttle, this is the main part. Similar to a MIRC channel, it binds all users inside, most of the data is sent to people within the same channel like chat and the games you can join created by other players, hosted by server. This is a console application and serves as a hub for players. Player info is held in "Client" class which includes a TCPClient and some other properties. Each client runs with a thread and makes async calls which are handled by the server. Also these "Client" objects are held in a collection class named "ClientCollection". Channel gets frozen when there are roughly 40-70 players inside. There is a maximum limit of 100 players permitted per channel.
* ServerNET is the body and does all other general stuff related by the whole system, not channel spesific. This is a form-application and runs stuff like server options.

2. Client (.NET 2.0)

* Runs with TCPClient, mostly single thread whereas server is multi-thread.
* Must use .NET 2.0.
* Mostly consisting of visuals and other non-important stuff.

When there are 40+ clients connected to a single channel, it starts to get frozen totally randomly (or that's what we have right now, got no evidance or enough datas to point out what's totally wrong). We really don't think network traffic is the issue (not quite sure yet) since we have tried it on different server machines with various setups. All the server machines we have used are capable of handling that much of process hardware-wise. So it is about the approach and what's going on code side.

The reason why we are struggling to address the issue is we are not exactly sure what could be causing it. Please check out the following example:
System A has 55 people online in their Channel #1 and it doesn't get frozen anyhow. System A uses A1 IP and the channel is on 16xxx port.
System B has 25 people online in their Channel #4 and it gets frozen like one or two minutes randomly. System B uses B1 IP and 18xxx for the channel port. It's on the same machine with System A which doesn't get frozen.

As a conclusion, it looks irrelevant with the number of online people but it occurs more often when numbers rise.

We tried rolling an Application.DoEvents() in an endless do-loop in Channel project thinking that some X process causing the channel to go frozen state for a few minutes, thus resulting a pause in channel. Then it performs every action which was queued while it was frozen, in a few seconds. CPU usage is averagely between 7%-20% per channel, it looks like it is getting better. However it was no permanant and effective solution.

Things we suspect:

* ClientCollection that holds players and TCPClients is inherited from CollectionBase. Maybe this is causing some chaos during sync'ing. This used to be an array back in the day and we were having less of these problems. Maybe it shouldn't be inherited from CollectionBase, but something else?
* We are using SyncLock (lock in C#) to sync ClientCollection. (although we had this problem before we started using locks)

Server info
Intel Xeon X3460 2.80GHz
64-bit Windows Server 2008 Enterprise

I know it is impossible to address the issue without seeing the whole code, but I regret that I'm unable to post the codes. Instead I'm looking for an idea to put me into some direction. However we are happy to share any other info for resolving this problem.

Thanks to everyone helping!
Mar 1 '11 #1
Share this Question
Share on Google+
1 Reply

Expert 100+
P: 797
I have a number of strategies for debugging this kind of situation. You mention it's just frozen a channel so I'm assuming your server app isn't hanging or crashing entirely. I'm not entirely sure how your server code is written, if it's a windows app, console app, windows service etc. What is the nature of the use of your ClientCollection? Are you doing frequent lookups? Would this perhaps be better served as a dictionary which would increase performance of client lookups?

I've written apps (for different purposes) with similar architectures, threaded listener/tcp client collections etc.

I would spin up your server app inside a console app which will allow you to debug any commands being sent from the client to make sure it really does have to do with size of the collection and not some arbitrary command that is freezing the system.

I would also mock up a client application that can spin up lots of connections to load test your server application, mimicking realistic use. This will allow you to mimic a more realistic load than 2 or 3.

In cases like this where load appears to be causing the problem, my first thought leads to race conditions, make sure your locking mechanisms or at the most granular level possible. Make sure they're locking objects correctly such that the act of locking/unlocking isn't in itself causing a race condition at that kind of load.


Have you checked that your read mechanism that's reading the data from the buffer is passing off the data quick enough so that the tcp client isn't hanging waiting for the data from the buffer to be processed? That can cause problems, though you'd likely see this even with only a single client. Just thinking out loud...

I will have a bit more thought and if I come up with strategies to debug this, I'll add to this post.
Mar 2 '11 #2

Post your reply

Sign in to post your reply or Sign up for a free account.