Quantcast
Channel: Adobe Community : Popular Discussions - Archived Discussions
Viewing all articles
Browse latest Browse all 21156

CPU usage increases when web server is down (ubuntu)

$
0
0

Hi All,

 

I have a strange bug that has been perplexing me for a few weeks now. It only seems to occur on Linux/Ubuntu and it involves http requests.

 

To give you a little background we've built a little http status monitor that pings a URL to check if a server is up or down. The problem is when we turn the server of the CPU usage of the application increases indefinately irrespective of the number or URLStream objects created. I this is leaft over a period of time the application locks up because of the high CPU usage.

 

Steps to reproduce:

 

1. Create a URL Stream or URLLoader to a domain name that can be resolved to an IP but the web server is unavailable and make a request (this can be done by either disabling the web server, or pointing to a fictitious domain and adding a fictitious IP to a local hosts file for this domain).

 

2. Request the URL repeatedly over a cycle of N seconds;

 

Results:

 

When attaching an IOErrorEvent listener the client will recognize that the URL cannot be reached and trigger the IOErrorEvent however when looking at the CPU usage level over a period of time. Given that N is 10 seconds or less you will notice a gradual and permanent increase in CPU usage (around 0.1% per 4 minutes). If the web server is then enabled the CPU usage will drop to a stable and lower level.

 

NB. When using the local hosts table the application will need to be restarted before CPU usage returns to a normal level.

 

Expected results:

 

When this scenario is tested on a Mac or Windows based machine the CPU usage level remains the same despite the same.

 

Attached is a simple project to demonstrate this.


Viewing all articles
Browse latest Browse all 21156

Trending Articles



<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>