I'm starting to move from visualvm to java mission control. I use wildfly 8.2, which doesn't work with JMC out of the box, but have modified the JMC ini to include the required class path as per my answer in the jboss community https://developer.jboss.org/message/955742#955742
I've been able to connect successfully, however the time it takes to connect is ridiculous. Often waiting 1-10 minutes (during which time jmc often becomes unresponsive) when connecting to a local wildfly instance, it's almost instant, the delay happens only when connecting to a remote instance.
Does anyone know what would cause this to be so slow, and if anything can be done to sort it?
Related
I have a large application that takes a significant amount of time to publish on the JBoss server. Once this job is done, I can start/stop the server as many times I want. But the problem revolves round the fact that I have to republish every time I relaunch the eclipse even though the files are already present in the server, i.e. the state of the server is lost.
It would be of great help to know of any method that could relieve me from the overhead of publishing my application every time I start a new instance of eclipse and instead just start the server.
I recently set up Eclipse debugging for developing Bukkit plugins. It's really awesome being able to see what variables are changing and on what breakpoints, but my main concern is, when debugging with clients connected to my test Minecraft server, they timeout. This is due to Eclipse, when encountering a breakpoint, pauses the servers main thread and timeouts the clients after 30 seconds. I've set the server so it doesn't shut down after 60 seconds, but the clients always disconnect. Is there a client-side option to set timeout periods? Or is it embedded in the way TCP packets work? - Or just another workaround to this?
I know this is 3+ years later but I've been struggling with this myself (Using Intellij Idea instead of Eclipse though) and I've found a solution to this, which I'll post here in case anyone else comes across this. Since the readTimeout was not behaving as I wanted it to do, what I did is the following:
a) To set timeout on the client side:
Install Forge for the spigot server version you are developing (You can still join spigot based servers for plugin development).
Install RandomPatches mod (allows you to edit hard-coded vanilla minecraft variables)
Edit RandomPatches config file (randompatches.toml) and set "read_timeout_seconds" based on your liking (This should only happen for the client-side when debugging a plugin. You should not change it for production servers ofcourse).
b) To set timeout on the server side:
Edit spigot.yml file and set "timeout-time" based on your liking. If you only change the client-side, server is going to crash once this value is surpassed and you'll need to start the server again.
c) Voila! Enjoy debugging for as long as you like!
I am using jboss-4.0.5.GA for project deployment, but it is hanging at startup by displaying message
Trying to install an already registered mbean: jboss.management.local
it is showing this for 6 minutes and it takes total 7-8 minutes to start.
If I start server without deploying application, then also it's taking 7 to 8 min.
You seem to have corrupted the JBoss installation, highly reccomended to use a fresh installation and deploy your application rather than debug this.
I use Jboss 4.2.2 which is taken from a zipped folder. Today I have renamed the existing setup and tried unzipping again and use a fresh copy. This didn't work out as expected and server is in starting mode for several minutes and thrown error in the end. So removed this fresh setup and used the old one and it worked like a gem, though not sure where is the problem here.
I'm running on Win 7 using Eclipse 4.2 starting a web app on a Tomcat 7 server and using Derby database. I have tried many approaches but run consistently into a common problem:
Everything works just fine the first time I start up and run.
When I redeploy my application after a change, all database connections hang (any kind of restart).
If I stop Eclipse and restart Eclipse, that clears up the problem and the next run works fine again.
Having done some investigation, it appears that the problem is that the Derby port (1527) is not released from one execution of the server to the next. That seems very strange to me since Derby is started by the Tomcat instance which is a separate javaw process.
I've tried:
Configuring the Derby connection as a Tomcat resource
Establishing the connection within my code (rather than via Tomcat resource)
Both the embedded and the network driver
Starting / stopping the network driver from a servlet on startup and shutdown of the Tomcat server
Shutting down the embedded driver via servlet on shutdown of Tomcat
Again, every approach works fine to connect the first time.
One other symptom that doesn't appear to be related (except for as a possible indicator of whether or not shutdown completes correctly) is that the db.lck file for my database never gets deleted. However, whether or not it exists has no bearing on whether or not I can reconnect (only stopping/starting eclipse has an impact).
Any insight would be appreciated.
Thanks!
After some further investigation I'm going to call this a duplicate of: Cannot create JDBC driver of class ' ' for connect URL 'null' : I do not understand this exception. It's not quite the same thing, but that solution (creating META-INF/context.xml) allows it to proceed to failing calls rather than hangs, which is a significant improvement and suggests it's largely related.
I did finally figure this out. It turns out I had the derby jars in the Tomcat lib folder (for Tomcat) and in the deployment assembly for my application in Eclipse (rather than just in the build path). So Tomcat was using the built-in libs, while my app was using the embedded libs, and this resulted in conflicts. Leaving the libs as part of Tomcat and removing them from my war file solved the problem completely.
I want to know how can I speed up RSA 7.5( which is an IDE by IBM having eclipse under the hood with websphere server runtimes) mainly server start. The first time I start it after computer reboot it loads after, but after that it takes for ever to start/stop the server. The debug mode for server takes for ever to start.
I am using server 7 run time for IBM RSA 7.5.
So bascially RAD/RSA has websphere run times which allows to configure the server runtime start/stop within RAD/RSA. The run time allows you to develop webapps and test time on the server on deploy it on the websphere run time.
The problem I am facing is with the websphere runtime which works fine after computer reboot but is very slow after several deployments/publishing of the same web app.
I would be grateful you give performance tips for speed up RSA server start/shutdown and overall performance tips. I have plenty of memory like 12 GB with i7 Core 6 cores on Win7.
Of course of your are running the server in debug mode it's going to be a lot slower, but you have a few options like putting the server in development mode or doing some fine tuning as to which applications should start. Take a look at these articles:
Rational Application Developer Performance Tips- Case study: Tuning WebSphere Application Server V7 and V8 for performance
Performance tuning WebSphere Application Server 7 on AIX 6.1
WebSphere tuning for the impatient: How to get 80% of the performance improvement with 20% of the effort
WebSphere Performance Monitoring & Tuning
Some of these are a bit dated but they have some good information that may still be relevent to your issues; especially the first one.
Make sure that the workspaces are stored on a local disk.
edit - forgot this: buy a SSD disk. It makes a huge difference when developing.
If you have a virus scanner, disable on-access scan in the SDP installation directory including the server plugin, and in all your workspaces.
Uninstall any applications (ears) you are not using - the more you have installed the longer the server takes to startup. If your server is taking too long to start, RAD/RSA will assume it has timed out and stop it before it finishes starting - if this happens then increase the start timeout limit by double-clicking your server in the Servers tab and modifying the values in the Timeouts section.
Oh, and If you have a lot of datasources defined, and autostarting connection pools with alot of connections, it may also take a while to start the pools.
But that can't explain it all... I haven't tested, but since WAS and RSA seems to spend a lot of time doing absolutely nothing, I am starting to suspect it's trying to download schemas or something. If you have the time, you could try to trace and see if you find something like that...
I came across this post while trying to troubleshoot my RSA performance. I figured I would update it with a recommendation for improving performance on RSA 8.0.4.
http://publib.boulder.ibm.com/infocenter/radhelp/v8/index.jsp?topic=/com.ibm.performance.doc/topics/cperformance.html has some excellent tips on improving performance in the "Performance Tips" section. After implementing just some of the "Always" tips I've found my memory reducing significantly and performance being much faster.
You should start with the "Always" tips and then move to the "Sometimes" and "Rarely" ones for finer tuning.