I'm running a Railo project on my local Ubuntu box, Eclipse Indigo, Tomcat 7, Fusebox 4 and AWS Toolkit for Eclipse
I have my project running smoothly on my local dev box. Trying to deploy the project at times takes under 10min (very rare) and other times it never happens, just have the loading bar showing and/or eventually fails.
I've tried publishing a new project which works at times, but when trying to do incremental deployment it just almost never happens (it worked once). In fact after the one time when I deployed the project I've not been able to do it again.
Unable to upload application to Amazon S3: Unable to calculate MD5 hash: /home/workspace/.metadata/.plugins/org.eclipse.wst.server.core/tmp1/aws-eclipse-1365821331354619151.war (No such file or directory)
Unable to calculate MD5 hash: /home/workspace/.metadata/.plugins/org.eclipse.wst.server.core/tmp1/aws-eclipse-1365821331354619151.war (No such file or directory)
Rightly so, there is no such file in that location. But why? Is it a permissions issue? I gave myself root right for GUI file browser but still no joy (gksu nautilus).
I'm new to AWS and Ubuntu environment and not sure what i should do in order to deploy.
So one of your questions seems to be problems in uploading files to S3 via the AWS java SDK right?.. on line 1011 https://github.com/amazonwebservices/aws-sdk-for-java/blob/master/src/main/java/com/amazonaws/services/s3/AmazonS3Client.java#L1011
So what I think your doing here is trying to execute a putObject S3 command which had a Content-MD5 hash saying the request could not be authenticated. I had this issue I found out that the MD5 hash needs to be base64 encoded and is required by Amazon to upload files.
However, looking at your error "No such file or directory" this could be a different issue. Doing a quick Google I found a post that could be of interest...
https://forums.aws.amazon.com/message.jspa?messageID=143497
Hope some of this helps.
Related
I'm using two laptops and stored my C# code in OneDrive.
I am aware that sharing code via OneDrive is not be the best approach, but that's what I'm dealing with now.
I noticed that on laptop 1 I have to define the following path to the data file (mdf):
C:\Users\ Diet\OneDrive\Personal\VisualStudio2019\Repos\project\project\App_Data\data.mdf
On laptop 2, the path is different because the user I'm logged in with has a different name (or at least that's what I believe is the cause)
C:\Users\ Dieter\OneDrive\Personal\VisualStudio2019\Repos\project\project\App_Data\data.mdf
Updating this in the Web.config fixed the connection to the database, BUT building the solution still returns an error, also related to a cloud operation, hence why I think it is caused by the path in OneDrive...
The error message:
CSC : error CS0041: Unexpected error writing debug information -- 'The cloud operation was unsuccessful.
I welcome your insights. Thank you for helping me out.
I have my projects stored in OneDrive and had this same issue. The fix was to set the entire Project folder contents to "Always keep on this device".
Seems that building the solution in VS was attempting to write to files that were not cached locally from OneDrive. As soon as I changed the setting, the build worked!
I was also storing my project on OneDrive, got the same error after installing a new ssd.
Rebuilding the solution was enough for me.
I downloaded the latest Unreal Engine repo from github and opened the setup.bat file as per the instruction on the Official Page. The command-line opened-up and started downloading dependencies and everything looked fine until it reached 30% progress, after which it threw an error:
Checking dependencies...
Updating dependencies: 30% (12936/41030), 1702.0/9123.9 MiB | 5.04 MiB/s...
Failed to download 'http://cdn.unrealengine.com/dependencies/3215544-fbbce13ceb6f4aaea59e20af2d659d08/36799aacce147eba397f6eacc7a0f6c9dac946d8': Can't read from pack stream (CorruptPackFileException)
Initially I thought that it was a connection issue even though I was on a 40-Mbps connection. But it seemed to get stuck on 30% with the same error everytime. So, I started searching for solutions on the internet but could not find any, except for the clue that it was possibly a connection issue.
Furthermore I also tried to clone the repo again and again with the same result (stuck on 30%)!!!
Solution:
Goto
dir-containing-the-repo \ UnrealEngine \ Engine \ Build
Open the file (which should be in the Build folder) with your favourite editor (i.e. VS Code, Notepad):
Commit.gitdeps.xml
Search for this text:
BaseUrl="http://cdn.unrealengine.com/dependencies"
Just replace the http:// with https:// and leave the rest as it is:
Your BaseUrl should now be:
BaseUrl="https://cdn.unrealengine.com/dependencies"
Save the file.
Run Setup.bat from the folder: dir-containing-the-repo \ UnrealEngine
This happened to me as well, but it turns out I ran out of disk space on my Linux machine in AWS. I was using 8GB.
Having a bigger volume (200GB) fixed the error for me on the next run
I'm following the tutorials and samples of the official WSO2 ESB documentation. When I create (or load) a synapse configuration in the WSO2 Developer Studio I'm getting the error as shown below:
Save problems
Save could not be completed.
Reason:
resources\sunErrorHandler.esb_diagram (The system cannot find the path specified)
This error shows up when I want to open anything in the Design view. Even when I'm not trying to save anything.
So far I have tried to add the path specified in the error, but without success. I already found a workaround to be able to save and deploy my ESB configurations. By using "Save as" I'm able to keep the changes I made in the XML files. However the error stills shows up when I open the new version of the file.
Does anyone know a better solution to permanently resolve this error message?
I had this problem as well, running eclipse.exe as administrator fixed it.
Issue remains unresolved for my Eclipse + DevStudio 3.8.0 install on a Windows 10 environment. However since I was experiencing some other issues with the WSO2 ESB I decided to migrate to a Linux (Ubuntu) environment. This seems to work much better, the eclipse does not seem to have this issue on Linux.
Eclipse want to write in C:\Program Files\eclipse\resources. I solve problem by adding full permissions to group users on folder C:\Program Files\eclipse
If this happens in Mac you can provide the required permission to the "/Applications/IntegrationStudio.app/Contents/MacOS/resources" folder.
for the past several days I've been experiencing this error, while publishing to either JBoss EAP 6.3 or Wildfly 8.2 from Eclipse.
Error renaming D:\Servers\wildfly-8.2.0.Final\standalone\tmp\tmp9064011157118650757.jar
to D:\Servers\wildfly-8.2.0.Final\standalone\deployments\BusinessService.war\WEB-INF\lib\spring-web-4.2.3.RELEASE.jar.
This may be caused by incorrect file permissions, or your server's temporary deploy
directory may be on a different filesystem than the final destination. You may adjust
these settings in the server editor.
The problem occurs when I "Add and Remove..." projects from the server, then try to publish them, so the server can start.
I've experienced this issue on two different machines (home (Wildfly) and work (JBoss EAP)).
I'm using:
Windows 7 / 10
Eclipse Mars / Luna
JBoss Tools plugin 4.3 / 4.2
JDK 1.8.0.66 / 1.8.0.65
Maven
Building with maven from Eclipse and from the command line makes no difference. The server is configured to deploy projects as compressed archives. On both machines my user has administrator rights and has full rights on the server directory.
So far I've tried:
recreating the server multiple times with different configurations
using a newly created workspace
reinstalling JBoss Tools
reinstalling Eclipse
using different JDK versions
I'm really at a loss here and I don't know how to proceed in resolving this issue. Please help.
If you are using Windows, the path could get too long and can cause this error. A simple fix is to move WildFly closer to the root.
I had the same problem and solved it like this:
First of all, stop Server (Servers->WildFly(rigth click)->Stop), than clean. So you can run server again.
I had this problem several times in my new windows 10 machine that my employer gave me. Since I did not have admin rights it was a hectic process to troubleshoot this issue. Simple fix would be moving JBOSS_HOME closer to root. However, you need to do a proper restart of your eclipse. I rather recommend a complete restart of your computer because after all you are going to change JBOSS_HOME in windows environmental variables.
This is related to permissions issue on wildfly folder. Allow full control to the wildfly folder.
https://issues.jboss.org/browse/JBIDE-18697
I have moved the wildfly home to reduce the overall path length, and also removed any non-alphanumeric characters from the folder name (like "-" and "." ) . This worked for me, everything else (removing tmp, deployment, rebooting wildfly, rebooting eclipse, rebooting computer) failed.
I also suspect that the issue was stemming from running Wildfly from a ConEmu and/or git bash shell. Running from a plain CMD shell seems more robust.
I also got stuck with the same problem. I tried the below steps and it worked:
Clear the deployments and tmp folder in standalone folder in wildfly folder.
Delete the server and again add the server
Make a build of the project and start the server after successful build.
This is a terribly annoying error that either the Eclipse team or Redhat need to fix.
The solution is to close Eclipse, right click on the icon -> Run As Administrator. This solved it for me.
When deploying my project to SpringSource dm Server, every once in a while a JAR fails to deploy with the following message:
/mnt/myproject/springsource/work/com.springsource.server.deployer/packed/my.project.0.1.10.M.jar' cannot be unpacked.
java.util.zip.ZipException: error in opening zip file
There are 5 .war files in the project. If one of them fails, it's always the same one (which is also the last one to be copied into the pickup directory). However, usually all 5 will deploy without issues. It is the exact same set of files in all instances, taken from a maven repository, just deployed to new server instances.
The file that fails can be opened just fine by 7-Zip. If I stop Spring, clear the pickup directory, start Spring and copy the .war files to pickup again, it will usually work.
The usual deployment process is:
Start Spring
Wait until it reports Open for business with profile 'web'
Copy all 5 projects with a 2 second delay between each copy (scripted).
Similar issues java-util-zip-zipexception-error-in-opening-zip-file and jboss5-cannot-deploy-due-to-java-util-zip-zipexception-error-in-opening-zip-fil do not seem to apply.
You don't say which version of dm Server you are running, so I would recommend upgrading to 2.0.x to pick up fixes if you haven't already. You may also like to upgrade to Eclipse Virgo which is the continuation of the dm Server project.
My guess is that the heuristic in dm Server for determining when a file copy into pickup has terminated is playing up, possibly due to a slow or irratic copy operation. Is there anything unusual about your disk, such as encryption or remote mount, which may interfere with the copy operation?
One way to rule out the heuristic would be to place the files in the pickup directory when dm Server is not running and then start dm Server when the copy operation has definitely completed. If the problem reproduces, then there may be a problem in the JRE you are using.