I'm using the following Talend Exchange components in Talend v6.3.1 to interact with a Windows shared directory (from a Linux server):
* tSambaPut
* tSambaGet
* tSambaList
* tSambaFileProperties
* tSambaDelete
* tSambaWaitForFile
SMBv2 and v3 are on the Windows server, but when v1 is disabled the Samba components cannot communicate with the server. We need to remove v1 and have a large number of jobs we'd like to avoid refactoring. Any suggestions are appreciated.
SmbException: the remote file cannot be found!
Exception in component tSambaPut_1
java.lang.Exception: SmbException: the remote file cannot be found!
at edhproduct_611.testsamba_0_1.TestSamba.tSambaPut_1Process(TestSamba.java:2385)
at edhproduct_611.testsamba_0_1.TestSamba.tFixedFlowInput_2Process(TestSamba.java:2156)
...
Indicated line numbers do not map properly to the generated code.
Related
I am trying to use JasperReports integration for the first time. I am using the included Jetty server, Oracle database XE 18c and Windows 7.
I am following the quick start guide https://github.com/daust/JasperReportsIntegration/blob/main/src/doc/github/installation-quickstart.md
I downloaded the zip folder, configuired database access through adding schema credentials in application.properties file as follows...
[datasource:default] type=jdbc
url=jdbc:oracle:thin:#localhost:1521:XEPDB1 username=hr password=hr
this parameter is limiting access to the integration for the specified
list of ip addresses, e.g.:
ipAddressesAllowed=127.0.0.1,10.10.10.10,192.168.178.31 if the list is
empty, ALL addresses are allowed.
Then I deployed the jri.war file successfully. Then I started the server successfully as well. But when I tried to test it through http://localhost:8090/, I got the following page, and I do not know if that's the norm or there's something wrong...
I need to know if testing is successful, and what's meant by "context" here?
Thanks
You deployed the jri.war to the context path /jri, this isn't an error, and is quite normal.
Just access your webapp via http://localhost:8080/jri/
I am trying to test for the first time Vault with Oracle Database on windows.
I made first steps of the tutorials but while I execute
vault write D:\Applications\vault_1.4.0_windows_amd64\oracle-database-plugin sha256="DEDDFSQ23EF" command=vault-plugin-database-oracle
I have got an error
Error writing data to D:\Applications\vault_1.4.0_windows_amd64\oracle-database-plugin: Error making API request.
URL: PUT http://127.0.0.1:8200/v1/D:%5CApplications%5Cvault_1.4.0_windows_amd64%5Coracle-database-plugin
Code: 404. Errors:
* no handler for route 'D:\Applications\vault_1.4.0_windows_amd64\oracle-database-plugin' here
I can't understand what's the issue.
what I did to fix this:
Inside Vault directory, I have created plugins directory where I have dwnloaded the right oracle plugins. Then I have used this command on cmd or Powershell on windows
vault write sys/plugins/catalog/database/oracle-database-plugin sha256="8A8ABE17E7A2A75BD871RE43D" command=vault-plugin-database-oracle
I create a step of database script which access to SQL Database Service in Workload Scheduler Service. When I run the process, the step encountered the error below.
error message
AWKDBE018E Cannot access required JDBC Driver folder
message information
http://www-01.ibm.com/support/knowledgecenter/SSGSPN_9.2.0/com.ibm.tivoli.itws.doc_9.2/common/src_ms/awsmsawkdbe.htm?lang=en
AWKDBE018E Cannot access required JDBC Driver folder
Explanation
The job was not able to access a JDBC Driver folder, you might not
have enough permissions.
System action
The operation is not performed.
Operator response
Verify that you have enough permissions.
This message seems to ask me to grant the proper authority to the job user. But there is no property to specify the job user of Workload Automation Agent. I use a Workload Automation Agent provisioned by Bluemix automatically.
Could you teach me which parameters are needed ?
Database script step information
JDBC driver class path info
I checked the path by the following "ls -lR" command step's log.
it seems to have a problem with the agent, I tried to replicate the same job type but it is not working with the same error message (even using different solutions for jdbc driver path).
If you are using the Workload Automation Agent that is created for you then you could open a support ticket to have the Workload team look at that agent.
Edit after having support from service team:
in the jar classpath field for a predefined workload scheduler process you have to put only the path to the directory containing jar files, without putting the jar file name to use.
So, according to current Workload Scheduler documentation, you have to use the following value:
/home/wauser/utils
By this way the database script works fine.
(screenshot added)
It looks like it is having issues referencing the location to the JDBC class path for DB2. Can you please double check the location for the class path for the DB2 driver?
Even though old, I wanted to make some quick checks.
This is tested on a 9.5 FP1 dynamic agent, part of the container delivery. The path values are the standard values for the container.
Try 1 - full path - SUCCESS
<jsdldatabase:driverPath>/opt/wa/TWS/jdbcdrivers/db2/</jsdldatabase:driverPath>
= Status Message: Success
= Exit Status : 0
Try 2 - relative path - FAIL
<jsdldatabase:driverPath>./jdbcdrivers/db2/</jsdldatabase:driverPath>
Job status : FAIL
===============================================================
AWKDBE018E Cannot access required JDBC Driver folder
===============================================================
Try3 - variable in path - FAIL
<jsdldatabase:driverPath>${UNISONHOME}/jdbcdrivers/db2/</jsdldatabase:driverPath>
===============================================================
AWKDBE018E Cannot access required JDBC Driver folder
===============================================================
Try4 - variable in path - FAIL
<jsdldatabase:driverPath>$UNISONHOME/jdbcdrivers/db2/</jsdldatabase:driverPath>
===============================================================
AWKDBE018E Cannot access required JDBC Driver folder
===============================================================
So put short you need an absolute path into that parameter.
BUT, you can set the path in a config file global to the agent
Try5 - variable in agent config -
Inside IWSDATA Home : wadata/JavaExt/cfg/DatabaseJobExecutor.properties, write the following line
jdbcDriversPath=/opt/wa/TWS/jdbcdrivers
then remove the xml element about driver from the job, so no line
<jsdldatabase:driverPath>/opt/wa/TWS/jdbcdrivers/db2/</jsdldatabase:driverPath>
===============================================================
= Exit Status : 0
Note that in this case the jdbcdrivers/db2 is not needed. It will search for subdirectories.
I have one application that has preset of the app server installed.
I just want that be copied as an image to GCS.
I tried 'gcimagebundle' that said 'platform is not found, you can configure the platform in the python_factory.py file'.
Is there any way that I can copy my snapshot to GCS?
More Information:
sudo gcimagebundle -d /dev/sda -o /tmp/ --log_file=/tmp/abc.log
***************************************************************************
* *
* !!!! WARNING: YOU ARE INSIDE LVE !!!! *
*IF YOU RESTART ANY SERVICES STABILITY OF YOUR SYSTEM WILL BE COMPROMIZED *
* CHANGE YOUR USER'S GROUP TO wheel to safely use SU/SUDO *
* MORE INFO: *
* http://docs.cloudlinux.com/index.html?lve_pam_module.html *
* *
***************************************************************************
Starting logging in /tmp/abc.log
Platform is not supported. Platform rules can be added to platform_factory.py.
os which i used is CentOS 6
n1-standard-1 (1 vCPU, 3.8 GB memory)
us-central1-b
centos-6-v20140415
There is a recently added capability in Compute Engine that allows you to create an image directly from a persistent disk. Please see the docs at https://developers.google.com/compute/docs/images#creating_an_image_from_a_root_persistent_disk
If you only have the snapshot, you would have to create a persistent disk from your snapshot first, and then use the 'gcloud compute images create' command with that disk as the source-disk
Last night, at home, where I do not work with a proxy, updated a NuGet repository using NuGetPowerTools / Enable-PackageRestore.
Worked great.
Came into the office, where i am working with the same Repository under the following conditions:
in a VMWare based VM,
using Bridged Networking,
Reinstalled NuGet today, restarted (to be sure I have latest version)
runnning under an account in a DEV Domain (ie, DEV\Me)
Accessing the outer world via a Proxy that requires a corp Domain Account (ie, CORP\Me)
Ie 9's Configuration/Internet Options/Connection is setup as:
Use Proxy:
Address: yadayada1
Port: 80
Bypass proxy for local addresses
Automatically detect Settings
Above settings are correct in as much that I can access the web via IE9, Chrome, etc.
NOTE: Can list and download Nuget packages...it's just the new Build process that can't.
As per suggestion on the web I have looked at DevEnv.exe.config and have the following settings:
-<system.net><settings><ipv6 enabled="true" /></settings></system.net>
I also tried with these settings as
<system.net><defaultProxy useDefaultCredentials="true"><proxy autoDetect="True" usesystemdefault="True" /></defaultProxy></system.net>
so that DevEnv.exe falls back for to same connection settings that IE 9 is using. Didn't do it.
What I get for all this is:
------ Build started: Project: XAct.Core, Configuration: Debug Any CPU ------
Please provide proxy credentials:
UserName: Password:
C:_Workspaces\XAct\CS.FF.XAct.Lib2.nuget\NuGet.targets(43,9): error : Cannot read keys > when either application does not have a console or when console input has been redirected > from a file. Try Console.Read.
C:_Workspaces\XAct\CS.FF.XAct.Lib2.nuget\NuGet.targets(43,9): error MSB3073: The
command ""C:_Workspaces\XAct\CS.FF.XAct.Lib2.nuget\nuget.exe" install
"C:_Workspaces\XAct\CS.FF.XAct.Lib2\XAct.Core\XAct.Core\packages.config" -source "" -o
"C:_Workspaces\XAct\CS.FF.XAct.Lib2\packages"" exited with code 1.
Found (but unfortunately lost again) a thread somewhere on the net where I saw NuGet developers thrashing it out -- referring to downloading the latest build from their TeamCity (which I could not access/join) -- but the thread did not clearly say Eureka! at the bottom.
Thanks.
Reiterating what Sky and Pranav mentioned - this issue seems to be fixed in NuGet 1.7 - see http://nuget.codeplex.com/workitem/1655:
pranavkm wrote Jan 31 at 8:17 PM
Fixed in changeset 6c156e0dc214