I created a Datastage Parallel job with Parallel Shared Container. The job was working fine. I did not make any change to the job. Suddenly the job has started failing with below error since 1 day:
main_program: Failed to create collation sequence from IBM InfoSphere
DataStage Enterprise Edition 9.1.0.6791 . Failed to create collation
sequence from IBM InfoSphere DataStage Enterprise Edition 9.1.0.6791.
Has anyone has faced similar issue?
Please help and let me know for further clarification.
The above mentioned error came due to incorrect NLS in the present transformers in the Container. I changed the NLS to ProjectDefault(OFF) and the issue got solved.
Related
When Migrating Cluster from M5 to M10, I am receiving error as show below.
"Configuring analytics nodes specific auto-scaling is not yet supported.".
Has anyone encountered this error? If yes, how to resolve this?
Is there any way to migrate to bigger cluster plan without the "click to update" button?
Update 1 (20 Sept 2022)
Have posted same Q on MongoDb Forum
https://www.mongodb.com/community/forums/t/not-able-to-upgrade-from-m5-to-m10/187868
The issue got resolved.
Note for future visitor to this Question 🙋: This was an internal issue with Atlas. The same was fixed by support team. There was no action required from my part.
In the last few hours I am no longer able to execute deployed Data Fusion pipeline jobs - they just end in an error state almost instantly.
I can run the jobs in Preview mode, but when trying to run deployed jobs this error appears in the logs:
com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Selected software image version '1.2.65-deb9' can no longer be used to create new clusters. Please select a more recent image
I've tried with both an existing instance and a new instance, and all deployed jobs including the sample jobs give this error.
Any ideas? I cannot find any config options for what image is used for execution
We are currently investigating an issue with the image for Cloud Dataproc used by Cloud Data Fusion. We had pinned a version of Dataproc VM image for the launch that is causing an issue.
We apologize for you inconvenience. We are working to resolve the issue as soon as possible for you.
Will provide update on this thread.
Nitin
I created a server job which is supplied from an oracle table
the job is related to a Master. When launching the master, it abort with the following error message "Abnormal termination ", I found a trick to solve the problem, recompling the job before launching the master was the solution.
I'd like to solve this issue, any help will be appreciated
Thank you
Hi i am beginer in Talend Open Studio 5.3.1 version.
currently i am facing issue in project i.e. schedule a job to run every 10 seconds and it monitor the other job and display output as status of another job which means the job is running or idle state.
Currently i am using Talend Open Studio 5.3.1 version by using this version it is possible or not .
explain me how to schelude a job for 10 seconds and display output as status of another job.
can anyone suggest and help me to solve my problem.
We should think a bit out of the box here. I'd solve this by using Project level logging: https://help.talend.com/display/TalendOpenStudioforBigDataUserGuide520EN/2.6+Customizing+project+settings
You'll have the jobs status stored in a database table, you just have to check whether the last execution of the job is still running or not. (Self join the stats table)
Monitoring jobs is not supported in Talend Open Studio, but there is some workaround:
Use a master job that launch the job to be monitored using tRunJob component, and your master job will have an idea whats going on.
Use empty files to synchronize your jobs, an empty file with a tricky name created by monitored jobs and the master job check them to get other jobs states.
Much easier is to use Quartz.
I am trying to configure a 5 node cassandra cluster to run Spark/Shark to test out some Hive queries.
I have installed Spark, Scala, Shark and configured according to Amplab [Running Shark on a cluster] https://github.com/amplab/shark/wiki/Running-Shark-on-a-Cluster.
I am able to get into the Shark CLI and when I try to create an EXTERNAL TABLE out of one of my Cassandra ColumnFamily tables, I keep getting this error
Failed with exception
org.apache.hadoop.hive.ql.metadata.HiveException: Error in loading
storage
handler.org.apache.hadoop.hive.cassandra.CassandraStorageHandler
FAILED: Execution Error, return code 1 from
org.apache.hadoop.hive.ql.exec.DDLTask
I have configured HIVE_HOME, HADOOP_HOME, SCALA_HOME. Perhaps I'm pointing HIVE_HOME and HADOOP_HOME to the wrong paths? HADOOP_HOME is set to my Cassandra hadoop folder (/etc/dse/cassandra), HIVE_HOME is set to the unpacked Amplad download of Hadoop1/hive, and I have also set HIVE_CONF_DIR to my Cassandra Hive path (/etc/dse/hive).
Am I missing any steps? Or have I configured these locations wrongly? Any ideas please? Any help will be very much appreciated. Thanks
Yes, I have got it.
Try https://github.com/2013Commons/hive-cassandra
whick is working with cassandra 2.0.4, hive 0.11, hadoop 2.0