Selecting appropriate tool for replacement of IBM DataStage ETL tool - datastage

we are looking for replacement of existing IBM DataStage platform . it has had around 1500 + mappings/datastage jobs on-premise . these mapping have also some complex transformations and mappings.
It is a complete ETL architecture on-premise. If it needs to get replace (Datastage) with microsoft platform (SSIS or Azure data factory) , what are the options to get replace with SSIS ETL tool .
Option:
if SSIS ETL tool selected , it needs to re-write all datastage scrips (with etl transformation) to SSIS
and then SSIS packege can be run through Azure Data factory. this will also incur new license cost for SSIS.
Can we re-write all ETL datastage jobs/scripts in Azure Data factory , would it be recommended as
Azure Data factory is orchestration tool and mainly used for Data driven movement. Azure data factory is not a traditional ETL tool
if anyone can through some lights on DataStage ETL architecture and its rich in build transformation.
and advised which would best option for datastage replacement.
Regards,
mangesh
mangesh7632#gmail.com

Sorry to hear you're looking for a DataStage replacement. If you're still interested, come take a look at what we're building in our new DataStage SaaS product. Here, you'll get a true subscription pay-as-you-go model (if pricing was the issue)
https://developer.ibm.com/tutorials/getting-started-using-the-new-ibm-datastage-saas-beta-service/
Happy to chat with you. Let me know if you'd like me to email you.
Best wishes with your data integration endeavors.
Kevin Wei - Product Manager, IBM

Related

How to establish the connection to SAP Solution Manager using Azure data factory

I have a requirement to extract the data SAP solution manager to azure blob using ADF as ETL tool.
I do not find any direct connector to create a linked services on source(SAP SOLUTION MANAGER).
Please advice me how to create linked services to SAP solution manager using ADF. Appreciate your effort.
Regards,
Rajesh K
Here is a list of supported data stores and formats in Azure Data Factory
How to create linked services to SAP solution manager using ADF
You can follow this official documentation to Create a linked service to an SAP table using UI

ArrayOutOfBoundsException on Deployment on Watson Studio

I am working on a predictive on IBM SPSS modeler with an SVM and then uploading the PMML to the cloud to provide it as a service, but even when I use the automated system to send the request it gives me ArrayOutOfBounds exceptions. Please let me know if you need more details

Which talend product to download?

Which talend tool should i download . I need to use enterprise technologies like REST/SOAP/JSON/XML/JDBC/SFTP/XSD (Oracle).
My use cases are :
Exposing services(REST/SOAP)
Reading from file like MT940,CSV,flat files and storing in database (Oracle)
Using SFTP ,File Movements frequently.
What is the difference between TALEND Data integration and Talend ESB.
Currently i have download Talend Big Data Open studio .
Will this suffice?
It depends on what you are wanting to do with it. I'm thinking Talend Data Integrator, the free version should suffice, if all you want to do is pull data.

Automated way to see Queries in all Oracle Connectors of all jobs in datastage

Is there a way to see all the queries that are there in my Oracle COnnector stages of my datastage project? I am using DS 11.3.
No not natively. You could export your project and parse the export for all of the SQL staments (this could be done by a DataStage job of cause) or you might be able to query it if you have IGC Information Governance Catalog) in place.

Azure - SSMS - PowerShell

I am working through my first Azure HDInsight tutorial. Can I do this without installing Azure Remote PowerShell on my local computer?
Can I use SSMS (2008R2) to run the PowerShell? My first attempt at that led me down the path of using a Database in Azure, but I do not think that is what I want to do (the tutorial describes setting Storage (not a Database) and then an HDInsight instance to interact with that Storage).
I am doing this tutorial: http://www.windowsazure.com/en-us/manage/services/hdinsight/get-started-hdinsight/
Thank you.
While you can use SQL Server and HDInsight together as part of a full pipeline, for the purposes of the getting started tutorial you want to think of them as two very different things.
The Storage referred to, is a standard Windows Azure Storage account, based on blobs. These then form a backing file system for the HDInsight cluster.
As far as using PowerShell goes, it is definitely the best, and easiest way to submit jobs to an HDInsight cluster. I would also recommend using a regular PowerShell console, or the PowerShell ISE to work with HDInsight as well, rather than the one available through SSMS, since the SSMS version won't load all the Azure modules by default.
There are other ways to submit jobs if PowerShell is not your thing (if you are on OS X or Linux for instance). You can use the REST API provided by WebHCAT (documentation). If you're on Windows, and prefer C# to PowerShell, you can also use the Windows Azure HDInsight Management Client from the Microsoft Hadoop SDK to submit jobs (available on codeplex and nuget). These will need you to break out Visual Studio and write a short console program to submit your job, so may be a bit heavy unless you're doing full on C# streaming Map Reduce, and so are already there.
If you're after a GUI based approach to job submission to HDInsight, you're out of luck at the moment, but your might like to check out what my team is working on at Red Gate, which will help you with submitting Hive and Pig jobs.