Google DataFusion : HUB creation - google-cloud-data-fusion

How to create and host private HUB under DataFusion and how to configure DataFusion make reference to such private HUB.

#alex:
CDAP 6.2 (release notes) added the ability to support multiple hubs.
You would need to configure the cdap-site property (market.base.urls) with list of hubs (string with + as delimiter) and restart CDAP. This property will be passed on to the UI which should surface the multiple hubs.
The design doc for this feature is available here for more information. We are working on documentation and this should soon be available via CDAP docs.

Related

How to use Azure Data Factory, Key Vaults and ADF Private Endpoints together

I've created new ADF instance on Azure with Managed Virtual Network integration enabled.
I planned to connect to Azure Key Vault to retrieve credentials for my pipeline’s source and sink systems using Key Vault Private Endpoint. I was able to successfully create it using Azure Data Factory Studio. I have also created Azure Key Vault linked service.
However, when I try to configure another Linked Services for source and destination systems the only option available for retrieving credentials from Key Vault is AVK Linked Service. I'm not able to select related Private Endpoint anywhere (please see below screen).
Do I miss something?
Are there any additional configuration steps required? Is the scenario I've described possible at all?
Any help will be appreciated!
UPDATE: Screen comparing 2 linked services (one with managed network and private endpoint selected and another one where I'm not able to set this options up):
Managed Virtual Network integration enabled, Make sure check which region you are using unfortunately ADF managed virtual network is not supported for East Asia.
I have tried in my environment even that option is not available
So, I have gathered some information even if you create a private endpoint for Key Vault, this column is always shown as blank .it validates URL format but doesn't do any network operation
As per official document if you want to use new link service, instead of key vault try to create other database services like azure sql, azure synapse service like as below
For your Reference:
Store credentials in Azure Key Vault - Azure Data Factory | Microsoft Docs
Azure Data Factory and Key Vault - Tech Talk Corner

How to run a preview on private instance?

Our pipeline fetch data from internet.
The preview mode doesn't work on my private cloud data fusion instance, I have a timeout each time.
The same jobs work when deployed.
Note I am obliged to have a private instance.
How can I get a preview that connects to the internet?
To successfully run preview on a private instance, you need to setup VPC peering to the tenant project. For more documentation on how to do this, you can see the documentation: https://cloud.google.com/data-fusion/docs/how-to/create-private-ip#set_up_network_peering

Why am I seeing a "Reseller channel 2c95500b-ea86-4b13-8bb5-b2f0c2fa8200 is invalid" error when attempting to create storage

I am attempting to create Cloud Object Storage to work in conjunction with a Watson Studio service instance. However, when I attempt to create it, the following error message is displayed:
“Reseller channel 2c95500b-ea86-4b13-8bb5-b2f0c2fa8200 is invalid.”
How can I create the storage that I need?
When you attempt to create the cloud object storage instance through the Watson Studio service, a link sends you to a storage provisioning page. However, you are unable to provision the storage instance. As a workaround, you can create the Cloud Object Storage service manually through the IBM Cloud Catalog at the following URL: https://cloud.ibm.com/catalog/services/cloud-object-storage
I had this issue as well. Turns out when you create the new account it will ask you for a location for services and by default it's set to "Dallas". However changing that to London/Tokyo fixed this issue for me. I needed it for IBM Course on Coursera. I would recommend making a new account with services location set to London / Tokyo. It'll instantly fix this problem.
I managed to solve this by first creating the object storage before creating the Watson Studio service.
After you have created a cloud object storage service with the direct link to the service
While creating a new project On Watson Studio, you will see an option to Add a cloud Object Storage service under Define storage to your project in the project creation page. Once you click Add, you can either use an existing Service as you have now or create a new COS service.
For more details instructions check this section of the solution tutorial

Can we use different run-time in Azure Data Factory v2 (ADFv2) Copy Activity?

I have copy activity, where source is Oracle on premises connected through Self-hosted IR and destination is Microsoft Synapse connected via Azure Run-time. These run-time is defined in connections (Linked Services).
But while execution pipeline is using Self Hosted Run-time through-out and overriding the run-time of Azure Synapse. And because of that connection is failing.
Is this default behavior? Can't I run pipeline with 2 different run-time.
Thanks #wBob but I am sorry that is not true, I found the answer at Microsoft documentation.
Copying between a cloud data source and a data source in private network: if either source or sink linked service points to a self-hosted IR, the copy activity is executed on that self-hosted Integration Runtime.
Ref: https://learn.microsoft.com/en-us/azure/data-factory/concepts-integration-runtime#determining-which-ir-to-use
Integration runtimes are defined at the linked service level. So you should have a linked service definition for your Oracle database and a separate linked service definition for your Azure Synapse Analytics (formerly known as Azure SQL Data Warehouse). So you can specify different integration runtimes, eg here's an example:
NB Azure Synapse Analytics is using the Autoresolve runtime and does not need a self-hosted integration runtime (SHIR) as it is a native PaaS service. Are you getting a specific error? If so, please post details.

Bluemix: Are devops services available on Bluemix local?

Does the Bluemix local provide devops services like Delivery Pipeline and Active Deploy?
Bluemix Local includes a private syndicated catalog that displays the local services that are available exclusively to you. It also includes additional services that are made available to you to use from Bluemix Public. The syndicated catalog provides the function to create hybrid applications that consist of public and private services.
Bluemix Local comes with all included Bluemix runtimes and a set of services and components available. Take a look at the Table 1. Local Services in Bluemix Local Docs.
As you can see, for example the Auto-Scaling service is already included in the local environment. However you have the option to decide which public services meet the requirements for your business based on your data privacy and security criteria.