KingswaySoft Google Cloud Storage Source component missing parameter - google-cloud-storage

I cant find a solution on how to provide the missing parameter in KingswaySoft Google Cloud Storage Component Editor. Anyone
KingswaySoft Google Cloud Storage Component Editor
KingswaySoft.IntegrationToolkit.GoogleCloudStorage.GoogleCloudStorageServiceException: Non-negative number required.
Parameter name: count (SSIS Productivity Pack, v21.1.0.1551 - DtsDebugHost, v15.0.2000.170)System.ArgumentOutOfRangeException

Related

Why am I seeing a "Reseller channel 2c95500b-ea86-4b13-8bb5-b2f0c2fa8200 is invalid" error when attempting to create storage

I am attempting to create Cloud Object Storage to work in conjunction with a Watson Studio service instance. However, when I attempt to create it, the following error message is displayed:
“Reseller channel 2c95500b-ea86-4b13-8bb5-b2f0c2fa8200 is invalid.”
How can I create the storage that I need?
When you attempt to create the cloud object storage instance through the Watson Studio service, a link sends you to a storage provisioning page. However, you are unable to provision the storage instance. As a workaround, you can create the Cloud Object Storage service manually through the IBM Cloud Catalog at the following URL: https://cloud.ibm.com/catalog/services/cloud-object-storage
I had this issue as well. Turns out when you create the new account it will ask you for a location for services and by default it's set to "Dallas". However changing that to London/Tokyo fixed this issue for me. I needed it for IBM Course on Coursera. I would recommend making a new account with services location set to London / Tokyo. It'll instantly fix this problem.
I managed to solve this by first creating the object storage before creating the Watson Studio service.
After you have created a cloud object storage service with the direct link to the service
While creating a new project On Watson Studio, you will see an option to Add a cloud Object Storage service under Define storage to your project in the project creation page. Once you click Add, you can either use an existing Service as you have now or create a new COS service.
For more details instructions check this section of the solution tutorial

Use only a domain and disable https://storage.googleapis.com url access

I am newbie at cloud servers and I've opened a google cloud storage to host image files. I've verified my domain and configured it, to view images via my domain. The problem is, same file is both accessible via my domain example.com/images/tiny.png and also via storage.googleapis.com/example.com/images/tiny.png Is there any solution to disable access via storage.googleapis.com and use only my domain?
Google Cloud Platform Support Version:
NOTE: This is the reply from Google Cloud Platform Support when contacted via email...
I understand that you have set up a domain name for one of your Cloud Storage buckets and you want to make sure only URLs starting with your domain name have access to this bucket.
I am afraid that this is not possible because of how Cloud Storage permission works.
Making a Cloud Storage bucket publicly readable also gives each of its files a public link. And currently this public link can’t be disabled.
A workaround would be implement a proxy program and running it on a Compute Engine virtual machine. This VM will need a static external IP so that you can map your domain to it. The proxy program will be in charged of returning the requested file from a predefined Cloud Storage bucket while the bucket keeps to be inaccessible to the public.
You may find these documents helpful if you are interested in this workaround:
1. Quick start to set up a Linux VM (1).
2. Python API for accessing Cloud Storage files (2).
3. How to download service account keys to grant a program access to a set of services (3).
4. Pricing calculator for getting a picture on how much a VM may cost (4).
(1) https://cloud.google.com/compute/docs/quickstart-linux
(2) https://pypi.org/project/google-cloud-storage/
(3) https://cloud.google.com/iam/docs/creating-managing-service-account-keys
(4) https://cloud.google.com/products/calculator/
My Version:
It seems the solution to this question is really a simple, just FUSE Google Cloud Storage with VM Instance.
After FUSE private files from GCS can be accessed through VM's IP address. It made Google Cloud Storage Bucket act like a directory.
The detailed documentation about how to setup FUSE in Google Cloud is here.
There is but it requires you to do more work.
Your current solution works because you've made access to the GCS bucket (example.com), public and then you're DNS aliasing from your domain.
An alternative approach would be for you to limit access to the GCS bucket to one (possibly several) accounts and then run a web-server that uses one of the accounts to access your image files. You could then also either permit access to your web-server to anyone or also limit access to it.
More work for you (and possibly cost) but more control.

How to take Backup of the Object Storage created in IBM Bluemix (IBM Cloud)?

I have a Cloud Object storage created from bluemix (IBM Cloud) console. Now for some reason, I need to backup all the data present in the existing object storage into another object storage.
I'm not able to find anything proper on this. Any help on this would be much appreciated.
You can use a tool like rclone or s3cmd to sync the Cloud Object Storage buckets to a linux server and then use those same tools to send them to the new Cloud Object Storage account.
https://mpcarl.github.io/rclone-cos
https://knowledgelayer.softlayer.com/procedure/connecting-cos-s3-using-s3cmd

Connecting to Object Storage on Node-Red

I am trying to retrieve an image from my Object Storage service using Node-Red.
I imported the Object Storage nodes in my Node-Red instance running on IBM Bluemix and I am trying to use the "get" node but I cannot figure out how to correctly connect it to the Object Storage service.
I set the option to get the configuration details from the Bluemix Service, but when deploying I would always get an error saying "Object Storage Put (err): No object storage configuration found!".
If instead I use the option for API based configuration information, I need to fill out fields which I cannot find on my VCAP_SERVICES or the Object Storage service credentials. The fields that need to be filled in are shown
but my service credentials only include:
"apikey"
"endpoints"
"iam_apikey_description"
"iam_apikey_name"
"iam_role_crn"
"iam_serviceid_crn"
"resource_instance_id"
Can anyone point me in the right direction?
The node-red object storage contribution available in this link:
https://flows.nodered.org/node/node-red-contrib-objectstore
was designed to be integrated with the Swift Object Storage in Bluemix (IBM Cloud Platform).
If you go the Catalog link below:
https://console.bluemix.net/catalog/infrastructure/object-storage-group
there are 4 options, the one you previously selected and is trying to connect is probably the first on the left side and will not integrate with node-red-contrib-objectstorage. T
To integrate with node-red you need to create an instance of the Object Storage service on the far right: Object Storage OpenStack Swift for Bluemix

Authorizing GCE to Access GCS

I have a django app running in my Google Compute Engine, and it needs to upload video files to my bucket in Google Cloud Storage. When searching for authentication methods, I found this doc. Under Setting the scope of service account access for instances section, it says I need to enable the Cloud Platform access in the settings when creating the VM. I wonder if it is a must and if there's any other way that I can access my cloud storage bucket from my apps in the compute engine. Because creating a new VM and set up the environment is very time-consuming. Any input would be greatly appreciated. Thanks in advance.
As documented on the page you linked to, to authenticate from Google Compute Engine to Google Cloud Storage, you have several options:
Use VM scopes: this must be set before creating the VM, because scopes are immutable once the VM is created. If you want read-only access, you need to add the scope devstorage.read_only (short form) or https://www.googleapis.com/auth/devstorage.read_only (full path). If you want read-write access, you should use the scope devstorage.read_write (short form) or https://www.googleapis.com/auth/devstorage.read_write (full path).
Note: there's also a feature gcloud beta compute instances set-scopes to update GCE VM scopes at runtime.
An alternative to using scopes is to use JSON authentication tokens, such as via Service accounts which can be used by Google API client libraries to connect to Google Cloud Storage.