google-cloud-dataprep: Importing Excel from GCS bucket gives gibberish characters - google-cloud-dataprep

I am trying to import an excel spreadsheet from a GCS bucket as a dataset into a flow.
But it's giving me gibberish data in the preview and also when I try to wrangle the data.
The same spreadsheet works fine when I upload from my computer.
Any idea how I can fix the issue when loading form the GCS bucket ?

Related

How to retrieve some information without using file

I'm searching a way to retrieve data (csv format) from something which is not a file, let me explain.
In the company i'm in actually they quit Fileserver to Google Drive - So i'm actually trying to read information from a google spreadsheet but it didn't work nicely so far.
Any idea how to do this ? Any other idea on a way to centralise datas (CSV format) which is neither fileserver or using GDrive ?
The data source should be easily modified (no database..)
Thanks.

tableau is using unidentified source

Hi I am using extract to view a workbook in tableau server.
I deleted the local extract as well as the extract in the server.
I am still able to see that data is being pointed.
Struggled a lot to understand this.Does tableau create temp folder to store?
FYI It is not in live mode(since the data is not being changed until I refresh).
How do I point to server extract?(is it automatic)?
If a workbook on Server is pointing to a server hosted data source and that data source is delete, then it would not render. If you're still able to view the data, then it sounds like the workbook was uploaded with the extract. Check the file size, if it's large, that's probably what happened.

How do I upload files to S3 instead of my hard drive using PyroCMS 2.1.5?

I can't seem to figure out how to post files (images, etc) to amazon S3.
I'd like for this to be default behavior for ALL media upload areas.
Jerel Unruh put together a video all about this:
http://www.youtube.com/watch?v=Te61OzHK400
Really you just enter your Amazon credentials into the settings area, then hook a folder up with the bucket. All images will be sent to S3 and cached locally, and work identically to local files.

Google Drive: Automatically convert files on upload?

Is it possible to get Google Drive to automatically convert uploaded documents to the native format?
I know it works with manual upload (i.e. Google Drive can auto-convert files you upload via the website), but I want to avoid having to upload every file by hand.
I'd prefer to use the API, or better yet, dump the files in my ~/Google Drive folder.
Using the API, you can pass the convert=true parameter to files.insert. The uploaded file will attempt to be converted to a native Google Docs format.
Sure, see this answer. Note you can upload a text file or a csv file and set its content type to google doc or google sheets respectively, and google will attempt to convert it. I have tested this for text -> doc and it works. You will need to set the const contentType at the start of the code to one of the supported google mimetypes.

Choosing right tool

I have following need:
1) Users will upload .xls or .csv files in "uploads" folder.
2) "uploads" folder have to be constantly monitored, and with each new file added to him, a job has to be started.
3) Job will process data from .xls or .csv file so they meet DB table structure, and write this data into DB table.
This have to be automated process, and I'm looking for all-in-one solution tool.
You didn't tell on which operating system, and you didn't tell if the user upload the files on a different server, or not. If the upload goes thru a web application (using an HTTP POST request), it is also different.
And I'm not sure that your wish scales well with many users.
You should take a look at Pentaho Data Integration, a.k.a. Kettle: http://sourceforge.net/projects/pentaho/
With Kettle you can desing a Job that pools the upload directory and once a file is found makes all the needed transformation and input on the desired database table.