Tableau Auto-refresh local on-premises Database/files on Mac - tableau-api

I wonder anyone has succeeded in refreshed publishing workbook from local on-premises database/files on Mac? I am using Tableau Online and I made Tabcmd working on mac. But seems I need Tableau Data Extract Command-line Utility to do the job.
I do not want to use VMs or Windows machines as it may cause trouble for my database management, which I built with OS X. I also tried Wine/Crossover to run Tableau.exe on Mac, which did not work.
Any better suggestions? Please kindly help. Thanks!

It's not clear exactly what you are trying to do.
If you want to create a Tableau data extract programmatically, you can write a script or program that calls the Tableau Data Extract API.
If you want to upload something to Tableau server, you can write a script or program that calls the Tableau REST API

Related

Import Data from IBM Connection Docs Excel URL in Tableau

My company recently transitioned to IBM Connections and has closed down a few features like Box.com.
IBM Connections can also store files, but when uploading them, I cannot connect Tableau to them directly as the file path is a URL and Tableau cannot get files from URLs.
I've tried to load it in Power Query and then load it in Tableau but then I need to find a way to refresh the file while the original Excel file gets edited.
Any suggestions how can I get a live connection from IBM Connections Docs > Tableau Server?
I don't believe a live connection is possible. However, it looks like IBM provides a robust api. Depending on your requirements, you could script something that writes out a json file. Then use tableau to refresh the json as an extract on a schedule. Another possibility is the Web Data Connector.

Tableau extracts update

I would like to know if there's a way to automatically update my extracts?
I have a live connection with redshift and use tableau desktop and publish some workbooks on tableau online.
I like to share some reports using extracts and tableau reader and I really need a way to update the extracts everyday
if I understand your question correctly, you should be able to refresh your Tableau Data Extracts by setting up "Refresh Schedules" on Tableau Online. Tableau Online supports variety of Data Sources to set up Extracts instead of using Live Connection.
Check this link on setting up refresh schedules.
Hope this helps!
Yes, you can do set refresh schedule that will help you to update your extract automatically without any manual effort but there are some terms & conditions you have to follow:
For this you have to set your dashboards/workbooks/project in "Extract" mode & not in LIVE.
I personally recommend "extracts" instead of LIVE until the person who is looking for this need REAL TIME updates
Also, you need to set this schedule from Admin's ID (Publisher access) which you use to publish your work from Desktop to Online
Note: You only can choose from lot of options already available
enter image description here

Creating automated Tableau packaged workbooks (for Tableau Reader) using Alteryx TDE output

I am running an Alteryx workflow on a dedicated PC, 4X a day. The Alteryx workflow creates a TDE output file that is the data source for a dashboard I would like to save as a packaged workbook (openable in Tableau Reader) and distribute 4X a day.
I have a Tableau online account and tried publishing an extract of the TDE file created by Alteryx to Tableau online, setting up a refresh schedule for that extract using Tableau online sync, and then linking my workbook to this Tableau online data source.
While that solution has worked well in terms of having 4X a day updated Tableau workbooks online, when I try to download the TWBX file, it does not contain the data extract (as the Tableau workbook has a "live" connection to the published Tableau data extract online) and thus cannot be opened in Tableau reader.
My main goal is to be able to automate production of Tableau packaged worbooks (a macro that opens Tableau and hits save would do the job), and would appreciate any creative suggestions / ideas.
This can't be accomplished. You would need to leverage on-premise Tableau Server.
FWIW...Although many companies use Reader (due to the cost of server), it was never intended to be used in a production environment. Data is not governed or secure.

Which kind of Google Cloud Platform mobile backend client is appropriate?

THE PROBLEM
I'm writing a mobile app which will allow a user to log in, save some preferences that must be stored in a database, and display congressional bills to the user.
I've only written simple RESTful services with PHP and MySQL in the past. I'd like to take advantage of newer technologies, and am a little lost on general direction.
The bill data (formatted as JSON) can be gathered by running the scrapers found here. Using docker, I managed to set a working directory and download the files on my local machine.
I've designed a MySQL database for holding the relevant bill and user data.
I started to mess around in Google Cloud Platform, and read the doc that describes different models. I'm thinking of a few different ideas, but aren't familiar with GCP or what I can actually accomplish.
QUESTIONS
1) What are App Engine, Compute Engine, and Container Engine each for? I get the gist that Container Engine holds different instances of stuff you load up with docker, and that Compute Engine sets up a VM, but I don't really understand the relationships. How should I think of them?
2) When I run those scrapers from the shell, where are the files being stored, and how can I check on them? On my computer, I set a working directory, but how do directories work in GCP? Is it just a directory in the currently selected VM, or is this what Buckets are for?
IDEAS
1) Since my bill data already comes as JSON, should I skip the entire process of building a database for the bills and insert them into Firebase somehow? Is this even possible? If so, am I stuck using Firebase's NoSQL, or can I still set up a relational database?
2) I could schedule the scrapers to run periodically, detect new files, and run a script to parse the JSON and insert new bill data into my a database (PostgrSQL?/MySQL?). Then I would write an API.
3) Download the JSON files to a bucket, and write an API that reads from them. Not sure how the performance would compare to using a DB.
I'm open to other suggestions as well.
For your use case (stateless web application), App Engine is probably your best choice. The Google documentation has severalcomparisons of your computing options
You can use App Engine with PHP and cloud-hosted MySQL if you want, which could be a good way to get your toes wet without going in over your head.

automating downloads of facebook insight data

I'm looking for a tool or process for exporting facebook insights data for a facebook page and a facebook app. Currently I am just manually downloading csv files from their Insights interface but ideally I want to automate this process and load the data into Pentaho Kettle, so I can perform some operations on the data.
Is there some way to automate the downloading and input of csv files? Or will I have to use the facebook graph api explorer? I am currently looking at a set-up where I use NetBeans and RestFB to pull the data I want, and then access that data using Pentaho Kettle. I am not sure if this will work, or if it is the best approach.
As Codek says, a Kettle plugin is a very good idea, and would be very helpful to the Kettle project. However, it's also a serious effort.
If you don't want to put in that kind of effort, you can certainly download files with a Kettle Job as long as the files are available through a standard transfer method (FTP, SFTP, SSH, etc). I've never used RestFB, so I don't know what's available. You might be able to get directly from a web service with the REST Client transform step.
After downloading the files, you can send them to a transform to be loaded. You can do this with either the Execute for every input row? option on the Transformation job step, or you can get the filenames from the job's result set in the transform with Get files from result.
Then you can archive the files after loading with Copy or Move result filenames. In one job, I find only files that are not in my archive using a Get File Names and Merge Join, and then a Set files in result step in a transform, so that can be done if need too.
To automate it, you can run your job from a scheduler using Kitchen.bat/Kitchen.sh. Since I use PostgreSQL a lot, I use PGAgent as my scheduler, but the Windows scheduler or cron work too.
Hope that helps.