How can I keep Watson Discovery’s collection synchronized with a custom list on SharePoint using the data crawler?
Is there a simple way to automatically update Watson’s data every time new data is added to SharePoint?
Related
My company has lots of data(Database: PostgreSQL) and now the requirement is to add search feature in that,we have been asked to use Azure cognitive search.
I want to know that how we can transform the data and send it to the Azure search engine.
There are few cases which we have to handle:
How will we transfer and upload on index of search engine for existing data?
What will be the easy way to update the data on search engine with new records in our Production Database?(For now we are using Java back end code for transforming the data and updating the index, but it is very time consuming.)
3.What will be the best way to manage when there's an update on existing database structure? How will we update the indexer without doing lots of work by creating the indexers every time?
Is there anyway we can automatically update the index whenever there is change in database records.
You can either write code to push data from your PostgreSQL database into the Azure Search index via the /docs/index API, or you can configure an Azure Search Indexer to do the data ingestion. The upside of configuring an Indexer to do the ingestion is that you can also configure it to monitor the datasource on a schedule for updates, and have those updates reflected into the search index automatically. For example via SQL Integrated Change Tracking Policy
PostgreSQL is a supported datasource for Azure Search Indexers, although the datasource is in preview (not get generally available).
Besides the answer above that involves coding on your end, there is a solution you may implement using Azure Data Factory PostgreSQL connector with a custom query that tracks for recent records and create a Pipeline Activity that sinks to an Azure Blob Storage account.
Then within Data Factory you can link to a Pipeline Activity that copies to an Azure Cognitive Search index and add a trigger to the pipeline to run at specified times.
Once the staged data is in the storage account in delimitedText format, you can also use built-in Azure Blob indexer with change tracking enabled.
I have few custom modules created using Module Builder. I need to fetch data using external REST API (around 200-300k record per API) and without storing data in suite database I need to display data on custom modules (using UI provided by SuiteCRM only. NO CUSTOM UI). and also I have operations to be perform on data which will again call External API.
So, as per requirement, Data should not be saved at database. How to achieve this in SUITECRM (7.11.8 version)?
I have tried using Process record hook. But things aren't working.
I am a newbie to Sitecore and I want to ask if it possible to get data from sitecore via. some REST API or if I can intercept when form data changes and save it to a json file.
There are a few ways to do this depending on what version of Sitecore you are using, but the most robust at the moment is to use the Sitecore Services Client Framework:
https://sdn.sitecore.net/upload/sitecore7/75/developer%27s_guide_to_sitecore.services.client_sc75-a4.pdf
You can use the ItemService REST api to get data out of Sitecore in json format of if you need more control over the data you can use the EntityService. This is specific to Sitecore 7.5 and above.
If you're in an earlier version then you have the Item Web api:
https://sdn.sitecore.net/upload/sdn5/modules/sitecore%20item%20web%20api/sitecore_item_web_api_developer_guide_sc65-66-usletter.pdf
If you planning on intercepting data on Web Forms for Marketers forms then you're best off using a Custom Save Action to save form data to a json file:
http://www.roundedcube.com/Blog/2016/creating-a-custom-save-action-for-sitecore-wffm
I am installed and added data source (web data connector) inside tableau
Web data connector URL : http://localhost/datasouceexample/example.html
we data connector holds some data.
import data from web data connector and created graphs using the data's from the web data connector URL .Generated graph and saved in "tableau public" option.
Embed the code inside my web application . Graphs shows perfectly.
My question is about the automatic data updation .
1 ) I want to update graph automatically based on the value from web data connector URL
2 ) If any new updations are inside input datas (new inputs) , automatically sync with tableau and update the graph,so no changes in the embed code and graph updated automatically
Any settings available inside tableau for do this section? if it is possible ? Thanks
Only dashboards with a google sheets datasource can be automatically updated on Tableau Public. Maybe you could redesign your web service to write to a google sheets doc. If so, you could republish your dashboard to sync with that doc.
The scenario you are describing is not possible using Tableau Public (the free version of Tableau software) and the web data connector. Tableau Public does not support updating web data connector data once published on Public, you would need to refresh the data from your desktop application and republish.
The closest supported path to your use case would be, as Kara_F mentioned in the comments, to use Tableau Server. With that product, you can create web data connector datasources and publish them to your Server. You can then schedule a refresh job which will update your data at a certain interval. More info for keeping WDC data up to date can be found here: http://onlinehelp.tableau.com/v0.0/server/en-us/datasource_wdc.htm.
I am new to SugarCRM, i have created SugarCRM's web to lead form on my website which is having four fields i.e. frist name, last name, address and phone . On submit this info get updated in SugarCRM in Leadlist. Now the requirement is whenever any update happens on SugarCRM, I have to update other data source, in this case marklogic. I want to keep same data on sugar and marklogic.
Whenever any update triggers the SugarCRM it should send some webservice or other method with updated information to marklogic, so that marklogic data should be in sync.
How can I do that? Any suggestion would be a great help.
Thanks.
Check out the new REST interface in MarkLogic 6, specifically the CRUD APIs. Using these you should be able to pretty easily make webservice updates to the database without writing any xQuery.
http://developer.marklogic.com/try/rest/index