As I want to set up a BAM solution, I'm wondering if there's any Open Source BAM solution that accepts REST requests as events and provide a REST interface to query for events ?
I found a BAM Open Source solution created by Jos Dirksen author of the SOA Governance in Action but I couldn't get the maven dependencies from http://mvn.riptano.com/content/repositories/public/ repository (ref: mvn.riptano.com not authorized connection, connection require authentication).
In [1] you can find how to publish data to WSO2 BAM 2.4.0. There, first you should create a stream (which is a predefined set of fields to be published) using REST API. Then you can send data to BAM via REST API. All the data will be saved in Cassandra where you can analyze data in it using a Hadoop cluster and summarize data. In most cases summarized data is stored in RDB where these data can be visualizes in a dashboard or generate a report.
[1] http://docs.wso2.org/display/BAM240/Sending+Events+through+the+REST+API
Related
I am working on creating a flow where I get JSON data from a rest api authenticating with the managed identity of my ADF instance and copy the data to a Kusto cluster. To do this, I am following the instructions here: https://learn.microsoft.com/en-us/azure/data-factory/connector-http?tabs=data-factory
However, when I go to create a linked service I do not see any way to authenticate against the API using my managed identity: http linked service auth options
I was expecting something like the options given in the Web task in ADF, where I am allowed to selected managed identities:web task auth options
HTTP connector supports only Anonymous, Basic, Digest, Windows, and ClientCertificate as authentication types. To use Managed Identity authentication type, you can use Rest linked service.
Search for REST in the available list of linked service and select REST connector.
You can select Managed Identity as auth type.
Reference: MS doc on Data Fcatory - REST connector.
We have Scheduler Jobs transfer data from Source to Destination (Employee Data to Systems or Filemovers (Configures in property files))
We have Rest API exposes data related to entity (For example Employee or Organisation)
We have Wildfly on Linux (cluser), Oracle and Java Integrations deployed.
Manager ask to provide IAAS i.e Integration as a Service. As we are already exposing Rest API; what more comprises IAAS?
I need to ingest data into my application. There is a Kafka stream already build for data ingestion. One of the client's requirement needs an API interface for accepting the data. So I need an API management solution which provides API gateway functionality and is able to produce message on a Kafka topic.
I have analyzed Kong + Kong-upstream plugin. But looking for any other similar solution
You can check WSO2 Api Manager.
check this article for wso2 kafka integration. https://ei.docs.wso2.com/en/7.2.0/micro-integrator/references/connectors/kafka-connector/kafka-connector-producer-example/
Also you can write your custom mediator for WSO2.
Another Solution;
If you need different solution else that api gateway. You can check Apache Nifi. It's application that you create your flow. You can do anything what you want.
Apache Nifi's Documentation
For your problem you need to check this Apache Nifi's processors;
-PublishKafka + ConsumeKafka
-ListenHTTP
Edit after #OneCricketeer's comment:
Kafka Solution Without any Integration
Kafka REST Proxy
I am using a JBoss based vault to secure sensitive data such as the database credentials.
I use a Java based HTTP REST client to create distributed Kafka connectors but ended up with a security concern such that a request for the connector's "config" exposes the sensitive credentials in the response.
I referred this official documentation but could not get much help in the context of JBoss vault.
Any pointers or references that directly addresses this specific problem is very much appreciated.
Any references to alternate open source (and free to use) Vault based solutions would also be of great help.
You'd have to write code that implements the ConfigProvider interface of the Connect API, mentioned there.
You can browse Kafka source code on Github to see the existing File one, but that KIP (which references Hashicorp Vault) and the source files are the only such documentation for now.
Connect doesn't use JBoss, either, so you'd have to find a way around that
I have a REST service that I call from my client app. The service use JSON to manage data.
I want insert SAP NetWeaver Gateway between client and REST server to expose the REST data as oData.
The REST service have get and put methods to read and write data from/to db.
Now I have to decide the way in which starting:
Translate data from/to REST server using ABAP code to serialize data in the two ways (get REST data and create oData in response to a url get oData call and create the REST call in response to a url post oData call)
Use SMP - SAP Mobile Platform (Eclipse plugin) to write integration code (in javascript) to consume REST service in reading and writing exposing this service ad oData. http://scn.sap.com/community/developer-center/mobility-platform/blog/2015/04/08/integration-gateway-rest-data-source-overview-of-blogs
The 2nd solution seems to be the best way (no-require ABAP programming, use js high level language, parse-libs to manage oData and json ...) but i don't know if SMP was created to do this work.
And what is the result of the 2nd method? It seems to be a zip file (similar to a war) that I can push to the SAP Gateway to deploy the integration-logic. Right? How can I test my code without deploy every time the zip on SAP Gateway?
The recommended approach is to create a create Odata service in SAP Gateway and consume it in your app using SMP 3 or HCPms (SAP's mobile solution on cloud).
Use SMP SDK to consume the Odata service.