Is it possible to store Audit4j audit events into MongoDB? - mongodb

We want to save audit4j events/logs (which are usually stored in text file) to mongodb.
Is it possible with some existing adapter/plugin? Or need to write on? If we need to write any documentation which can be referred to?
Most of the article talks about auditing mongodb changes itself, hence we are confused.
Any pointer will be appreciated.
Thanks and regards

There's a plugin available on github https://github.com/nipunthathsara/Audit4j-MongoDB

Related

What template was used to create my PostgreSQL database

I have a PostgreSQL database that I think was create using a custom template database.
Is there a way to see what template was used to create the database? There doesn't seem to be any documentation that describes this.
As far as I know, no, there is no system catalogue that tracks this. I'm not sure it would be terribly useful to know that database "X" was created from "template2" unless you could prove exactly what the contents of "template2" were at the point of copying. If you're tracking things to that level of detail then you will already have the creation logged.

Use PowerShell to write Data direct to OMS

I'd like to write my custom data direct to our Microsoft Azure OMS using powershell. Does anyone know how this is possible?
There's alot of information about configuring instances, adding new datasources, or querying the data using Azure​RM.​Operational​Insights:
https://blogs.technet.microsoft.com/privatecloud/2016/04/05/using-the-oms-search-api-with-native-powershell-cmdlets/
https://learn.microsoft.com/de-ch/powershell/module/azurerm.operationalinsights/?view=azurermps-4.0.0
Has anyone found a way to directly write data to OMS instead of storing it in a log-file and then use custom-log-import? I know this is possible, i saw it in a keynote, but cannot find any information about this.
Thank you for your Inputs!
This should do the trick - writing to Log Analytics:
https://learn.microsoft.com/en-us/azure/log-analytics/log-analytics-data-collector-api#sample-requests
Really not much to add, it's a copy/paste process. Entries will show using the type you define in $LogType appended with '_CL' because you essentially create a Custom Log entry.

OneDrive Custom Properties

I would like to store some additional metadata with documents that I am uploading to OneDrive using the REST interface. Does anyone know if this is possible at this time? An example might be a hash value that would help me detect alterations without relying on time stamps.
I can't use "Comments" because they are inside the user's field of view which would cause confusion.
Embedding data in the documents themselves is not an option. I don't control document formats.
Any help would be greatly appreciated.
Custom/extended metadata is not currently supported by the Live API for OneDrive.

Creating Metadata Catalog in Marklogic

I am trying to combined data from multiple sources like RDBMS, xml files, web services using Marklogic. For this as I see from MarkLogic documentation on Metadata Catalog (https://www.marklogic.com/solutions/metadata-catalog/), Data Virtualization (https://www.marklogic.com/solutions/data-virtualization/) and Data Unification it is very well possible. But I am not able to get hold of any documentation describing how exactly to go about it or which tools to use to achieve this.
Looking for some pointers.
As the second image in the data-virtualization link shows, you need to ingest all data into MarkLogic databases. MarkLogic can then be put in between to become the single entry point for end user applications that need access to that data.
The first link describes the capabilities of MarkLogic to hold all kinds of data. It partly does so by storing them as-is, partly by extracting text and metadata for searching, partly by conversion (if you needs go beyond what the original format allows).
MarkLogic provides the general purpose MarkLogic Content Pump (MLCP) tool for this purpose. It allows ingesting zipped or unzipped files, and applying transformations if necessary. If you need to retrieve your data from a different database, you might need a bit more work to get that out. http://developer.marklogic.com holds tutorials, blogs, and tools that should help you get going. Searching the MarkLogic Mailing List through http://marklogic.markmail.org/ can provide answers as well.
HTH!
Combining a lot of data is a very broad topic. Can you describe a couple types of data you'd like to integrate, and what services or queries you would like to build on that data?

Umbraco: Storing spatial data

I'm researching Umbraco for use as a base in a large CMS project, however the project calls for the SQL Server 2008 database to store spatial data against content.
Being new to Umbraco I'm still reading through the documentation and slowly building up an idea of it's architecture. However so far it doesn't look like Umbraco supports the storage of spatial data.
There only appears to be four database datatype options: date, integer, ntext, nvarchar
Is it possible to store spatial data to the database?
Update: Futher research into how Umbraco works has showed me I was on the wrong track. It seems the way to do this is store the lat/long data in the data inside the usual XML format Umbraco uses.
Then to use the Spatial.net extensions that have been built on top of Lucene.net, rather than use the limited search capabilities Examine exposes.
However this is all still theoretical, I've just not been able to achieve this. If I do before someone answers this question, I'll post my findings here to help others.
You could take a look at how to make user controls (with Visual Studio) in Umbraco.
It is also possible the versatility in Umbraco 'Document Types' is enough for you.
It is possible to extend Umbraco in any sort of way to get the solution you want. I don't know how you want the spatial data to interact with your frontend - so it is difficult to provide a direct solution.
Although there are ways to store spatial data and perform queries against it using Spatial.net, it's not a very elegant solution.
Instead I've created an additional table in SQL Server 2008 with the geometry/geography datatype and a reference to the Umbraco content it's connected with.
I've then got a event hook which updates this whether content is added/updated/deleted.