Why there is no Tridion page template system metadata published to broker database? - metadata

We are using Tridion 2011 SP1 HR1 and New UI.
We would like to generate new UI JSON tags using our Java framework based on the Java content delivery API: Tridion delivery framework (TDF).
The JSON tag for the page should contain the PageTemplateModified parameter.
<!-- Page Settings: {"PageID":"tcm:48-84852-64",
"PageModified":"2013-02-28T15:04:30",
"PageTemplateID":"tcm:48-24343-128",
"PageTemplateModified":"2013-02-27T17:11:31"}
-->
But when a Tridion page or a Tridion page template is published to the broker database, the zip sent to the deployer does not contain any information about the page template. In the broker DB, ITEMS table does not contain any rows for page templates.
It is easy to verify using the following SQL query on the broker DB:
select * from items where item_type = 128
Is there a reason, why?
[edit]
I am speaking about system metadata.

This is very subjective,and not really a fit for Stack Overflow consider asking it at https://tridion.stackexchange.com/grant-beta-access?key=63880f9cbf452c62f726770209533893
I think the only way you will be able to get the modification date of a PT into your rendered page output would be to add the data to your page (or page meta) at publish time.
To answer the Why? Part of your Q, probably nobody thought of a reason for it, so it was never designed that way.

I was going to reply to the thread under Chris' answer, but it became too long...
The Last modified date setting in the XPM JSON cannot be the real last modified date, and I'll explain why.
When you load Experience Manager, it will quickly go through the various ""LastModified" settings in the JSON to compare with what is stored in the CM and determine if your page is up to date. If you published this page in December last year, then modified the template in January, the dates should not match - and XPM will let the editor know that "this page is not up to date, click here to reload".
If you would now change that last modified date "on the fly" to be the real last modified date of the template as opposed to the last modified date of the template at the time this page was published then XPM will not detect that there was a change made to the template on the CM and the editors may have unexpected results, like changing the value of a field, reload, and suddenly their sidebar is on the left instead of being on the right.
Hope this helps

Related

TYPO3: export list of pages with info about last editor

I need to create an export of all pages and get information about the last modification date and who did the last modification.
Last modification date is easy: that's stored in SYS_LASTCHANGED in the pages table. But how can I find the information about who did the change? For changes to pages that's easy as well, as I can check sys_log and sys_history and search the pid in the field recuid in combination with tablename. For tt_content records I can do a select first to get all records for a specific page. And then search recuid and tablename accordingly.
But how about all the other records, e.g. from plugins? Do I really need to iterate over all tables and the pid field to get all possible records?
I think you have a lot of work to do, as the information SYS_LASTCHANGED is not very trustworthy.
Depending on the rendering process multiple records are used for building up a page. and more and more information does not come from records but from files, which are versioned by chance in a git repository (or similar).
Just in case all(!) information is stored in records (tsconfig, typoscript, fluid templates in records (sys_file for files)). There are thousands of ways a page is generated and which information influences the current page rendering.
just some examples:
the typoscript is changed. does it result in a rendering change of the current page?
a record (e.g. tt_content) is deleted. so this record will no longer show in the page. will you consider all invisible (delted, hidden, time restricted) records?
TYPO3 has constructions like: show content from page X, show content records X,Y,Z from other pages
translations: does a change in another language than the current is a change to the current language (a fallback might be possible)
if you consider links in menus or text: if another page which is linked in the current page gets disabled TYPO3 does not generate a link any longer

Kibana - what logs are not reporting

I am currently using kibana 5.0 almost 45 log sources are integrated with kibana like iis,vpn ,asa etc.now my question is how to create a visualization to check what logs sources are not reporting to kibana.can anybody help on this?
Quick and dirty solution...
Make sure each log source is given a unique and meaningful tag as soon as their data enters the logstash workflow.
As each document is processed write an entry to a separate index, call it masterlist.idx (do not give this index a date suffix). Use the tags you assigned as the document ID when you write entries to masterlist.idx.
The masterlist.idx should really just contain a list of your log sources with each entry having a timestamp. Then all you've got to do is visualise masterlist showing all the entries. You should have 45 documents each with a timestamp showing their latest updates. I think the default timepicker on Kibana's discover tab will do the job. Any sources that haven't been updated in X days (or whenever your threshold is) have a problem.

CRM 2011 and SSRS - Generating a report for a single record

Is it at all possible using CRM 2011 and SSRS to generate a report on a single record, and only get results for that one record?
EDIT
Additional Info - Must Use:
Custom SSRS report
Custom entity in CRM
Here's a more specific link to your question: link. You're probably looking for pre-filtering (look for "3. Pre-filtering Element" in the link provided) if you want the report to be record specific (context sensitive).
Here's a link describing the 2 types of pre-filters (CRM 4.0 but the theory applies to CRM 2011): link. And here's an example of prefiltering in CRM 2011: link
I have done this successfully in CRM 2011 with a completely custom report made in BIDS, on a custom entity, with full context sensitivity.
Make sure to learn fetchXML as it's going to be the going forward technology for these reports. The existing reports are using SQL which make them bad examples to copy off of.
Here's an example on how to extract fetchXML from an advanced find: link It also has more information on pre-filtering.
Take a look a the report Account Overview.rdl. It could be executed for a single account record or multiple records.
See Reporting for Microsoft Dynamics CRM Using Microsoft SQL Server Reporting Services
Create an embedded connection to the CRM database engine for the environment you want to target.
Create an embedded dataset to query the current record. This going to be kind of weird since experience will tell you that you are going to get tons of records, but because of the clunkiness behind CRM it will actually only get the current record. For example, if you wanted to get the current quote you would use "SELECT quoteid FROM FilteredQuote AS CRMAF_Quote"
Add a parameter to store the reference to the entity you just queried. In keeping with this example I created #QuoteFilter which is type text, could store multiple values (even though that's not what we're using it for), and gets its default value from the dataset in step 2. Also, probably ought to make this hidden since GUIDs aren't end user friendly.
Finally, use the parameter discovered in the where clause of the other datasets. For example, a search on quote products for the current quote would look something like SELECT * FROM FilteredQuoteDetail WHERE (quoteid = #QuoteFilter)
As a final note, you should keep in mind that CRM loves to remember everything even when you don't want it, too. On one of my reports I messed up my datasource and CRM was forever convinced that the report should run against all records. I fixed my datasource, but uploading the report did not trigger a refresh and correct the problem. In the end, I deleted the report from CRM, created a new one, uploaded the same exact file with no changes, and everything worked. Go figure.

Orchard CMS BlogML old entries

I want to import old blog posts into Orchard CMS's blog module.
These are not contained in any xml format, if I try to add them manually then my posts display by creation date which ruins the order of original creation date of my blog. I noticed with Windows Live Writer I can set the Publish date but this still does not effect the Creation date.
There is a module called BlogML, but my blogs are not in RSS format. I see the BlogML.xsd is here http://blogml.org/, but is there some software I can use to generate the correct XML which I can then import into Orchard? And will this fix the creation date at all? I think I may be left in creating a custom type say 'imported blog' and add extra fields on which I can sort by.
If you are so worried about the creation date, you could just go into the database manually and change the creation date for each blog entry. Would take a lot less time than anything else.

Eclipse Birt Reports, Creating report from SQL database, (user key?)

I'm fairly new to using the Birt Report Designer and need to figure out how to generate a report from a SQLite database. I have suceeded in getting it to connect to the DB but am now unsure how to generate a report and the tutorials that I have found aren't of much help so far.
I have a template that was given to me by my employer that has a few fields, I'm wondering if these fieldnames (in the template) are supposed to match field names in the DB.
Also, when I go to Run->View Report-> As PDF I am unsure what I am supposed to enter for the field "User Key", does this correspond to a table name in the DB or something along these lines?
As of now, I have tried entering a table name but just a blank report is generated.
If anyone can point me to a good resource or help with this I would greatly appreciate it. Thanks
There are two books i could really advice:
BIRT - A Field Guide to Reporting
Integrating and Extending BIRT
and the Eclipse Help containing BIRT documentation.
I suppose the User Key could be report parameter (listed in Data Explorer window), which is passed to Data Set to select appropriate data. If I'm guessing right, check within a Data Set editor ("Parameters" tab and "Query" tab) where the User Key parameter goes in - probably to one of the table field in a WHERE clause. Parameters in a query are represented by question marks: SELECT * FROM fooTable WHERE barColumn = ?. Hope tracking this would lead to find out, what to enter to the parameter.
Additionally, ensure if your Data Set(s) is(are) connected correctly to your SQLite Data Source ("Data Source" tab in a Data Set editor).
Being as new as you are to BIRT, I would suggest building a couple of reports with the sample DB (Classic Models). There are many, many samples out there for you to use as a guide. Additionally, most tutorials will use the Classic Models data so you can follow right along. After you create a couple of practice reports (this should not take more than 30-45 minutes) the template you have been given will likely make A LOT more sense and allow you to make progress almost immediately.
If you are looking for a nice collection of tutorials and samples, be sure to check out Birt Exchange for Dev Share (samples) & tutorials.
As for the "User Key" this is almost certainly a report-level parameter used to filter the data set (as the previous answer points out).
Good Luck!