Is it safe to edit the schema of a blog post in Sanity, after I already have added 4-5 long posts (using the same schema)? - content-management-system

I currently have a blog running on Next.js + Sanity. I have had a lot of success, but in the past when setting it all up encountered a scary error while editing the blog post schema, that basically crashed and caused me to lose my most recent blog post. I assume it had to be an error with trying to change schema of a blog post as it was being written?
Anyway, I now have a few long posts published, and I don't want to lose them. They are in production and getting views, but now I'm realizing I want to add alt tags to all of their images, which I will do via editing the schema.
Is this known as a safe thing to do with Sanity? I can't afford to have these posts go down on me.
Thank you.

Yes, this is safe to do. If you want to add fields in your schema you won’t run into problems with the existing posts. It’s only when you change the schema type or rename schema fields where you might run into problems. However, these can be solved by exporting your data, reshaping it, and then importing it into your project again. You can find more info about that here: https://www.sanity.io/docs/migrating-data#1-export-c4665bde1f66

Related

How do we share full NetLogo models here when working on a problem?

I'm pretty new on this list, and trying to answer a question posted here on StackOverflow and i am wondering if there is some standard way to post entire models here, not just the "CODE" tab portion. Or for that matter even an image of the view screen.
I don't see any place to put attachments on posts or answers. I suppose I could upload the model to the NetLogo-users Google users-group, which does have a place to upload files, and cross-reference it here. I suppose I could try to open an email channel to the user and send them email with an attachment. I suppose I could post it on GitHub.
But is there some way to attach a full .nlogo file right here that I'm missing? Some interfaces are really complicated and only looking at the CODE tab is not adequate.
And, yes since the .nlogo model is pure text I could paste the entire thing into a window here ( which would object to having code in a text window of course) but that's a lot of extra characters in the post.
Unfortunately there is not. This is actually more of a problem for the questioning than answering as it can be difficult to get the person asking the question to post the relevant bit of code, including the other bits of code that lead up to the problem and give key information like what the contents of a variable may be. NetLogo does not lend itself to MWE at all, and beginners simply don't have the experience to replace interface variables with global variables etc.
Uploading to Google users group and cross referencing is likely to get the question/answer closed as it's not complete. But StackOverflow has a different purpose than the users group - it is supposed to be focussed on specific questions with specific answers - such things as syntax problems and bugs, not design. The last thing we want is interfaces or full models. We tend to be more lenient than other areas of StackOverflow because we know NetLogo has a very high proportion of beginners without support and that MWE in NetLogo doesn't really make sense, but questions that require full models are definitely out of scope.

Kentico Import Tool inconsistent/buggy when updating documents

I've had a number of problems using the provided Kentico Import Toolkit, namely when using the "Import new and overwrite existing pages" option to update my existing/already imported pages. I'm using a custom SQL query to import and have had a profile saved for each import I've needed (client has article based site so a few tables of similar information) to try and keep each as consistent as possible between imports.
Here's the problems I've encountered thus far (in no particular order):
the tool tries to guess which fields from the query correlates to the fields of the page type in Kentico for you, which is a nice idea, but seems poorly implemented. If I'm not very careful and reload the profiles every time I import I've had instances where fields changed inexplicably when testing imports because the tool thought it knew which field I wanted
this is more the problem when importing/reimporting multiple times in a session and choosing to go back and load the same profile (without reloading)
the NodeAlias field is only seemingly required on update/reimport rather than on initial import. I'm sure there's an internal cleaning of the document's title to generate a NodeAlias and this is generated fine when importing documents while NOT providing the NodeAlias. After importing the items initially and wishing to update however the NodeAlias is seemingly required as you'll get errors with text asking it be included. This implies to me that there's matching of the NodeAlias along with the given ID field, which should be fine in theory but isn't specifically mentioned anywhere in the tool as best I can tell.
I've had instances where reimporting items will change/strip their NodeAliasPath. I've gotten around this by specifically setting the NodeAliasPath (which only shows after selecting "Show Advanced Columns") but like NodeAlias path before it, I'd think the tool should be smart enough to know to keep the path if not specifically given for updated items.
it seems very odd that in order to match on ID for previous items you have to provide the name of the new column instead of the old one. My example: client was using just a field named 'id' and the new one is 'OriginalID' to clearly differentiate it from the Kentico derived ID fields. To match the items I have to use 'OriginalID' rather than 'id'
A couple of notes/niceties or potential updates along with the above:
it would be nice if there were some way to select if the page should
be published or not through a single query. Currently having the
"Automatically publish pages under workflow" toggle checked seems to always publish
the items. I have an instance where the client has old documents in
the provided DB dump that they don't want visible on the site but
want preserved in the DB if they change their mind later. Currently I
have to perform 2 imports, 1 for the unpublished and a second for the
published items, to accommodate this, which is quite cumbersome
I'll likely edit/add to this as I get responses. This isn't really a specific problem (as I managed a workaround to the NodeAliasPath stripping problem, which inspired this post initially) but more just me asking if these are bugs,if I'm not using the software as intended, etc.
You've stated all the problems you're having/experiencing and possibilities why they are happening but didn't ask a particular question. If you suspect they are bugs, then I'd go to directly to Kentico Support and report the issues there since these are things that have been part of the KIT for as long as I have worked with it.

Spammy structured data Issue (Markup Ok but considered spammy by Google)

I received the message in Search Console about spammy structured data regarding one of my websites. Mind you - It's an older website that has not been up to the standards.
So I had all the error and warnings fixed and I submitted website for reconsideration. Reconsideration has been refused. Problem is I do not know why? Website pass the structured data testing tool test. There is no errors or warnings. At the moment markup is done within HTML ( not json-ld). Could someone take a look and advise me what I can do to have it fixed according to google guidelines and manual action removed.
I', not so sure if this is important or not, but index page is not the only one to contain markup for LocalBusiness. There are more pages similar to the index one, based on different location.
Website in question is http://www.man-van.biz
Thank you for your help in advance.
Problem has been solved. Problem was with duplicated schema markup on aggregate ratings (the same feedback's on all pages). Once I removed it from other pages and kept it only on index page and fix some other issues with markup reconsideration request has been approved and manual action removed.

Content merging AEM

We're looking for a solution on how to best deal with the situation where multiple authors are working on the same page. If the first author pushes in the content, the second should have a way to merge it when he tries to publish. Launches appears to be a way to take care of this but it doesn't seem to be handling content merging. Is there any way an author can view the diff(and or do merge) of the content that might have been pushed by another author while they were working concurrently ?
Please help with any pointers.
Page modifications happen in real time to the underlying structure. They also happen at as small a level as possible.i.e. If you go into a text area and modify the text there, the text node is changed on the server, you aren't saving the entire page.
The only way that person A could interfere with what person B is doing is if they were working on the exact same area of the page. Which, honestly is a process issue. I say this because the answer to your question is that there is nothing out of the box to handle this type of scenario and if you are on 6.0 or higher and looking at the JCR3. JCR3 handles this far worse than the older version did. Last time I checked it didn't support nodes at all
Adding to what Bailey said, AEM OOB allow multiple users to edit same page in real time, though if multiple users are working on same node will be a reason of conflict. Such cases can be managed by defining a process like:
1. Take a lock of page and edit page or
2. Create versions of page and publish versions

List of possible 'changes' values in Facebook Realtime Update

I've implemented this for my site:
https://developers.facebook.com/docs/graph-api/real-time-updates/v2.0
and I am receiving updates successfully. The format of these updates is fairly easy to understand by understanding the JSON response. However, I can't find a list of possible types of changes (possible values for the JSON response) so I can handle / replicate all the scenarios that may come up.
Is this published anywhere?
The possible values for 'item' and 'verb' are not documented. Here's everything I have observed when receiving updates for pages only:
Items:
like
share
comment
photo
video
post
Verbs:
add
remove
edited
hide
unhide
Can't claim this is comprehensive, but that's the best of my knowledge based on reviewing logs of many thousands of updates. I would guess that user, permissions, and payments RTUS have their own sets of possible values.
Not sure what you mean with types of changes exactly. There's
https://developers.facebook.com/docs/graph-api/real-time-updates/v2.0#subscribefields
https://developers.facebook.com/docs/graph-api/real-time-updates/v2.0#receiveupdates
If you click on the first links down arrows on the right side, you'll get the list of fields that can change for the user and the page objects.
After some searching, I came across https://developers.facebook.com/docs/public_feed#updates There's a reference to http://activitystrea.ms/ in the code:
<activity:verb>http://activitystrea.ms/schema/1.0/post/</activity:verb>
Unfortunately, the website doesn't work apparantly, but one can view an old version here:
https://web.archive.org/web/20140720095121/http://activitystrea.ms/registry/verbs/
This seems to contain the list of possible verbs wihtin this schema. Which of those Facebook actually uses is not clear to me unfortunately. There's the specs as well: http://activitystrea.ms/specs/json/schema/activity-schema.html#verbs