How to remove special characters from URL permalinks - special-characters

I am starting with jekyll and i'll appreciate any kind of help.
I am making a blog site but not in english language.
Default language will be Czech so in post tittles there will be characters like ĚŠČŘŽÝÁÍ.
I want to use pretty permalinks with post tittle in it, but actualy it doesn't work
properly in Safari browser. I am getting error 500 from server.
How to resolve it? Is there any plugin which can convert these characters in ascii symbols escrzyai and how to install it?

I believe permalinks are determined by settings + post file names, not settings + post titles.
So you should have files names like
_posts\2014-04-12-hello-world.md
instead of
_posts\2014-04-12-je-každoročně-tak-veliký.md
Inside the post, you can still use post title in Czech.
Permalinks' documentation is here, note if you have Czech categories and you use them in permalinks, you will still get into trouble, so it's better to avoid Czech categories in permalinks.

Related

What are these characters (�pret�) and how do I fix this?

I just don't know where I'm supposed to ask this, I don't know much about computers, feel free to direct me to the right forum, I searched for this but could not find out where to post.
Anyhow, I was doing search and came across this website:
http://tonioc.free.fr/dico/dico.txt
I was using Firefox. How do I fix this, use a different browser or is there an addon or something?
Thank you for your help
Check this site: http://apps.timwhitlock.info/unicode/inspect?s=%EF%BF%BD
This question mark character is a replacement character. Meaning that the website you are visiting is either using characters not supported by the font itself of the browser does not understand them. I tried with chrome and didn't get any result too. The characters are unknown.
You have to change the character encoding to Occident (windows-1252). There is an option inside Firefox.

What API does facebook use for displaying the full Wikipedia articles?

I couldn't find any api that return the article in a usable HTML form. Most of them return extracts which have very poor HTML formatting which makes them useless for anything.
There is no way to tell what Facebook did exactly, but the easiest way to grab the HTML contents of an article is by using the render action, i.e. by appending action=render to the URL:
https://en.wikipedia.org/wiki/Cooking?action=render
This produces the exact same HTML you can see on Wikipedia, but omits the non-content part (sidebar etc). If you need to reproduce the layout of an article more faithfully, you need to reuse parts of Wikipedia's CSS, and there is no easy way to do that.
Since just a few days there is a REST API for getting the html. It is available at https://rest.wikimedia.org/
Since it is so new, Facebook is probably not using it (yet) but if you want to get it for yourself I suggest you start exploring there.

AngularJS Forms and I18N support - not reading Japanese characters

A quick question regarding Angular Forms and Japanese characters. Am using Angular 1.2.17, and modern Chrome web browser on Mac OSX (latest).
Am writing an AngularJS application for Japanese market. Everything works great displaying Kana etc on the HTML pages etc. There are no issues with the web server, or Database etc, UTF8 support is throughout the application.
However, for the AngularJS forms, it does not read the Kanji / Hiragana / Katakana unless the word or sentence starts with latin character. Angular $scope appears to be unable to distinguish the fact that the JP characters have been typed at all unless prefixed with a latin character.
Example:
こにちわ does NOT register when typed into the input field, and hence form validation fails because it will think a required field is empty.
Whereas:
adsfこにちわ does register and the form can be submitted successfully. End to end, the JP characters are handled correctly, and get stored into DB correctly. So Angular / JS is parsing the UTF8 text correctly. The issue is likely something to do with how Angular binds to the data ($scope) when only JP characters are provided. It doesn't handle this properly by default.
Does anyone know of any HTML or Angular configuration etc - required Angular module or params, meta tags etc etc, that would coerce the Form to behave properly. Have not tested, but am pretty certain this issue is not specific to JP characters - it is likely anyone working with a non latin alphabet might have experienced the same behaviour.
Must be missing something obvious here.
Thank you for any help at all!
OK, updating this question very late - actually solved it very shortly after asking.
This turned out to be a timewaster question. Apologies.
But if anyone should come across a similar problem then please check for any REGEX declarations in the Form fields. For instance, a ng-pattern="/^[a-zA-Z]/"
Yes, this will do what it says, and exclude Kanji. Surprisingly it does NOT then put helpful validation error on the form field so from UI perspective it appears that the foreign language characters simply weren't registered.

TTStyledText that recognizes URLs without http:// scheme

I am using TTStyledText to display text with potential markup on a TTStyledTextLabel.
If text contains something like "http://stackoverflow.com" TTStyledText autogenerates the markup and linkifies said text.
The problem is I also want "www.stackoverflow.com" to appear as valid hyperlinks, as well as anything that reasonably resembles an http URL, even if the scheme is missing.
Any way to do that with TTStyleText or do I need to recognize the URL and reformat the text myself?
I've always found TTStyledText to be woefully underpowered and missing a lot of functionality. This is a good example. I think in order to make this work you'll need to as you say pre-format the text portion before you pass it in.

Sanitizing title URLs in ExpressionEngine 1.6.x

I run a blog where the blog title is either an external link or an internal link to a longer piece similar to what you’ve seen on similar blogs. For some reason, ExpressionEngine (1.6.x) does nothing to sanitize such things as ampersands in the URLs provided.
I use Markdown in the body text, which seems to do a great job of sanitizing all URLs. Yet, ExpressionEngine’s own handling of the titles doesn’t cut it. I have tried formatting the “title URLs” in Markdown and failed miserable, and damn if I know what the hell it is in ExpressionEngine that prevents me from using it.
So the question boils down to what other ExpressionEngine 1.6.x users do and have done, or whether someone can come up with a MacGyver-esque solution. Because I’ve been stumped upwards of half a year.
The XML Encode Plugin for EE1 from Rick Ellis of EllisLab will convert your special characters to HTML entities.
The plugin was originally designed to convert reserved XML characters to HTML entities in the ExpressionEngine RSS templates, but should work for what you need.
To use the plugin, wrap your {title_link} custom field in between its tag pairs:
{exp:xml_encode}
{title_link}
{/exp:xml_encode}
This would result in:
http://www.google.com/search?q=nytimes&btnG=Google+Search
Being converting into:
http://www.google.com/search?q=nytimes&btnG=Google+Search
Other EE1 Plugins which offer more similar but advanced features are Cleaner by utilitEEs (Oliver Heine) or Low Replace by Lodewijk Schutte.