I had an application that used a Sybase ASA 8 database. However, the application is not working anymore and the vendor went out of business.
Therefore, I've been trying to extract the data from the database, which has Arabic characters. When I connect to the database and display the contents, Arabic characters do not display correctly; instead, it looks something like ÇáÏãÇã.
Which is incorrect.
I tried to export the data to a text file. Same result. Tried to save the text file with UTF-8 encoding, but to no avail.
I have no idea what collation the tables are set to. Is there a way to export the data correctly, or convert it to the correct encoding?
the problem was solved by exporting the data from the database using "Windows-1252" encoding, and then importing it to other applications with "Windows-1256" encoding.
When you connect to the database, use the CHARSET=UTF-8 connection parameter. That will tell the server to convert the data to UTF-8 before sending it to the client application. Then you can save the data from the client to a file.
This, of course, is assuming that the data was saved with the correct character set to begin with. If it wasn't, you may be out of luck.
Related
I started using HeidiSQL to access Postgres, where the base is under WIN1252.
When opening a table, where it has data like ~ç and other special characters, it is not showing correctly.
I couldn't find where to change the display configuration.
I am loading data into QlikView report from different sources, one of them is Sybase db. Seems like Sybase db is using ISO 8859-1 encoding, but there are also Russian characters there, and QlikView just don't display them properly.
I don't see the way to manually define encoding in Qlikview. Is there any?
I tried to specify cyrillic charset in ODBC settings, but it also doesn't help. Funny thing is ASE isql (tool to run queries on Sybase) there is no issue with encoding. Can I specify encoding when select stuff in Sybase?
Sounds like a charset conversion issue. My guess is that your isql has a charset conversion option enabled, but your qlikview session has not.
I have installed postgreSQL.
I use pgAdminIII as admin panel.
I am trying to watch tables content and see following:
How to avoid encoding problem?
For a UTF8 database, pgAdmin should always display strings correctly. The most likely explanation is that the data itself is incorrect.
This generally happens when a client application sends data in a format which doesn't match its client_encoding setting. If this is the case, setting client_encoding correctly would prevent this from happening (provided the client application's code page is supported by Postgres). This wouldn't fix the existing data, but it might be possible to repair it with the convert function.
I have recently started using PostgreSQL for creating/updating existing SQL databases. Being rather new in this I came across an issue of selecting correct encoding type while creating new database. UTF-8 (default) did not work for me as data to be included is of various languages (English, Chinese, Japanese, Russia etc) as well as includes symbolic characters.
Question: What is the right database encoding type to satisfy my needs.
Any help is highly appreciated.
There are four different encoding settings at play here:
The server side encoding for the database
The client_encoding that the PostgreSQL client announces to the PostgreSQL server. The PostgreSQL server assumes that text coming from the client is in client_encoding and converts it to the server encoding.
The operating system default encoding. This is the default client_encoding set by psql if you don't provide a different one. Other client drivers might have different defaults; eg PgJDBC always uses utf-8.
The encoding of any files or text being sent via the client driver. This is usually the OS default encoding, but it might be a different one - for example, your OS might be set to use utf-8 by default, but you might be trying to COPY some CSV content that was saved as latin-1.
You almost always want the server encoding set to utf-8. It's the rest that you need to change depending on what's appropriate for your situation. You would have to give more detail (exact error messages, file contents, etc) to be able to get help with the details.
I'm trying to import a large dataset into Google Fusion Tables using the csv-import function. The data contains Danish æ-ø-å characters. The original encoding of the data seems to be ANSI (or "windows-1252"). Data uploadet in that encoding is not displayed correctly. I've tried to reencode the various strings in most other relevant encodings (Encoding.(Unicode|ASCII|UTF8) etc.) but nothing seems to please Fusion Tables.
I'm using FileHelpers to generate the csv and I have tried explicitly setting the encoding there too, but to no avail.
UTF 8 should work. I had uploaded this table as a test: http://www.google.com/fusiontables/DataSource?dsrcid=276537