I have many files on my webpage, which I wanted to standardize to utf-8. To automatically detect which ones contain problems I made a small script which detects the actual encoding (using chardet
for the detection).
A file containing the ñ
is detected by chardet
as ISO-8859-9 (as I understand it - correctly).
The file contains a 0xF1 byte (coming from the ISO encoding or Latin-9)
If I use geany
to convert the document to UTF-8, the file doesn't seem to change (even though the tab changes to indicate the change). The 0xF1 is still present.
I suspect that geany (or the called library) doesn't feel 0xF1 is not utf-8, hence doesn't translate it.
Is there any work-around?
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you are subscribed to this thread.