Is there a way to recognize characters other than Japanese and the alphabet?
Is there a way to recognize characters other than Japanese and the alphabet?
To what language do you want to use?
I have a plan to develop a feature that enables users to inject user-customized language processer, so I want to know your use case.
chinese
@wtetsu I want to use it for eng -> turkish, is this only eng -> jp?
@toprak En -> Any language would work perfectly. However, you have to find or make dictionary data and import it.
Is it not necessary to rewrite the program just by importing the dictionary data?
@yamadayo Please try it out. Maybe, it's not smart like English/Japanese parsing, but partially work.
If you need smarter parser for other languages, program improvement is required.
Greetings. I've tried Spanish and didn't work at the accent alphabets e.g. "programación"
The example dataset:
{
"lenguaje": "language as in choice of words\nprogramming language",
"lenguaje de programación": "A programming language",
"lengua materna": "mother tongue"
}
Screenshot:

I would suggest a workaround something like enabling user to map unsupported characters e.g. "ó" -> "o", "ñ" -> "n";
@yuis-ice Thank you for the information. Let me consider it.