Joined: Nov 2008
Posts: 19
Thanks:
0
Hello,
I was wondering if there was a software that could help me make glossaries by extracting the words from the texts I am using (newspaper articles, web-pages, ebooks etc). I remember JWPce could extract individual characters, but what about compounds and kana words?
Thank you.
Joined: Feb 2008
Posts: 1,322
Thanks:
0
I'm not quite sure I'm understanding what you're looking for exactly...
But it seems like Copy/Paste would suit this quite fine...?
Joined: Oct 2009
Posts: 3,944
Thanks:
11
There's a program called "chasen" but I don't know if it does exactly what you want.
Joined: Jan 2008
Posts: 131
Thanks:
0
I think you can try the technique used in the "Ultimate Study pack" ...is somewhere in the forum...
Joined: Aug 2009
Posts: 710
Thanks:
0
I feel like I should be excited by these programs, but my tool-using brain functions aren't working properly at the moment. I can still use a stick to poke an ant-hill, though.
Edited: 2010-03-17, 5:33 pm
Joined: Nov 2008
Posts: 19
Thanks:
0
Thank you all for the new suggestions on how to "skin" the text.
I also tried wakan, but I could not easily figure out how to make the wordlist with my text (although I could with a smaller, only Japanese text).
Cheers.
One last question: is there a way to change kana to romaji in open office?
I tried using chasen, but it's a bit complicated.
Joined: Aug 2008
Posts: 56
Thanks:
0
Chasen is the old way to do this. There's also kakasi - but it too is an outdated way.
The best way to segment words now is to use MeCab, which was written by the same author as Chasen. It has a mode which will split sentences into their constituent parts.