Joined: Jul 2007
Posts: 2,313
Thanks:
22
Ok, just updated the front page. Also requested Ropsta add the credits to the lists he compiled to identify who helped here. I removed the minor lists that were completed to aid use.
Out of interest, anyone know a program or a way to use a spreadsheet to prioritize sorting? What I mean is, given a group of kanji, the program looks for words in a list with those kanji (anywhere in the word) and organizes that list based on that.
Joined: Jul 2007
Posts: 2,313
Thanks:
22
Ok, I guess star office calc has a way to sort of do it, but it's on the first character only. It's a matter of making a user defined sort order. For a larf, I'm going to try to organize 401-2000 of Core Basic, and 2001-4000 of Core Intermediate.
I'll post the order in Google Docs with requests for opinions on how the order came out. To make it a useful list, it'll have numbers to help you organize it via Smart.fm or pseudo KO2k1 method. That way, it's a simple matter of pasting to the full list on google docs, resort to the ko2k1, then import to Anki.
Speaking of which, which thread had a link to the kanji in the KO2k1 order?
Joined: Jul 2007
Posts: 2,313
Thanks:
22
pubbie,
Not quite sure how you're doing this as I have a slow connection to check out your work. Now, are you organizing the core2k and core6k vocabulary words using the KO2k1 order then using that order sort the sentences attached to the vocabulary words?
Reason I ask is that it sounds like you script is looking for kanji in the entire sentence then putting that in order. That could mean a sentence can be used upto 10 times if it has 10 kanji in it.
A better method may be: Sort Core2k and 6k Vocabulary words with your script into KO2001- put sorted list in spreadsheet and add a number column (for resorting). With that, it's not too much to sort both Core2k and 6k spreadsheets that have everything (photo, audio, sentence, translations, etc.).
Edit: Can someone compile the next set of completed sentences (361 to 450 I believe). Don't forget to credit who compiled which set.
Edited: 2009-04-01, 4:52 pm
Joined: May 2008
Posts: 327
Thanks:
0
The problem is that we DON'T have a list of the Kanji in KO2001 order... and it would take AGES to make one up manually. I asked in another thread if someone with a scanner or one of those pens could do this, but no-one responded...
I think this is worthwhile, because there are a lot of people who don't use iKnow/Smart.fm but use ANKI instead...
Edited: 2009-04-02, 5:22 am
Joined: Jul 2007
Posts: 2,313
Thanks:
22
Pubbie, I don't think it's just me, but others like to learn vocabulary via the sentence method. Others like to be concerned with the entire sentence, I'm concerned mainly with the word then secondarily with the sentence.
Plus, you can be systematic with vocabulary words. I have the top 60,000 words in Japanese which was developed by scanning 250 million characters on websites. Removing kana, romaji, numerals and special characters leaves you with a useful list. However, iKnow sort of kind of has done this already by using vocabulary lists developed by frequency in newspapers. What I wanted was take those frequency based lists, then organize them using the KO2k kanji order. Reason being people say they learn words much, much faster.
Anyway, I recommend that with your program (if it doesn't do this already) is all but the last dupe gets deleted. By last, I mean the sentence or vocabulary word that has the more "difficult" kanji ie latter in the list. So if 水道 comes up, there's an entry by 水 and by 道. So the word will be put later in the list by 道 kanji while the one by water is deleted. If it comes across all kana words or sentences, then those get evenly spaced out in the list.
For Core2k and 6k sentences, recommend the following "Bunches"
Core 2000
First 400
Last 1600 (401-2000)
Core 6000
Next 1000 (2001-3000)
Next 1000 (3001-4000)
Next 1000 (4001-5000)
Last 1000 (5001-6000)
My reasoning: The first 400 are very common words that presented no problem to people based on comments. The next set became a bear very fast. So these 1600 can be organized. After that, I say organize the next groups of 1000 on the idea that you have nice controlled blocks of words for quick learning and usefulness.
Joined: Mar 2008
Posts: 1,533
Thanks:
0
You know, I don't think it even has to be the 2001ko order... It just has to be an order that builds from easy to hard. The official Japanese schooling order would work better than nothing, in fact.
I think I'm going to try to find some time this weekend to write a script that will sort things like you are suggesting, Nukemarine. I don't think it'll be too hard, actually... It'll be harder to optimize it than anything.
Joined: Jul 2007
Posts: 2,313
Thanks:
22
Umm, what happened to pubbie's posts? Did I miss something or was this behind the scenes?
Joined: Aug 2006
Posts: 1,022
Thanks:
1
I just wanted to thank everyone for their efforts making these lists.
I went through KO, and now I'm using the smart.fm lists for listening practice by importing them into Anki. I can read most of the sentences but listening is often really hard for me 1st time round, so I think it will be worth while doing this listening practice.
Joined: Apr 2008
Posts: 1,635
Thanks:
0
I agree that Python & Ruby are nice. But their unicode support kinda lets you down.
Perl's unicode support is flawless. But then you gotta take care not to write "write-only" code.
How is PHP with unicode?
Edited: 2009-04-05, 7:18 am
Joined: Apr 2008
Posts: 1,635
Thanks:
0
Lol, and why not COBOL then?