Authors Need Help: Brainstorming Session
Thursday, March 29th, 2012 08:15 amOK, I don’t have much experience coordinating a project this big. My general approach, when confronted with an Enormous Project is to break it down into bite-sized pieces, but! I’m not a project manager and have none of that Foo.
Below, is what I’ve got — comments and advice welcome. I’ll also be looking for volunteers, but that’ll come later.
I’m soliciting ideas on how to implement in order to achieve the goal without loss of life, and without anyone having to bear an enormous burden of work.
The Goal: A list, for each novel, of all the Liaden-and-other-Weird-Words that appear in that novel, AND a list of Liaden-and-other-Weird-Names that appear in that novel.
What the lists would look like:
1. Title of Book, Edition
a. Word One, Page Number
b. Word Two, Page Number
Lather, rinse, repeat
2. Title of Book, Edition
a. Name One, Page Number
b. Name One, Page Number
Lather, rinse, repeat
I’m guessing that there ought to multiple eyes on each novel, in order to make sure that the maximum number of Weird Words (henceforth WW) are captured. Some of the WW will be English words (we use a smattering of obsolete English words, Just Because), some of the WW will be Terran slang, Delgadan words, and the ever-popular etcetera.
For the names — I’m guessing another buncha eyes for each book, so that the maximum number are captured.
Question: Planet and ship names — Different lists? Or folded into the Names List?
Also needed, someone or someplace to receive, and coordinate, the lists.
Ultimately, the lists will be used by Lee and Miller for Something Really Cool, and will play an important role in the Web Pronunciation Guide Project.
There is some time limitation on getting this together, but at the moment, the deadline is squishy.
So! What’s the best way to set this up?
EDITED TO ADD: Please nobody build anything yet. We're still in brainstorming mode. I appreciate everyone's input.Originally published at Sharon Lee, Writer. You can comment here or there.
no subject
Date: 2012-03-29 12:27 pm (UTC)no subject
Date: 2012-03-29 03:33 pm (UTC)The only thing this doesn't find are names or words that are spelled like standard English words but used for something different, like a person named Cat. The name Jeeves is not flagged by spellcheck, for instance.
There's a free-trial program that will find all the unique words in a document or text file. I haven't downloaded it or tried it yet. You'd wind up wading through a lot of and, or, but, the, a, and said, but it would catch the names that are in the dictionary.
ETA: The free-trial software is Word Patterns by Mysoftwarefactory.net.
no subject
Date: 2012-03-29 11:08 pm (UTC)no subject
Date: 2012-03-29 04:16 pm (UTC)The only part you'd have to do manually is to separate Names from Words.
Author Needs Help
Date: 2012-03-29 01:11 pm (UTC)Book Title
Page Number Word/Name Definition
Toni
Re: Author Needs Help
Date: 2012-03-29 01:28 pm (UTC)Re: Author Needs Help
Date: 2012-03-29 02:02 pm (UTC)Toni
Re: Author Needs Help
Date: 2012-03-29 03:21 pm (UTC)no subject
Date: 2012-03-29 01:16 pm (UTC)I happily volunteer too!
Date: 2012-03-29 01:26 pm (UTC)Google Docs Form?
Date: 2012-03-29 01:34 pm (UTC)no subject
Date: 2012-03-29 02:10 pm (UTC)I would go with a spreadsheet for collating. I would not, however, go with Google docs - too many cooks, etc
I volunteer to work on the project.
no subject
Date: 2012-03-29 02:26 pm (UTC)Most words and names will appear repeatedly in books; I presume you want the first time each is used? Or?
no subject
Date: 2012-03-29 06:56 pm (UTC)Author Needs Help
Date: 2012-03-29 02:28 pm (UTC)Just once? or every appearance
Date: 2012-03-29 02:30 pm (UTC)Since I also agree that using volunteers to spread the work is the way to go, the next complication will be to determine the "format" of each book so that page numbers make sense. That means to me, that either you supply them, or there is another task to determine if various electronic versions, e.g. ePub and Kindle editions, match on pagination.
I also strongly recommend a trial run, after we THINK we have all the bases covered. That is - set up the spread sheet a volunteer will work with, the final repository, and try it with a handful of folks for ONE book, before letting loose the hordes of volunteers.
I, also, would be happy to be one of those - volunteers, that is...
Werid Word Coordination
Date: 2012-03-29 02:33 pm (UTC)You could give each book a page and assign two or three or however many people you like to each page. they in turn list the words and such on that page.
I would recommend this as I've used PB works wiki and it works well. It's not perfect but it does well with this sort of collaboration with people who aren't in the same geographic location.
Here's a link to the wiki of my class from last Spring: http://digitaltextuality.pbworks.com/w/page/35113284/Syd%20loves%20Learning
no subject
Date: 2012-03-29 03:24 pm (UTC)no subject
Date: 2012-03-29 04:58 pm (UTC)http://www.anindexer.com/about/sw/swindex.html has a good overview of other software options. An interesting one is HTML Indexer at http://www.html-indexer.com/
Now, I'm off to a meeting...
no subject
Date: 2012-03-29 04:57 pm (UTC)2. And so, we called it it database and that is what we should use for this task. It seems a perfect relational database application. Something like Microsoft Access (is there still a Microsoft Access?) will be able to do the job very handily. Some programming will be required.
I am not volunteering because I have been retired for way too long and I are obsolete and too old to play anymore. And "way too long" in this business could be matter of months, not years.
So what is needed, aside from humans to implement this, is a computer accessible to all those humans and on which this database (LiadenWord) resides.
Good Luck, and as Groucho said, "Stay Warm."
no subject
Date: 2012-03-29 09:27 pm (UTC)brainstorming
Date: 2012-03-29 11:36 pm (UTC)Sounds scary now doesn't it? Start small. Think first about what your readers want (pronouncations and definitions) and what you and Steve want/need (hey, why shouldn't you benefit from this adventure) in terms of what you NEED right now and a Wishlist for the future. Start with a chapbook and learn a bit about it and get someone design the database and entry forms (volunteers). Then populate the database with the results of a script that reads the electronic book. If you play your cards right, all you (the authors) will have to do is define and pronouce each unique item not in English and enter it into the database.
Once you have the database (such as MS Access as mentioned above), you can ask it to print out the lists Any Way You Want. The good news? I see you asking your readers for help on remembering when characters showed up and what you said about them. If this gets done properly, you'll be able to have the answers immediately because you will know what page and what book.
beth
no subject
Date: 2012-03-30 12:09 am (UTC)The same software that runs Wikipedia is open source. The only downsides would be the initial work needed to set it up, and dealing with spam (there might be a way to reduce this.)
For the list of words do you want every time it appears in the book, or just the first time?
no subject
Date: 2012-03-30 02:09 am (UTC)1) A nonstandard word list -- lots of false positives, but the stuff that isn't in a list of standard English words
2) The paragraph each word is in.
3) The word count for each word (i.e. word #49473 in a book.)
The advantage a Perl script has over other computerized means of accumulating a list is that it's more flexible and easier to tweak for exactly what you want -- if you wanted the first three times a word appeared, for example, that's very doable.
Very Large Project
Date: 2012-03-30 03:58 am (UTC)As for the other books....that would be too much like work. I will leave those books to your expert commenters who no doubt have your books on their e readers ....
no subject
Date: 2012-03-30 03:58 pm (UTC)Do a list of the books and a page count on each, that's your first database entry. (Ghost Ship, paperback, X pages, Ghost Ship, hardcover, X pages).
Take that and generate a list of links for each book, one link per page.
Phase 1--the users go to the links and click and get a form, and they put in the word "None" or write a comma separated list of the words and hit submit.
As pages are examined, we show ONLY pages that haven't yet been checked, so we insure that every page has been read by someone.
When there are no links left on any books, we go into Phase 2. Again, you see the list, but this time when you click the link, it shows you what was put in the first time. You click "yes, this is correct" or "no, this is incorrect". If it's incorrect, the link will show up on another page needing review. If it's correct, once two or three people verify the page is correct, the link is finished.
Once all links are checked and the ones that need reviewing get reviewed, we just ask the database to generate the list, and it can go into a spreadsheet or a word doc or whatever.
This would be about six hours of programming, it is, believe it or not, a fairly simple job and would let the maximum number of people help. And I have time this weekend... :D
no subject
Date: 2012-03-30 10:10 pm (UTC)And "philter", too. :-)
no subject
Date: 2012-03-30 10:18 pm (UTC)It will help if you also respectfully remove all processed comments to make it obvious where there is something new.
Whereas using overly-smart software will leave you forever worrying that you didn't do it quite right.
Volunteering, anytime, anywhen
Date: 2012-03-31 02:08 am (UTC)Project Suggestions
Date: 2012-03-31 11:04 pm (UTC)This would be in addition to actual readers. You would still need people to manually/optically search for the obsolete and unusual English words, which may not be picked up by spell-checker.
In this instance, the Hard Work would be the initial (manual) pagination. I know this because, once upon a time, this was a project of mine, intended as a surprise to the both of you. Then you asked the FoL to start working on a wiki, and I figured anything I could do would be duplication of that effort.
In any case, I agree with those suggesting Google Docs for the submissions. Rather than trying to keep the records separate, I further suggest you have somebody to compile all of the submissions into the same edition of the book, so you have a single mark-up that includes all of the submissions.
Authors Need Help: Brainstorming Session
Date: 2012-04-03 12:54 am (UTC)This gives you both the list and where the words/names appear.
There are concordance programs that handle multiple books, but last time I checked, those were pay software. Might have changed in the last 5 or years.
Anybody out there know more than me about this?
Thank you very much for all the enjoyment.
Judy