Regarding 1) I disagree with that measuring quality is impossible; rather, there are several different metrics to doing that, and what may be impossible is selecting the "best one".

Piotrus05:11, 30 November 2009

It depends. I think having clear, subjective feedback potentially really helps, but this exact type of feedback hasn't been used in statistics or project models yet. In such a situation there is no way to measure quality factors like "article completeness", "balance", "structure", etc. The only way in which the statistics currently are able to measure quality is by looking if an edit got reverted and comparison with the editor's other contributions. This tells us very little since the revert can (as far as I see) have 17 valid direct reasons, be neutral to quality, or/and even have a negative direct effect on quality (in 17 possible ways).

Woodwalker08:07, 30 November 2009

Per Woodwalker. Metrics we could develop include:

  1. Crudely measuring things like user tags, word to cite ratios, stability, and the like.
  2. Reporting user feedback.
  3. Setting standards for articles (newly created | baseline quality | good | featured) and measuring time taken and progression rates between these.

These aren't bad by themselves. But I'm not aware of any way to calculate useful metrics for genuine quality aside from those things.

FT2 (Talk | email)11:06, 30 November 2009
 

I just ran a reader survey on five articles that I wrote or had a hand in editing. I found the results to be extremely enlightening and helpful in my work as an editor, and also as an important input to policy disputes in the project I am working on. You can see the results of the survey here.

I performed the survey by creating the survey form at www.surveymonkey.com, and attaching a link at the top of each article (see, for example, this revision of one of the articles.

As an editor, I would love a tool that I could use to develop a survey with article-specific questions, attach it to the end of an article, and analyze the responses.

In numerous other forums, I have pointed out that Wikipedia is an editor-centric, rather than reader-centric, institution. All the mechanisms and rules of behavior are designed to create cooperation of a community of editors. In this dynamic, the reader is most often shunted aside; to the extent that, when I proposed my survey, there were editors who clearly didn't want to know what their readers were thinking.

This is something that has to change if Wikipedia is to move forward, and that change will occur only when features of the editing environment support the change. That is why I think a tool like this would be invaluable, not only to me but to the entire Wikipedia weltschaum. --Ravpapa 15:09, 8 March 2010 (UTC)

Ravpapa15:09, 8 March 2010