Abstracts presented at anesthesiology subspeciality and component society meetings are chosen by peer review. We assessed this process by examining selection criteria and determining interrater concordance. For the societies studied, the level of reviewer agreement ranged from poor to moderate, i.e., slightly better than by chance alone. We hypothesize that having clearer evaluation criteria, scoring systems with interval scales, and assessment based on quality can strengthen the peer review process.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1213/01.ane.0000200314.73035.4d | DOI Listing |
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!