Hanna Fjeldheim Dale, Milada Hagen, G. Malmstrøm, Jennifer T. Fiennes, M. Høivik, V. Kristensen, Jørgen Valeur
{"title":"评估硬结和软结 \"终点\":比较患者和专家对 2280 份粪便样本的布里斯托粪便量表评分结果","authors":"Hanna Fjeldheim Dale, Milada Hagen, G. Malmstrøm, Jennifer T. Fiennes, M. Høivik, V. Kristensen, Jørgen Valeur","doi":"10.12688/f1000research.152496.1","DOIUrl":null,"url":null,"abstract":"Background Stool consistency is an important outcome measure to evaluate in the investigation of several gastrointestinal diseases. The Bristol Stool Scale (BSS) is one of the most commonly used tools for evaluation of stool consistency. BSS ranges from 1-7 and each score is assigned to a given consistency of the feces. Self-reported characterizations can differ from an expert evaluation, and the reliability of BSS is unclear. We aimed to evaluate the reliability of BSS by comparing patient scores with expert scores. Methods Patients with inflammatory bowel disease collected stool samples throughout a 3-year follow-up. The stool´s consistency was evaluated with BSS by the patients and matched with an expert score. Agreement between patient and expert scores was assessed using Cohen’s kappa. Results BSS scores from 2280 fecal samples collected from 992 patients at up to five time points were included. When all samples were compared, there was good to substantial agreement between patient and expert scores (Cohen’s weighted kappa: 0.66-0.72). When the BSS scores were simplified and categorized as 1 (scores 1-2), 2 (scores 3-5) or 3 (scores 6-7), the agreement improved slightly (Cohen’s weighted kappa: 0.73-0.77). When the scores from the first sample per patient were compared, the experts were more likely to assign higher scores compared to the patient. The proportion of the lowest assigned scores (1-2) was 12.1% for patients and 8.1% for experts. Conclusions The agreement between patient and expert BSS scores is good to substantial, especially when the BSS scores are simplified into three categories.","PeriodicalId":504605,"journal":{"name":"F1000Research","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Assessing hard and loose “endpoints”: comparison of patient and expert Bristol Stool Scale scoring of 2280 fecal samples\",\"authors\":\"Hanna Fjeldheim Dale, Milada Hagen, G. Malmstrøm, Jennifer T. Fiennes, M. Høivik, V. Kristensen, Jørgen Valeur\",\"doi\":\"10.12688/f1000research.152496.1\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Background Stool consistency is an important outcome measure to evaluate in the investigation of several gastrointestinal diseases. The Bristol Stool Scale (BSS) is one of the most commonly used tools for evaluation of stool consistency. BSS ranges from 1-7 and each score is assigned to a given consistency of the feces. Self-reported characterizations can differ from an expert evaluation, and the reliability of BSS is unclear. We aimed to evaluate the reliability of BSS by comparing patient scores with expert scores. Methods Patients with inflammatory bowel disease collected stool samples throughout a 3-year follow-up. The stool´s consistency was evaluated with BSS by the patients and matched with an expert score. Agreement between patient and expert scores was assessed using Cohen’s kappa. Results BSS scores from 2280 fecal samples collected from 992 patients at up to five time points were included. When all samples were compared, there was good to substantial agreement between patient and expert scores (Cohen’s weighted kappa: 0.66-0.72). When the BSS scores were simplified and categorized as 1 (scores 1-2), 2 (scores 3-5) or 3 (scores 6-7), the agreement improved slightly (Cohen’s weighted kappa: 0.73-0.77). When the scores from the first sample per patient were compared, the experts were more likely to assign higher scores compared to the patient. The proportion of the lowest assigned scores (1-2) was 12.1% for patients and 8.1% for experts. Conclusions The agreement between patient and expert BSS scores is good to substantial, especially when the BSS scores are simplified into three categories.\",\"PeriodicalId\":504605,\"journal\":{\"name\":\"F1000Research\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"F1000Research\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.12688/f1000research.152496.1\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"F1000Research","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.12688/f1000research.152496.1","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Assessing hard and loose “endpoints”: comparison of patient and expert Bristol Stool Scale scoring of 2280 fecal samples
Background Stool consistency is an important outcome measure to evaluate in the investigation of several gastrointestinal diseases. The Bristol Stool Scale (BSS) is one of the most commonly used tools for evaluation of stool consistency. BSS ranges from 1-7 and each score is assigned to a given consistency of the feces. Self-reported characterizations can differ from an expert evaluation, and the reliability of BSS is unclear. We aimed to evaluate the reliability of BSS by comparing patient scores with expert scores. Methods Patients with inflammatory bowel disease collected stool samples throughout a 3-year follow-up. The stool´s consistency was evaluated with BSS by the patients and matched with an expert score. Agreement between patient and expert scores was assessed using Cohen’s kappa. Results BSS scores from 2280 fecal samples collected from 992 patients at up to five time points were included. When all samples were compared, there was good to substantial agreement between patient and expert scores (Cohen’s weighted kappa: 0.66-0.72). When the BSS scores were simplified and categorized as 1 (scores 1-2), 2 (scores 3-5) or 3 (scores 6-7), the agreement improved slightly (Cohen’s weighted kappa: 0.73-0.77). When the scores from the first sample per patient were compared, the experts were more likely to assign higher scores compared to the patient. The proportion of the lowest assigned scores (1-2) was 12.1% for patients and 8.1% for experts. Conclusions The agreement between patient and expert BSS scores is good to substantial, especially when the BSS scores are simplified into three categories.