instruction
stringlengths
10
664
context
stringlengths
1
5.66k
response
stringlengths
1
3.34k
category
stringclasses
1 value
Patient goals after incontinence procedures: does the single-incision sling satisfy them?
This study was undertaken to describe short-term postoperative achievement of subjective preoperative goals for single-incision MiniArc slings, in comparison with tension-free vaginal tape (TVT). Patients submitted to mid-urethral sling (TVT and MiniArc) procedures for stress urinary incontinence (SUI) in two centers were included in this prospective study. Before surgery, the patients completed a preoperative open-ended questionnaire, in which they described their personal outcomes goals for SUI surgery and the degree of severity of their symptoms. At the first postoperative check, they were asked to assess the degree to which their goals had been met and the degree of postoperative incontinence symptoms; their grade of satisfaction was evaluated with IIQ-7, UDI-6 and a 0-10 visual analog scale. One hundred and eight patients (TVT n=51, MiniArc n=57) were included in this study. Incontinence symptom relief and improvement of quality of life were the most commonly described preoperative goals. Six to eight weeks after surgery, 47 patients (92.1%) after TVT and 53 (92.9%) women after single-incision slings were objectively cured (P=1). After surgery, more than 90% of the patients in both groups achieved their preoperative goals. Symptom scores improved significantly and were comparable in both groups.
Our results show that self-reported achievement of preoperative goals of patients submitted to single-incision slings are comparable at the first follow-up with patients who have undergone the classic mid-urethral sling.
closed_qa
Improving stress echocardiography accuracy for detecting left circumflex artery stenosis: a new echocardiographic sign?
The accuracy and reproducibility of stress echocardiography (SE) for the detection of coronary artery lesions requires improvement, particularly in the left circumflex artery (LCx). To evaluate the feasibility and diagnostic value of a new sign: Rise of the Apical lateral wall and/or Horizontal displacement of the Apex toward the septum ("RA-HA") in apical echocardiographic views. Consecutive patients with normal left ventricular function at rest, positive SE and an indication for coronary angiography were included. SEs were analysed blindly by three independent cardiologists: two seniors (S1 and S2) and one junior (J). Of 81 patients, 58 had an exercise SE and 23 had a dobutamine SE. Significant coronary stenosis was found in 59 of 77 patients who underwent coronary angiography (76.6%). Interobserver reproducibility for the presence of RA-HA was very good between S1 and S2 (κ = 0.86), and good between S1 and J (0.67) and S2 and J (0.70). The sensitivity, specificity and positive and negative predictive values of RA-HA for the detection of significant coronary artery stenosis were, respectively, 39-41%, 83-89%, 88-92% and 29-31% for S1/S2; and 29%, 83%, 85% and 26% for J. To predict LCx stenosis (single or multivessel): 67-70%, 89%, 80-81% and 80-82% for S1/S2, respectively, and 50%, 89%, 75% and 74% for J.
With a short learning curve, RA-HA is easily diagnosed with a very good interobserver reproducibility. It has high specificity and PPV for the detection of a coronary artery stenosis, particularly in the LCx artery, during exercise or dobutamine SE.
closed_qa
Should computed tomography coronary angiography be aborted when the calcium score exceeds a certain threshold in patients with chest pain?
There is ongoing debate about whether a computed tomography coronary angiography (CTCA) should be aborted when the calcium score (CS) exceeds a certain threshold in patients with chest pain. The aim of this study was to discover whether specific "cutpoints" regarding coronary artery CS could be determined to predict severe coronary stenoses assessed by CTCA, thus identifying patients amenable to an invasive diagnostic approach. 294 consecutive patients with chest pain of uncertain cause who were referred for non-invasive diagnostic CTCA were included. Subjects underwent Agatston CS and CTCA using current 64-slice technology. Severe coronary stenoses were noted in 75 of 294 (25.1%) patients on CTCA. A very high prevalence of severe coronary stenoses was found in patients with CS ≥ 400 (87.0%). The CS had area under the ROC curve 0.86 to predict severe coronary stenoses on CTCA. The best discriminant cut-off point was CS ≥ 400 (sensitivity of 55.3%, specificity of 93.5, positive predictive value of 85.8%, negative predictive value of 84.0%). Multivariable logistic regression analysis controlling for traditional risk factors showed CS ≥ 400 remained an independent predictor of severe coronary stenoses on CTCA (OR 14.553, 95% confidence interval 4.043 to 52.384, p<0.001).
CS can be used as a "gatekeeper" to CTCA in patients with chest pain. Due to the very high prevalence of severe coronary stenoses in patients with CS ≥ 400, further evaluation with CTCA is not warranted as these patients should be referred to invasive coronary angiography, avoiding the repeated exposure to ionizing radiation and iodinated contrast.
closed_qa
Are there symptom differences in patients with coronary artery disease presenting to the ED ultimately diagnosed with or without ACS?
Symptoms are compared among patients with coronary artery disease (CAD) admitted to the emergency department with or without acute coronary syndrome (ACS). Sex and age are also assessed. A secondary analysis from the PROMOTION (Patient Response tO Myocardial Infarction fOllowing a Teaching Intervention Offered by Nurses) trial, an multicenter randomized controlled trial, was conducted. Of 3522 patients with CAD, at 2 years, 565 (16%) presented to the emergency department, 234 (41%) with non-ACS and 331 (59%) with ACS. Shortness of breath (33% vs 25%, P = .028) or dizziness (11% vs 3%, P = .001) were more common in non-ACS. Chest pain (65% vs 77%, P = .002) or arm pain (9% vs 21%, P = .001) were more common in ACS. In men without ACS, dizziness was more common (11% vs 2%; P = .001). Men with ACS were more likely to have chest pain (78% vs 64%; P = .003); both men and women with ACS more often had arm pain (men, 19% vs 10% [P = .019]; women, 26% vs 13% [P = .023]). In multivariate analysis, patients with shortness of breath (odds ratio [OR], 0.617 [confidence interval [CI], 0.410-0.929]; P = .021) or dizziness (OR, .0311 [CI, 0.136-0.708]; P = .005) were more likely to have non-ACS. Patients with prior percutaneous coronary intervention (OR, 1.592 [CI, 1.087-2.332]; P = .017), chest pain (OR, 1.579 [CI, 1.051-2.375]; P = .028), or arm pain (OR, 1.751 [CI, 1.013-3.025]; P<.042) were more likely to have ACS.
In patients with CAD, shortness of breath and dizziness are more common in non-ACS, whereas prior percutaneous coronary intervention and chest or arm pain are important factors to include during ACS triage.
closed_qa
Does vascular endothelial growth factor participate in uterine myoma growth stimulation?
Peptide growth factors play a role in the rebuilding of extracellular matrix in the course of leiomyoma growth, and exert a regulative effect on the cell only when they bind with a specific membrane receptor and transmit a signal into the cell. A high content of certain peptide growth factors and their receptors in leiomyoma suggests that in the course of the tumour growth hyperstimulation of cells takes place. A combined action of various peptide growth factors causes an amplification of signal paths in cells, inducing gene expression of proteins responsible for cell division and changes of metabolism. We therefore decided to evaluate the amounts and expression of VEGF, their receptor and mRNA levels. Studies were performed on human myometrium and uterine leiomyomas of various weights (small: i.e. less than 10 g, and large: i.e. more than 100 g). Expression and content of VEGF-A, D and VEGF R-1, R-2 were analysed with Western blot and ELISA methods, respectively. The RT-PCR method was used to determine VEGF mRNA levels. Our immunoblotting studies and immunoenzymatic assay, as well as RT-PCR technique, did not detect significant differences in the expression of VEGFs and their receptors in control myometrium and in uterine leiomyomas.
The increase in the amount of some peptide growth factors, especially FGFs and IGF-I, in large leiomyomas without any change in VEGF content means a decrease in the proportional relationship of the latter to other growth factors. Stimulation of extracellular matrix formation seems stronger than angiogenesis during myoma growth.
closed_qa
Is urolithiasis in children associated with obesity or malnutrition?
Although it is known that obesity predisposes to urolithiasis, a tendency for malnutrition in children with urolithiasis owing to recurrent urinary infections and abdominal pain also makes sense. In this study, we aimed to determine the nutritional status of infants and children with urolithiasis, and to observe whether obesity or malnutrition is more prevalent in that population. One hundred eighty-seven children aged 4 months to 17 years (mean, 4.9 ± 4.4 years) with urolithiasis, and 278 age- and sex-matched children without any chronic diseases were included. Anthropometric evaluations, including weight and height standard deviation score (SDS), body mass index, and triceps and subscapular skinfold thickness (SFT), were performed. Mean weight SDSs of the patients was statistically lower than that of the control subjects (P<.0001). Malnutrition rate was statistically higher in the patients with urolithiasis when evaluated according to weight SDS and percentiles of body mass index and SFT. When the age factor was taken into account, the percentage of malnutrition, determined by the percentiles of triceps and subscapular SFT measurements, was found to be higher in children younger than 2 years. Short stature was more prevalent in older children.
Malnutrition among children with urolithiasis is not as rare as thought previously. A careful anthropometric evaluation should be included in the clinical assessment of those children.
closed_qa
Is myocutaneous flap alone sufficient for reconstruction of chest wall osteoradionecrosis?
This study was carried out to determine whether the myocutaneous flap, alone, is sufficient to reconstruct a chest wall defect after osteoradionecrosis and provide satisfactory stability to the chest wall. This study involved five patients who were subjected to post-mastectomy radiotherapy as a treatment for breast cancer. Excision of the ulcer and all the necrotic ribs, with preservation of the parietal pleura and reconstruction with the latissimus dorsi flap, was done without the use of either an artificial prosthesis or autologous rib to reconstruct the chest wall defect. Clinical and radiological follow-up showed no complications regarding respiratory impairment or pleural complications.
The use of myocutaneous flap in patients with chest wall defect following osteoradionecrosis is satisfactory to cover the chest wall defect and provide satisfactory stability to the chest wall.
closed_qa
Does abdominoplasty have a positive influence on quality of life, self-esteem, and emotional stability?
In a previous prospective study, the authors evaluated the quality of life in patients undergoing aesthetic surgery. In this survey, the authors split up the operative indication and analyzed quality of life, self-esteem, and emotional stability after abdominoplasty alone. Sixty-three patients participated in the study. The testing instrument consisted of a self-developed questionnaire to collect demographic and socioeconomic data and a postoperative complication questionnaire developed especially for abdominoplasties. In addition, a standardized self-assessment test on satisfaction and quality of life (Questions on Life Satisfaction), the Rosenberg Self -Esteem Questionnaire, and the Freiburg Personality Inventory were used. Significantly increasing values in some items of the standardized self-assessment test on satisfaction and quality of life were found: sum scores of the General Life Satisfaction showed a significant improvement (p = 0.004) and the scores of the items housing/living conditions (p = 0.000) and family life/children (p = 0.000). Within the Satisfaction with Health module, a significant improvement in the items mobility (p = 0.02) and independence from assistance (p = 0.01) was found. Values in the module Satisfaction with Appearance (Body Image) increased regarding satisfaction with the abdomen (p = 0.001). Over 84 percent were very satisfied with the aesthetic result, 93.4 percent would undergo the same treatment again, and 88.9 percent would further recommend the operation. Data revealed that participants' self-esteem was very high and their emotional stability was very well balanced.
This study demonstrates that abdominoplasty increases most aspects of quality of life, particularly family life, living conditions, mobility, and independency from assistance. Also, patient self-esteem and emotional stability ratings are very high postoperatively.
closed_qa
Does epidural clonidine improve postoperative analgesia in major vascular surgery?
The prospective, single-blinded study involved 60 patients randomised into three groups (20 patients each): Group BM- bupivacaine 0.125% and morphine 0.1 mg/ml; Group BC-bupivacaine 0.125% and clonidine 5 μg/ml; Group MC-morphine 0.1 mg/ml and clonidine 5 μg/ml continuously infused at 5 ml/h. The quality and duration of the analgesia measured by the Visual Analogue Scale (VAS) at rest and on movement, additional analgesia requirements, sedation scores, haemodynamic parameters and side effects (respiratory depression, motor block, toxic effects, nausea and pruritus) were recorded. The average VAS scores at rest and on movement were significantly lower in Group MC at two, six and 24 hours following the start of epidural infusion (P<0.05). The duration of the analgesic effect after finishing the epidural infusion was significantly longer in Group MC (P<0.05). Patients from Group MC were intubated longer. Additional analgesia consumption, sedation scores and haemodynamic profiles were similar in all three groups. Pruritus was more frequent in morphine groups (P<0.05), but other side effects were similar in all three groups.
Under study conditions, clonidine added to morphine, not 0.125% bupivacaine, provided significantly better pain scores at two, six and 24 hours following the start of epidural infusion and the longest-lasting analgesia following the discontinuation of epidural infusion. However, patients from the Group MC were mechanically ventilated longer than patients from other two groups. Continuous monitoring of the patient is necessary after the administration of clonidine for epidural analgesia.
closed_qa
The vein collar: an anastomotic servant or a patency promoter?
Primary patency regarding the use of vein collar were re-analyzed in 345 patients from SCAMICOS with Kaplan-Meier life-table technique and Cox proportional hazards regression in a counting process notation to evaluate any interaction between time-period and the effect of a vein collar on the primary patency rate. No overall effect on primary patency of a vein collar at the distal anastomosis was found irrespective of the site anastomosis. However, during the first 30 days of follow-up the primary patency among the femoro-crural bypasses was 0.87 (0.79-0.95) and 0.72 (0.63-0.83) with and without vein collar respectively. The interaction between vein collar and time-period was not statistically significant (P=0.070) and neither was the Score test for the whole interaction analysis (P=0.091) for the patients with anastomosis to the crural arteries. No such initial differences were found for the patients with anastomosis to the popliteal artery below-knee.
A clinically relevant but not statistically significant better primary patency during the first 30 days was found for patients with PTFE-bypass to the crural arteries with a vein collar at the distal anastomosis. There were no long-term advantages of the vein collar irrespective of the location of the anastomosis.
closed_qa
Tibial angioplasty in diabetic patients: should all vessels be treated?
We retrospectively reviewed all consecutive diabetic patients with tibial disease with no concomitant proximal lesions who were treated by angioplasty. Among 82 patients with isolated tibial disease 48 patients were selected. All patients had to have more than one diseased tibial vessel that can be treated by angioplasty. Group A patients (N.=25) had only one tibial vessel treated while group B patients (N.=23) had more than one tibial vessel treated. We compared both groups with respect to patients' characteristics, lesion morphology, and limb salvage rate. Lesion morphology was worse in group A than B: anterior tibial artery showed more long lesions (17 vs. 8), more multiple lesions (22 vs. 11), and peroneal artery showed more long lesions (23 vs. 10), more multiple lesions (24 vs. 12), and more occlusions (18 vs. 10). Limb salvage rate at 12 months was similar (91%) in both groups. There were 5 complications in each group.
The lesion morphology was worse in group A. Simpler lesions in group B motivated performing more than one vessel angioplasty. There was no difference in the limb salvage rate in the medium term among both groups. Additional vessels angioplasty in less diseased arteries was not associated with substantial additional morbidity.
closed_qa
Gender and stroke lateralization: factors of functional recovery after the first-ever unilateral stroke?
The goal of this prospective study was to evaluate gender differences in rehabilitation outcome in patients after the first-ever unilateral stroke. A total of eighty right-handed patients were prospectively enrolled, 35 (44%) women, and 45 (56%) men. A degree of neurological deficit was quantified by the National Institutes of Health Stroke Scale. Functional outcome was assessed by the Motor Status Scale, Chedoke Arm and Hand Activity Inventory, Rivermead Mobility Index and Barthel Index. At the time of hospital admission there was no significant gender difference in clinical stroke severity. At discharge, we registered significantly better motor and functional recovery in men compared to women. Further, we found significantly better rehabilitation outcome in women with stroke in dominant left hemisphere (LH) than in women with stroke in subdominant right hemisphere (RH). Conversely, men with stroke in subdominant RH had significantly better rehabilitation outcome than men with stroke in dominant LH. Using a multivariate analysis we have found that men with stroke in RH had significantly higher probability to reach not only high response in mobility, but also more autonomy in ADL. The frequency of stroke in LH was significantly higher in both genders aged less than 51 years, as well as in women, while the frequency of stroke in RH was significantly higher in men.
This paper places particular emphasis on substantial gender-based differences in functional recovery of patients with their first-ever unilateral stroke.
closed_qa
Experiences from a randomised, controlled trial on cycling to school: does cycling increase cardiorespiratory fitness?
The objective of the present study was to investigate the effect of a 12-week randomised controlled cycling-to-school trial on cardiorespiratory fitness. A total of 53 10- to 13-year-old children from one public school were included. The children were randomised into either a cycling group or a control group. The cycling group was encouraged to cycle to and from school each day during a period of 12 weeks. Peak oxygen consumption (VO(2peak)) and anthropometrical data (weight and height) were measured at baseline and at the end of the 12-week period. No significant differences were observed in VO(2peak) change over the 12-week period between the cycling group and the control group (49.7 ml O(2)/min/kg vs. 50.6 ml O(2) /min/kg; effect size=-0.13, F=0.495, p=0.486). Within the intervention group, 69.2% (95% CI 50.1-88.2) started cycling, and within the control group 40.8% (95% CI 20.9-60.5) started cycling. Given that several children in both groups (intervention and control) started cycling to school, re-analyses were conducted between those starting cycling and those not starting cycling. At follow up, a significant difference between those starting cycling and those who did not starting cycling was observed in VO(2peak) (51.7 ml O(2)/min/kg vs. 47.9 ml O(2)/min/kg; effect size=0.49, F=8.145, p=0.007), after adjustment for baseline scores, gender and age.
This study indicates that cycling to school improves cardiorespiratory fitness.
closed_qa
Is there equity in use of healthcare services among immigrants, their descendents, and ethnic Danes?
Legislation in Denmark explicitly states the right to equal access to healthcare. Nevertheless, inequities may exist; accordingly evidence is needed. Our objective was to investigate whether differences in healthcare utilisation in immigrants, their descendents, and ethnic Danes could be explained by health status, socioeconomic factors, and integration. We conducted a nationwide survey in 2007 with 4952 individuals aged 18-66 comprising ethnic Danes; immigrants from the former Yugoslavia, Iran, Iraq, Lebanon, Pakistan, Somalia, Turkey; and Turkish and Pakistani descendents. Data were linked to registries on healthcare utilisation. Using Poisson regression models, contacts to hospital, emergency room (ER), general practitioner (GP), specialist in private practice, and dentist were estimated. Analyses were adjusted for health symptoms, sociodemographic factors, and proxies of integration. In adjusted analyses, immigrants and their descendents had increased use of ER (multiplicative effect 1.19-5.02 dependent on immigrant and descendent group) and less frequent contact to dentist (multiplicative effect 0.04-0.80 dependent on the group). For hospitalisation, GP, and specialist doctor, physical health symptoms had positive but different explanatory effects within groups; however, most immigrant and descendent groups had increased use of services compared with that of ethnic Danes. Socioeconomic factors and integration had no systematic effect on the use in the different groups.
The Danish healthcare system seems responsive to health across different population groups. We found no systematic pattern of inequity in use of free-of-charge healthcare services, but for dentists, who require co-payment, we found inequity among immigrants and descendents compared with ethnic Danes.
closed_qa
Socially differentiated cardiac rehabilitation: can we improve referral, attendance and adherence among patients with first myocardial infarction?
From 1 September 2002 to 31 December 2005, 388 first-incidence MI patients ≤75 years were hospitalised. Register check for newly hospitalised MI patients, screening interview, and systematic referral were conducted by a project nurse. Patients were referred to a standard rehabilitation programme (SRP). If patients were identified as socially vulnerable, they were offered an extended version of the rehabilitation programme (ERP). Excluded patients were offered home visits by a cardiac nurse. Concordance principles were used in the individualised programme elements. Adherence was registered until the 1-year follow up. 86% were referred to the CR. A large share of elderly patients and women were excluded. The attendance and adherence rates were 80% and 71%, respectively among all hospitalised patients. Among referred patients, the attendance rate was 93%. Patients were equally distributed to the SRP and the ERP. No inequality was found in attendance and adherence among referred patients.
It seems possible to overcome unequal referral, attendance, and adherence in cardiac rehabilitation by organisation of systematic screening and social differentiation.
closed_qa
Removal of industry-sponsored formula sample packs from the hospital: does it make a difference?
Most US hospitals distribute industry-sponsored formula sample packs. No research has examined outcomes associated with sample pack removal as part of a hospital intervention to eliminate sample distribution postpartum. To examine prospectively hospital-based and breastfeeding outcomes associated with removal of industry-sponsored formula sample packs from the hospital. We enrolled mothers postpartum at Cooper University Hospital, an urban New Jersey hospital, in 2009-2010. For the first 6 months, all women received industry-sponsored formula samples packs (control group); for the next 6 months, all postpartum women received hospital-sponsored bags with no formula at source (intervention group). Research assistants blinded to the design called subjects weekly for 10 weeks to determine feeding practices. We enrolled 527 breastfeeding women (284 control; 243 intervention). At 10 weeks postpartum, 82% of control and 36% of intervention women (P<.001) reported receiving formula in the "diaper discharge bag." Kaplan-Meyer curves for any breastfeeding showed the intervention was associated with increased breastfeeding (P = .03); however, exclusive breastfeeding was not significantly different between intervention and controls (P = .46). In post hoc analysis, receiving no take-home formula in bottles from the hospital was associated with increased exclusive breastfeeding in control (P = .02) and intervention (P = .03) groups at 10 weeks.
Although the hospital-branded replacement contained no formula at source, many women reported receiving bottles of formula from the hospital. Change in practice to remove industry-sponsored formula sample packs was associated with increased breastfeeding over 10 weeks, but the intervention may have had a greater impact had it not been contaminated.
closed_qa
Tissue banking in a regional hospital: a promising future concept?
Vital tissue provided by fresh frozen tissue banking is often required for genetic tumor profiling and tailored therapies. However, the potential patient benefits of fresh frozen tissue banking are currently limited to university hospitals. The objective of the present pilot study--the first one in the literature--was to evaluate whether fresh frozen tissue banking is feasible in a regional hospital without an integrated institute of pathology. Patients with resectable breast and colon cancer were included in this prospective study. Both malignant and healthy tissue were sampled using isopentan-based snap-freezing 1 h after tumor resection and stored at -80 °C before transfer to the main tissue bank of a University institute of pathology. The initial costs to set up tissue banking were 35,662 US$. Furthermore, the running costs are 1,250 US$ yearly. During the first 13 months, 43 samples (nine samples of breast cancer and 34 samples of colon cancer) were collected from 41 patients. Based on the pathology reports, there was no interference with standard histopathologic analyses due to the sample collection.
This is the first report in the literature providing evidence that tissue banking in a regional hospital without an integrated institute of pathology is feasible. The interesting findings of the present pilot study must be confirmed by larger investigations.
closed_qa
Improving quality of medical treatment and care: are surgeons' working conditions and job satisfaction associated to patient satisfaction?
Over the last decades, surgeons, researchers, and health administrators have been working hard to define standards for high-quality treatment and care in Surgery departments. However, it is unclear whether patients' perceptions of medical treatment and care are related and affected by surgeons' perceptions of their working conditions and job satisfaction. The aim of this study was to evaluate patients' satisfaction in relation to surgeons' working conditions. A cross-sectional survey with 120 patients and 109 surgeons working in Surgery hospital departments was performed. Surgeons completed a survey evaluating their working conditions and job satisfaction. Patients assessed quality of medical care and treatment and their satisfaction with being a patient in this department. Seventy percent of the patients were satisfied with performed surgeries and services in their department. Surgeons' job satisfaction and working conditions rated with moderate scores. Bivariate analyses showed correlations between patients' satisfaction and surgeons' job satisfaction and working conditions. Strongest correlations were found between kindness of medical staff, treatment outcome and overall patient satisfaction.
This study demonstrates strong associations between surgeons' working conditions and patient satisfaction. Based on these findings, hospital managements should improve work organization, workload, and job resources to not only improve surgeons' job satisfaction but also quality of medical treatment and patient satisfaction in Surgery departments.
closed_qa
Can amino acid carbon isotope ratios distinguish primary producers in a mangrove ecosystem?
The relative contribution of carbon from terrestrial vs. marine primary producers to mangrove-based food webs can be challenging to resolve with bulk carbon isotope ratios (δ(13)C). In this study we explore whether patterns of δ(13)C values among amino acids (AAs) can provide an additional tool for resolving terrestrial and marine origins of carbon. Amino acid carbon isotope ratios (δ(13)C(AA)) were measured for several terrestrial and marine primary producers in a mangrove ecosystem at Spanish Lookout Caye (SLC), Belize, using gas chromatography-combustion-isotope ratio mass spectrometry. The δ(13)C values of essential amino acids (δ(13)C(EAA)) were measured to determine whether they could be used to differentiate terrestrial and marine producers using linear discriminant analysis. Marine and terrestrial producers had distinct patterns of δ(13)C(EAA) values in addition to their differences in bulk δ(13)C values. Microbial mat samples and consumers (Crassostrea rhizophorae, Aratus pisonii, Littoraria sp., Lutjanus griseus) were most similar to marine producers. Patterns of δ(13)C(EAA) values for terrestrial producers were very similar to those described for other terrestrial plants.
The findings suggest that δ(13)C(EAA) values may provide another tool for estimating the contribution of terrestrial and marine sources to detrital foodwebs. Preliminary analyses of consumers indicate significant use of aquatic resources, consistent with other studies of mangrove foodwebs.
closed_qa
Is a cementless dual mobility socket in primary THA a reasonable option?
Dislocation after THA continues to be relatively common. Dual mobility sockets have been associated with low dislocation rates, but it remains unclear whether their use in primary THA would not introduce additional complications.QUESTIONS/ We therefore asked whether a current cementless dual mobility socket (1) reduced the dislocation rate after primary THA, (2) provided a pain-free and mobile hip, and (3) provided durable radiographic fixation of the acetabular component without any unique modes of failure. We retrospectively reviewed 168 patients who underwent primary THA using a dual mobility socket between January 2000 and June 2002. The average age at surgery was 67 years. We assessed the rate of dislocation, hip function, and acetabular fixation on serial radiographs. Of the 168 patients, 119 (71%) had clinical and radiographic evaluation at a minimum of 5 years (mean, 6 years; range, 5-8 years). A long-neck option left the base of the Morse taper uncovered in 53 hips. Four patients underwent revision for dislocation between the femoral head and the mobile insert (intraprosthetic dislocation) at a mean 6 years; all four revisions occurred among the 53 hips with an incompletely covered Morse taper.
A current cementless dual mobility socket was associated with a pain-free and mobile hip and durable acetabular fixation without dislocations if the long-neck option was not used. However, intraprosthetic dislocation related to contact at the femoral neck to mobile insert articulation required revision in four hips. Surgeons should be aware of this specific complication.
closed_qa
Are obstetrician-gynecologists satisfied with their maternal-fetal medicine consultants?
To survey generalist obstetrician-gynecologists about their satisfaction with and patterns of referral to maternal-fetal medicine (MFM) specialists. A survey was sent three times to 1030 randomly selected American Congress of Obstetricians and Gynecologists members across the country, and results were tabulated. A total of 516 surveys (50%) were returned; 68% of respondents were satisfied (S) with available MFM services and 31% were not satisfied (Not S). S and Not S respondents were similar with respect to age, gender, years in practice, type of practice, hours worked per week, proximity to MFM specialists, number of deliveries per year, and level of nursery in their hospital. Reasons for dissatisfaction included: MFM specialist not readily available (49%), during the day (26%), at night (35%), or on weekends (36%); MFM specialist unwilling to take care of hospitalized patients (26%); or MFM specialist does only ultrasound, chorionic villus sampling, and amniocentesis (32%). Although some generalists do not consult MFM specialists frequently, the majority of both S and Not S respondents would request an MFM consult or comanagement for 26 of 38 specific maternal, fetal, and obstetric diagnoses/complications.
The majority of obstetrician-gynecologists are satisfied with their MFM support. The dissatisfaction expressed by 31% of generalists might be ameliorated if individual MFM specialists increased their availability and/or broadened their scope of practice.
closed_qa
Depressive symptoms in older people with metabolic syndrome: is there a relationship with inflammation?
To investigate if there is a higher prevalence of depressive symptoms in older people with metabolic syndrome (MetS) compared with those without and whether dedpressive symptoms are independently associated to MetS and its single components and to the inflammatory markers. Physical parameters, standard blood analytes, high sensitivity C-reactive protein (hsCRP) and erythrocyte sedimentation rate (ESR) were assessed. Fifteen-item Geriatric Depression Scale and mini mental state examination (MMSE) were administered. One hundred thirty-three subjects were enrolled. MetS patients (57) exhibited higher prevalence of depressive symptoms (p < 0.0001), worse cognitive function (p < 0.0001), and higher levels of ESR and hsCRP were higher (p < 0.0001). The univariate analysis showed a linear strong correlation of depressive symptoms (p < 0.0001) with the MMSE score (r = -0.422), body mass index (r = 0.414), MetS (r = 0.582), number of MetS components (r = 0.663), fasting blood glucose (r = 0.565), ESR (r = 0.565), hsCRP (r = 0.745), central obesity (r = 0.269; p = 0.002), and high-density lipoprotein cholesterol (r = -0.241; p = 0.005). However, the multivariate analysis showed that only age (B = -0.093; p = 0.032), MetS (B = 1.446; p = 0.025), fasting blood glucose (B = 0.039; p = 0.005), and hsCRP (B = 7.649; p < 0.0001) were independently associated with depressive symptoms.
MetS and inflammation are independently associated with depressive symptoms in older people. Inflammation may explain cognitive decline too. Further investigations are needed to better understand the direction of these associations and to determine whether these can be reversible.
closed_qa
Cyclic Vomiting Syndrome (CVS): is there a difference based on onset of symptoms--pediatric versus adult?
Cyclic Vomiting Syndrome (CVS) is a well-recognized functional gastrointestinal disorder in children but its presentation is poorly understood in adults. Genetic differences in pediatric-onset (presentation before age 18) and adult-onset CVS have been reported recently but their clinical features and possible differences in response to therapy have not been well studied. This was a retrospective review of 101 CVS patients seen at the Medical College of Wisconsin between 2006 and 2008. Rome III criteria were utilized to make the diagnosis of CVS. Our study population comprised of 29(29%) pediatric-onset and 72 (71%) adult-onset CVS patients. Pediatric-onset CVS patients were more likely to be female (86% vs. 57%, p = 0.005) and had a higher prevalence of CVS plus (CVS + neurocognitive disorders) as compared to adult-onset CVS patients (14% vs. 3%, p = 0.05). There was a longer delay in diagnosis (10 ± 7 years) in the pediatric-onset group when compared to (5 ± 7 years) adult-onset CVS group (p = 0.001). Chronic opiate use was less frequent in the pediatric-onset group compared to adult-onset patients (0% vs. 23%, p = 0.004). Aside from these differences, the two groups were similar with regards to their clinical features and the time of onset of symptoms did not predict response to standard treatment. The majority of patients (86%) responded to treatment with tricyclic antidepressants, anticonvulsants (topiramate), coenzyme Q-10, and L-carnitine. Non-response to therapy was associated with coalescence of symptoms, chronic opiate use and more severe disease as characterized by longer episodes, greater number of emergency department visits in the year prior to presentation, presence of disability and non-compliance on univariate analysis. On multivariate analysis, only compliance to therapy was associated with a response. (88% vs. 38%, Odds Ratio, OR 9.6; 95% Confidence Interval [CI], 1.18-77.05).
Despite reported genetic differences, the clinical features and response to standard therapy in pediatric- and adult-onset CVS were mostly similar. Most patients (86%) responded to therapy and compliance was the only factor associated with a response.
closed_qa
The UK Clinical Aptitude Test: is it a fair test for selecting medical students?
The United Kingdom Clinical Aptitude Test (UKCAT) is designed to increase diversity and fairness in selection to study medicine.AIM: The aim of this study is to determine if differences in: access to support and advice, in modes of preparation, type of school/college attended, level of achievement in mathematics, gender and age influence candidate performance in the UKCAT and thereby unfairly advantage some candidates over others. Confidential, self-completed, on-line questionnaire of applicants to study on an undergraduate medical degree course who had taken the UKCAT in 2010. Differentials in access to support and advice, in modes of preparation, type of school/college attended, in level of achievement in mathematics, gender and age were found to be associated with candidate performance in the UKCAT.
The findings imply that the UKCAT may disadvantage some candidate groups. This inequity would likely be improved if tutors and career advisors in schools and colleges were more informed about the UKCAT and able to offer appropriate advice on preparation for the test.
closed_qa
Do combined alternating sessions of 1540 nm nonablative fractional laser and percutaneous collagen induction with trichloroacetic acid 20% show better results than each individual modality in the treatment of atrophic acne scars?
There have been no well-controlled studies evaluating the efficacy of combining 1540 nm nonablative fractional laser with percutaneous collagen induction (PCI) and trichloroacetic acid (TCA) 20% in the treatment of atrophic acne scars. We hypothesized that combined alternating sessions of both modalities would show better results than each individual modality. Thirty-nine patients with post acne atrophic scars were included in this study. Patients were randomly equally divided into three groups; group 1 was subjected to six sessions of PCI combined with TCA 20% in the same session, group 2 was subjected to six sessions of 1540 nm fractional laser and group 3 was subjected to combined alternating sessions of the previously mentioned two modalities. Scar severity scores improved by a mean of 59.79% (95% CI 47.38-72.21) (p<0.001) in group 1, a mean of 61.83% (95% CI 54.09-69.56) (p<0.001) in group 2 and a mean of 78.27% (95% CI 74.39-82.15) (p<0.001) in group 3. The difference in the degree of improvement was statistically significant when comparing the three groups using ANOVA test (p = 0.004).
The current work recommends combining 1540 nm nonablative fractional laser in alternation with PCI and TCA 20% in the treatment of atrophic acne scars.
closed_qa
Is it time to abandon paper?
A multidisciplinary primary care clinic in Sydney, Australia, was planning to use electronic questionnaires to measure patient-reported outcomes. Semi-structured interviews with 20 patients were undertaken to explore, among other things, practical issues regarding different questionnaire formats. The response rates and costs of email versus postal invitations were also evaluated. Compared with postal invitations, email invitations offered a cost-effective and practical alternative, with a greater proportion of patients volunteering for an interview. Assuming the interface is well-designed and user-friendly, many patients were happy to use the Internet to answer questionnaires. Most patients thought alternate formats should also be offered. Patients discussed advantages and disadvantages of the Internet format. Although more younger patients and females had given the clinic an email address; both sexes, and young and old patients, expressed strong preferences for either wanting or not wanting to use the Internet.
Researchers should consider using email invitations as a cost-effective first-line strategy to recruit patients to participate in health services research. Internet questionnaires are potentially cheaper than paper questionnaires, and the format is acceptable to many patients. However, for the time being, concurrent alternate formats need to be offered to ensure wider acceptability and to maximize response rates.
closed_qa
Three-dimensional conformal brachytherapy boost in locally recurrent or residual cervical carcinoma: does it impact clinical outcome?
Fourteen consecutive patients with recurrent or residual cervical cancer who were treated with interstitial brachytherapy as a boost were included in the study. All patients received 50.4 Gy external radiation (EBRT) to whole pelvis with conformal technique to reduce the dose to bowel. The clinical target volume (CTV) and organs at risk were contoured on CT scan with gold seeds being a surrogate marker of initial tumor extent implanted before commencing treatment. The median dose of prescription was 10.5-12Gy in 3 fractions. Dose volume histogram was calculated to evaluate the dose that covers 100% and 90% of the target volume and dose to the bladder, rectum and bowel (2 mL, 1 mL volume). The median follow-up was 12 months (range 6-18). The doses to CTV (D90, D100) ranged from 1141 to 2014 cGy, and 585 to 969 cGy, respectively. The mean cumulative 2-mL rectal, bladder and bowel doses were 66.70, 73.15 and 61.01Gy, respectively. Rectal toxicity of grade 2 or more had a strong correlation with the dose delivered (Spearman's correlation, 0.950). The local control rate at one year was 92% with failure seen in one patient only.
Conformal EBRT supplemented with 3D-IBT seems to be a practical and appropriate approach to give the most optimal therapeutic benefit with the least side-effects in postoperative recurrent and residual cervical cancer patients.
closed_qa
Beyond early intervention: can we adopt alternative narratives like 'Woodshedding' as pathways to recovery in schizophrenia?
Significant numbers of those developing a first episode of psychosis are on a path to a persisting and potentially life long condition. Constituting the schizophrenia spectrum disorders, such conditions demand the particular qualities and attitudes inherent within recovery-based practice. This paper explores some of these qualities and attitudes by examining the tension between a traditional 'clinical' narrative used by many health providers and a 'human' narrative of users of services and their families. We draw out key features and constructs of recovery practice as they relate to the EI paradigm. These include: woodshedding, turning points, discontinuous improvement models, therapeutic optimism, gradualism and narratives of story telling. We also highlight the role of family members and other close supporters and believe their potential contribution requires greater consideration.
The early intervention (EI) paradigm can resonate and indeed offer a stronghold for recovery-based practice where traditional mental health services have sometimes struggled. Conversely, failure of caregivers to provide such an approach in the early phase of illness can cause unnecessary and sometimes disastrous consequences.
closed_qa
Do premorbid impairments predict emergent 'prodromal' symptoms in young relatives at risk for schizophrenia?
Individuals at risk for developing schizophrenia (SZ) in the future frequently exhibit subtle behavioural and neurobiological abnormalities in their childhood. A better understanding of the role of these abnormalities in predicting later onset of 'prodromal' symptoms or psychosis may help in early identification of SZ. In an ongoing prospective follow-up study of young genetically at-risk relatives of patients with SZ, we studied the prevalence of problems in premorbid social adjustment and childhood psychopathology and examined their relationship with the presence and progression of 'prodromal' symptoms of SZ. Growth curve analyses showed that 'prodromal' symptoms, as measured by the Scale of 'Prodromal' Symptoms, increased during follow-up. Premorbid maladjustment and childhood behavioural disturbances were cross-sectionally correlated broadly with 'prodromal' symptomatology scores. Longitudinal analyses revealed that behavioural disturbances, but not childhood maladjustment at baseline, significantly predicted increases in 'prodromal' symptomatology during the 2-year study period.
Premorbid behavioural disturbance and maladjustment may predict the later emergence of 'prodromal' symptoms. 'Prodromal' symptoms in young at-risk relatives may define a subgroup worthy of follow-up into the age of risk for psychosis in order to cost-effectively characterize the predictors of psychotic symptoms and SZ.
closed_qa
Is there adaptation of the exocrine pancreas in wild animal?
Physiology of the exocrine pancreas has been well studied in domestic and in laboratory animals as well as in humans. However, it remains quite unknown in wildlife mammals. Roe deer and cattle (including calf) belong to different families but have a common ancestor. This work aimed to evaluate in the Roe deer, the adaptation to diet of the exocrine pancreatic functions and regulations related to animal evolution and domestication. Forty bovine were distributed into 2 groups of animals either fed exclusively with a milk formula (monogastric) or fed a dry feed which allowed for rumen function to develop, they were slaughtered at 150 days of age. The 35 Roe deer were wild animals living in the temperate broadleaf and mixed forests, shot during the hunting season and classified in two groups adult and young. Immediately after death, the pancreas was removed for tissue sample collection and then analyzed. When expressed in relation to body weight, pancreas, pancreatic protein weights and enzyme activities measured were higher in Roe deer than in calf. The 1st original feature is that in Roe deer, the very high content in pancreatic enzymes seems to be related to specific digestive products observed (proline-rich proteins largely secreted in saliva) which bind tannins, reducing their deleterious effects on protein digestion. The high chymotrypsin and elastase II quantities could allow recycling of proline-rich proteins. In contrast, domestication and rearing cattle resulted in simplified diet with well digestible components. The 2nd feature is that in wild animal, both receptor subtypes of the CCK/gastrin family peptides were present in the pancreas as in calf, although CCK-2 receptor subtype was previously identified in higher mammals.
Bovine species could have lost some digestive capabilities (no ingestion of great amounts of tannin-rich plants, capabilities to secrete high amounts of proline-rich proteins) compared with Roe deer species. CCK and gastrin could play an important role in the regulation of pancreatic secretion in Roe deer as in calf. This work, to the best of our knowledge is the first study which compared the Roe deer adaptation to diet with a domesticated animal largely studied.
closed_qa
Are perceived stress, depressive symptoms and religiosity associated with alcohol consumption?
The aim of this study was to investigate the association of perceived stress, depressive symptoms and religiosity with frequent alcohol consumption and problem drinking among freshmen university students from five European countries. 2529 university freshmen (mean age 20.37, 64.9% females) from Germany (n = 654), Poland (n = 561), Bulgaria (n = 688), the UK (n = 311) and Slovakia (n = 315) completed a questionnaire containing the modified Beck Depression Inventory for measuring depressive symptoms, the Cohen's perceived stress scale for measuring perceived stress, the CAGE-questionnaire for measuring problem drinking and questions concerning frequency of alcohol use and the personal importance of religious faith. Neither perceived stress nor depressive symptoms were associated with a high frequency of drinking (several times per week), but were associated with problem drinking. Religiosity (personal importance of faith) was associated with a lower risk for both alcohol-related variables among females. There were also country differences in the relationship between perceived stress and problem drinking.
The association between perceived stress and depressive symptoms on the one side and problem drinking on the other demonstrates the importance of intervention programs to improve the coping with stress.
closed_qa
Should risky treatments be reserved for secondary prevention?
Clinical intuition suggests that risk-reducing treatments are more beneficial for patients with greater risk of disease. This intuition contributes to our rationale for tolerating greater adverse event risk in the setting of secondary prevention of certain diseases such as myocardial infarction or stroke. However, under certain conditions treatment benefits may be greater in primary prevention, even when the treatment carries harmful adverse effect potential. We present simple decision-theoretic models that illustrate conditions of risk and benefit under which a treatment is predicted to be more beneficial in primary than in secondary prevention. The models cover a spectrum of possible clinical circumstances, and demonstrate that net benefit in primary prevention can occur despite no benefit (or even net harm) in secondary prevention.
This framework provides a rationale for extending the familiar concept of balancing risks and benefits to account for disease-specific considerations of primary vs. secondary prevention.
closed_qa
Is omentectomy mandatory in the operation for ovarian cancer?
To investigate whether omentectomy is required in the operation for ovarian cancer, in particular at the early stage. F344 nude rats were divided into two groups: one in which laparotomy and omentectomy were performed (primary omentectomy group, n=6) and one without omentectomy (n=12). Concurrently, DISS cells derived from ovarian cancer were transplanted intraperitoneally. After three weeks, the 12 rats without omentectomy were divided into two more groups: one in which the omentum was resected together with the tumor (sham operation/omentectomy group, n=6) and one without omentectomy (sham operation alone group, n=6). The survival of the sham operation alone group was shortest with a median of 35 days, while the median of the primary omentectomy group was 42 days. In the sham operation/omentectomy group, four rats survived beyond Day 90, which was significant compared with other two groups. The intraperitoneal findings in the primary omentectomy group revealed extensive disseminated foci on the mesentery and under the abdominal wall. The sham operation alone group was characterized by jaundice resulting from the compression of the biliary system at the liver hilum by the omental mass. Disseminated foci were not observed in the peritoneal cavity from the sham operation/omentectomy group.
This study suggests the possibility that the omentum has a role in capturing cancer cells and suppressing further peritoneal dissemination. Therefore, although omentectomy is rewarding if disseminated foci are present in the omentum, it is suggested that the timing of omentectomy requires reconsideration in the absence of omental metastasis.
closed_qa
Is Drosera meristocaulis a pygmy sundew?
South America and Oceania possess numerous floristic similarities, often confirmed by morphological and molecular data. The carnivorous Drosera meristocaulis (Droseraceae), endemic to the Neblina highlands of northern South America, was known to share morphological characters with the pygmy sundews of Drosera sect. Bryastrum, which are endemic to Australia and New Zealand. The inclusion of D. meristocaulis in a molecular phylogenetic analysis may clarify its systematic position and offer an opportunity to investigate character evolution in Droseraceae and phylogeographic patterns between South America and Oceania. Drosera meristocaulis was included in a molecular phylogenetic analysis of Droseraceae, using nuclear internal transcribed spacer (ITS) and plastid rbcL and rps16 sequence data. Pollen of D. meristocaulis was studied using light microscopy and scanning electron microscopy techniques, and the karyotype was inferred from root tip meristem. The phylogenetic inferences (maximum parsimony, maximum likelihood and Bayesian approaches) substantiate with high statistical support the inclusion of sect. Meristocaulis and its single species, D. meristocaulis, within the Australian Drosera clade, sister to a group comprising species of sect. Bryastrum. A chromosome number of 2n = approx. 32-36 supports the phylogenetic position within the Australian clade. The undivided styles, conspicuous large setuous stipules, a cryptocotylar (hypogaeous) germination pattern and pollen tetrads with aperture of intermediate type 7-8 are key morphological traits shared between D. meristocaulis and pygmy sundews of sect. Bryastrum from Australia and New Zealand.
The multidisciplinary approach adopted in this study (using morphological, palynological, cytotaxonomic and molecular phylogenetic data) enabled us to elucidate the relationships of the thus far unplaced taxon D. meristocaulis. Long-distance dispersal between southwestern Oceania and northern South America is the most likely scenario to explain the phylogeographic pattern revealed.
closed_qa
Monitoring performance for blood pressure management among patients with diabetes mellitus: too much of a good thing?
Performance measures that reward achieving blood pressure (BP) thresholds may contribute to overtreatment. We developed a tightly linked clinical action measure designed to encourage appropriate medical management and a marker of potential overtreatment, designed to monitor overly aggressive treatment of hypertension in the face of low diastolic BP. We conducted a retrospective cohort study in 879 Department of Veterans Affairs (VA) medical centers and smaller community-based outpatient clinics. The clinical action measure for hypertension was met if the patient had a passing index BP at the visit or had an appropriate action. We examined the rate of passing the action measure and of potential overtreatment in the Veterans Health Administration during 2009-2010. There were 977,282 established VA patients, 18 years and older, with diabetes mellitus (DM). A total of 713,790 patients were eligible for the action measure; 94% passed the measure (82% because they had a BP<140/90 mm Hg at the visit and an additional 12% with a BP ≥140/90 mm Hg and appropriate clinical actions). Facility pass rates varied from 77% to 99% (P<.001). Among all patients with DM, 197,291 (20%) had a BP lower than 130/65 mm Hg; of these, 80 903 (8% of all patients with DM) had potential overtreatment. Facility rates of potential overtreatment varied from 3% to 20% (P<.001). Facilities with higher rates of meeting the current threshold measure (<140/90 mm Hg) had higher rates of potential overtreatment (P<.001).
While 94% of diabetic veterans met the action measure, rates of potential overtreatment are currently approaching the rate of undertreatment, and high rates of achieving current threshold measures are directly associated with overtreatment. Implementing a clinical action measure for hypertension management, as the Veterans Health Administration is planning to do, may result in more appropriate care and less overtreatment.
closed_qa
Does occupational lifting and carrying among female health care workers contribute to an escalation of pain-day frequency?
The aim of the study was to investigate if different frequencies, loads and trunk postures of occupational lifting and carrying increases the risk of sub-chronic (1-30 days last 12 months) low back pain (LBP) to become persistent (>30 days last 12 months) among female health care workers. Female health care workers answered a questionnaire about occupational lifting or carrying frequency (rarely, occasionally and frequently), load (low: 1-7 kg, moderate: 8-30 kg and heavy:>30 kg) and trunk posture (upright or forward bent back), and days with LBP in 2005 and 2006. The odds ratio (OR) for developing persistent LBP in 2006 from these characteristics of occupational lifting and carrying was investigated with multi-adjusted logistic regressions among female health care workers with sub-chronic LBP (n = 2381) in 2005. Among health care workers with sub-chronic LBP, increased risk of persistent LBP was found from frequently lifting or carrying with forward bent back of moderate loads (OR: 1.63; 95% CI: 1.15-2.33) and heavy loads (OR: 1.56; 95% CI: 1.04-2.34). No increased risk for LBP to develop into a persistent condition was found for frequent lifting with upright back, frequent lifting or carrying of light loads, or occasionally lifting or carrying of any loads.
Preventive initiatives for sub-chronic LBP to develop into a persistent condition ought to focus on reducing frequent lifting and carrying of moderate and heavy loads with forward bent back.
closed_qa
Occupational solvent exposure and cognition: does the association vary by level of education?
Chronic occupational solvent exposure is associated with long-term cognitive deficits. Cognitive reserve may protect solvent-exposed workers from cognitive impairment. We tested whether the association between chronic solvent exposure and cognition varied by educational attainment, a proxy for cognitive reserve. Data were drawn from a prospective cohort of French national gas and electricity (GAZEL) employees (n = 4,134). Lifetime exposure to 4 solvent types (chlorinated solvents, petroleum solvents, benzene, and nonbenzene aromatic solvents) was assessed using a validated job-exposure matrix. Education was dichotomized at less than secondary school or below. Cognitive impairment was defined as scoring below the 25th percentile on the Digit Symbol Substitution Test at mean age 59 (SD 2.8; 88% of participants were retired at testing). Log-binomial regression was used to model risk ratios (RRs) for poor cognition as predicted by solvent exposure, stratified by education and adjusted for sociodemographic and behavioral factors. Solvent exposure rates were higher among less-educated patients. Within this group, there was a dose-response relationship between lifetime exposure to each solvent type and RR for poor cognition (e.g., for high exposure to benzene, RR = 1.24, 95% confidence interval 1.09-1.41), with significant linear trends (p<0.05) in 3 out of 4 solvent types. Recency of solvent exposure also predicted worse cognition among less-educated patients. Among those with secondary education or higher, there was no significant or near-significant relationship between any quantification of solvent exposure and cognition.
Solvent exposure is associated with poor cognition only among less-educated individuals. Higher cognitive reserve in the more-educated group may explain this finding.
closed_qa
Is it effective to perform two more prostate biopsies according to prostate-specific antigen level and prostate volume in detecting prostate cancer?
To evaluate the effectiveness of 2 more core prostate biopsy protocol in detecting the prostate cancer (PCa) by comparing 10-core prostate biopsy with 12-core according to the prostate-specific antigen (PSA) level and the prostate volume. A total of 474 men with elevated serum levels of PSA between 2.5 and 20.0 ng/mL, regardless of abnormal finding on digital rectal examination and transrectal ultrasonography, received transrectal ultrasound-guided prostate biopsies. The patients were prospectively randomized to undergo 10-core (group 1, n = 351) or 12-core (group 2, n = 123) biopsy. The PCa detection rates were assessed and compared according to the serum level of PSA and prostate volume. Of 474 men, 128 (27.0%) were diagnosed with PCa. The PCa detection rates of 10-core and 12-core biopsies were 26.4% and 28.4%, respectively (P = .378). There was no difference in cancer detection rates according to PSA level in both groups. Comparing the cancer detection rates according to the prostate volume (<40 mL and ≥ 40 mL), the patients with prostate volume ≥ 40 mL showed higher cancer detection rates in 12-core biopsy group (26.9%) compared with 10-core biopsy group (16.4%) (P<.05).
The overall cancer detection rates showed no differences in both groups. But the 12-core biopsy was a more efficient method in men with a prostate volume of ≥ 40 mL, compared to the 10-core biopsy.
closed_qa
Is routine echocardiography necessary after catheter ablation of atrioventricular nodal re-entrant tachycardia?
The aim of this study was to investigate whether pericardial effusion (PE) detected by transthoracic echocardiography (TTE) was clinically significant and whether routine echocardiography was necessary after catheter ablation of atrioventricular nodal re-entrant tachycardia (AVNRT). A total of 202 patients with AVNRT were included in the study from three centers. The patients received basic electrophysiology-guided therapy, followed by radiofrequency ablation (RFA). All patients underwent TTE before and after RFA therapy. The mean age of the study population was 46.2 ± 17.9 and 30.7% of the patients were male. Of these patients, six (3%) had postoperative PE, as detected by TTE. However, none of them had cardiac tamponade (CT). Four patients had minimal PE, while two had mild PE. Repeated TTE at one to three months showed resolved PE. No significant difference was seen among the patients with and/or without PE in terms of age, gender, the number of RFA applications, or RFA duration; however, significantly prolonged duration of fluoroscopy exposure was observed in the patients with PE.
PE was detected in 3% of the patients by TTE and associated with prolonged duration of fluoroscopy exposure. However, no patients with moderate or large PE or cardiac tamponade were found in the study. In conclusion, we suggest that TTE should only be performed in the presence of clinical indications following ablation of AVNRT.
closed_qa
Does obesity modify the association of supplemental folic acid with folate status among nonpregnant women of childbearing age in the United States?
Obesity is associated with an increased risk of having a pregnancy affected by a neural tube defect (NTD). It is not clear whether the amount of folic acid required by obese women to protect against NTDs is the same as that for nonobese women. We analyzed data from the National Health and Nutrition Examination Survey, representative of the noninstitutionalized civilian U.S. population, to assess whether body mass index (BMI; normal weight, overweight, and obese categories) modified the association between supplemental folic acid intake and folate status. We estimated the geometric mean concentration among nonpregnant women of childbearing age (15-44 years) during the postfortification period of: serum folate (2003-2008); red blood cell (RBC) folate (2007-2008); and plasma total homocysteine (tHcy; 2003-2006), adjusted for age, race and ethnicity, and total dietary folate expressed as dietary folate equivalents for strata of supplement use and BMI. BMI was inversely associated with serum folate among women who did not use supplements containing folic acid; no differences between women in different BMI categories were observed among supplement users. Regardless of supplement use, obese women had the highest RBC folate concentrations. There were no differences in tHcy by BMI, regardless of supplement use.
These results do not support a straightforward modification of the relationship between supplemental folic acid intake and folate status by BMI. In this population, BMI may affect the body distribution of folate, as reflected by lower serum and higher RBC folate levels in obese women who do not use supplements.
closed_qa
Are airborne refractory ceramic fibers similar to asbestos in their carcinogenicity?
Animal studies on refractory ceramic fiber (RCF) have led to the suggestion that RCF might resemble asbestos in carcinogenicity. Human data are available to test this hypothesis. We compared the occurrence of lung cancer and mesothelioma in 605 men engaged in the manufacture of RCF and followed since 1987 to cancer rates that would have been anticipated if airborne RCF were carcinogenic to the same degree as are crocidolite, amosite or chrysotile asbestos. We integrated the results of workplace exposure monitoring with mortality follow-up using formulas presented by Hodgson and Darnton (2000) to estimate hypothesized risks under different asbestos scenarios. During 15,281 person-years of observation, there were 12 deaths from lung cancer. General population rates predicted 11.8 cases expected for an observed/expected (O/E) ratio of 1.0. Anticipated numbers of deaths from lung cancer under hypotheses of carcinogenicity similar to that of amphiboles and chrysotile were 62 and 17, allowing for rejection of amphibole-like effects (p<10(-5)) but not chrysotile-like carcinogenicity (p = 0.15). There were no cases of mesothelioma, as compared to 4.9 anticipated under a crocidolite-like hypothesis (p = 0.007 to reject), 1.0 for amosite (p = 0.38) and 0.05 for chrysotile (p = 0.95).
There was no increase in lung cancer or mesothelioma in these workers exposed to RCF. If the cohort had the same exposure to crocidolite asbestos the number of lung cancer and mesothelioma cases would have been significantly greater than observed. The data do not yet permit a similar conclusion with respect to chrysotile asbestos.
closed_qa
Do apolipoprotein E genotype and educational attainment predict the rate of cognitive decline in normal aging?
We investigated suspected longitudinal interaction effects of apolipoprotein E (APOE) genotype and educational attainment on cognitive decline in normal aging. Our sample consisted of 571 healthy, nondemented adults aged between 49 and 82 years. Linear mixed-models analyses were performed with four measurement time points: baseline, 3-year, 6-year, and 12-year follow-up. Covariates included age at baseline, sex, and self-perceived physical and mental health. Dependent measures were global cognitive functioning (Mini-Mental State Examination; Folstein, Folstein,&McHugh, 1975), Stroop performance (Stroop Color-Word Test; Van der Elst, Van Boxtel, Van Breukelen,&Jolles, 2006a), set-shifting performance (Concept Shifting Test; Van der Elst, Van Boxtel, Van Breukelen,&Jolles, 2006b), cognitive speed (Letter-Digit Substitution Test; Van der Elst, Van Boxtel, Van Breukelen,&Jolles, 2006c), verbal learning (Verbal Learning Test: Sum of five trials; Van der Elst, Van Boxtel, Van Breukelen,&Jolles, 2005), and long-term memory (Verbal Learning Test: Delayed recall). We found only faint evidence that older, high-educated carriers of the APOE-ε4 allele (irrespective of zygosity) show a more pronounced decline than younger, low-educated carriers and noncarriers (irrespective of educational attainment). Moreover, this outcome was confined to concept-shifting performance and was especially observable between 6- and 12-year follow-ups. No protective effects of higher education were found on any of the six cognitive measures.
We conclude that the combination of APOE-ε4 allele and high educational attainment may be a risk factor for accelerated cognitive decline in older age, as has been reported before, but only to a very limited extent. Moreover, we conclude that, within the cognitive reserve framework, education does not have significant protective power against age-related cognitive decline.
closed_qa
Are obstetrical, perinatal, and infantile difficulties associated with pediatric bipolar disorder?
Despite increasing acknowledgement of bipolar disorder (BD) in childhood, there is a paucity of literature that has investigated obstetrical, perinatal, and infantile difficulties and their potential link with BD. To this end, we examined difficulties during delivery, immediate post-birth, and infancy and the association with BD in childhood. From two similarly designed, ongoing, longitudinal, case-control family studies of pediatric BD (N = 327 families), we analyzed 338 children and adolescents [mean (± standard deviation) age: 12.00 ± 3.37 years]. We stratified them into three groups: healthy controls (N = 98), BD probands (N = 120), and their non-affected siblings (N = 120). All families were comprehensively assessed with a structured psychiatric diagnostic interview for psychopathology and substance use. Mothers were directly questioned regarding the pregnancy, delivery, and infancy difficulties that occurred with each child using a module from the Diagnostic Interview for Children and Adolescents-Parent Version (DICA-P). Mothers of BD subjects were more likely to report difficulties during infancy than mothers of controls [odds ratio (95% confidence interval) = 6.6 (3.0, 14.6)]. Specifically, children with BD were more likely to have been reported as a stiffened infant [7.2 (1.1, 47.1)]and more likely to have experienced 'other' infantile difficulties [including acting colicky; 4.9 (1.3, 18.8)] compared to controls. We found no significant differences between groups in regards to obstetrical or perinatal difficulties (all p values>0.05).
While our results add to previous literature on obstetrical and perinatal difficulties and BD, they also highlight characteristics in infancy that may be prognostic indicators for pediatric BD.
closed_qa
Can targeted early intervention improve functional recovery in psychosis?
This paper assesses the impact of different models of early intervention (EI) service provision on functional recovery and inpatient hospital admission. The study compares the outcome of a comprehensive EI team with a partial model (community mental health team (CMHT) plus specialist support) and traditional care (generic CMHT) over a 10-year period. The design is in comparison with historical control. The study compares the functional recovery outcomes of three cohorts from the same geographical area over the period 1998-2007. The primary outcomes were partial and full functional recovery defined with respect to readily identifiable UK benefit system thresholds and psychiatric inpatient admission days at 1 and 2 years post-referral. Only 15% of individuals made a full or partial functional recovery at 2 years under the care of a traditional generic CMHT in 1998. In 2007, 52% of the cases were making a full or partial functional recovery under the care of the comprehensive EI team. A large reduction in inpatient admissions was associated with the EI strategy.
The implementation of comprehensive EI teams can have a major impact in improving functional recovery outcomes in psychosis and reducing inpatient admissions. Partial implementation using limited funding of specialist workers in collaboration with traditional care appeared to have a more limited effect on these recovery dimensions. The implementation of targeted EI in psychosis strategies can result in substantive functional benefits.
closed_qa
Dissatisfaction with residency: a discrepancy between preferences and expectations?
Medical residents play two roles that enter into conflict during their educational period: trainees and workers. This dual role can lead to dissatisfaction among residents that can affect both the quality of the services they provide to citizens and the proper functioning of the health services model itself.AIM: To analyse discrepancies between the preferences and expectations of first-year medical residents and whether these differences affect satisfaction with the residency. A questionnaire was administered on-line to the entire population of first-year medical residents of the Autonomous Community of Andalusia (Spain) in 2008. We performed a means contrast test between the indicator discrepancy (difference between preferences and expectations during the residency as a training or a working period), overall satisfaction with the residency and their relationship to other expectations of medical residents. Respondents showing greater discrepancy have a more negative opinion about the residency.
There is a gap between what residents prefer and what they expect from the residency, giving rise to dissatisfaction. This gap must be bridged to improve the quality of training received by these new physicians, their satisfaction and hence the delivery of health services to citizens.
closed_qa
Tracheostomy and laryngectomy survey: do front-line emergency staff appreciate the difference?
In an emergency scenario, it is vital to appreciate the difference between a laryngectomy and a tracheostomy so that oxygen can be administered in an appropriate manner. This survey aimed to ascertain the level of emergency healthcare personnel's knowledge with regards to distinguishing between a tracheostomy and a laryngectomy patient, and the emergency management of such patients. Forty-four accident and emergency staff (28 doctors, nine nurses and seven paramedics) within one Foundation Trust were invited to complete a questionnaire to ascertain (1) their confidence at differentiating between a laryngectomy and tracheostomy stoma; (2) knowledge of the appropriate site for oxygen delivery if needed; and (3) overall level of training on this subject. There were significant gaps in knowledge, particularly with regards to fundamental differences between a tracheostomy and a laryngectomy; less than 5 per cent were able to describe the anatomical difference. Only 41 per cent correctly identified the route of oxygen administration in laryngectomy patients.
In this cohort of emergency staff, the fundamental difference between a laryngectomy and a tracheostomy was poorly understood. This lack of awareness of front-line emergency staff needs to be addressed in order to maximise patient safety.
closed_qa
Is home management of fevers a cost-effective way of reducing under-five mortality in Africa?
To assess the cost-effectiveness of two strategies of home management of under-five fevers in Ghana - treatment using antimalarials only (artesunate-amodiaquine - AAQ) and combined treatment using antimalarials and antibiotics (artesunate-amodiaquine plus amoxicillin - AAQ + AMX). We assessed the costs and cost-effectiveness of AAQ and AAQ + AMX compared with a control receiving standard care. Data were collected as part of a cluster randomised controlled trial with a step-wedged design. Approximately, 12,000 children aged 2-59 months in Dangme West District in southern Ghana were covered. Community health workers delivered the interventions. Costs were analysed from societal perspective, using anaemia cases averted, under-five deaths averted and disability-adjusted life years (DALYs) averted as effectiveness measures. Total economic costs for the interventions were US$ 204,394.72 (AAQ) and US$ 260,931.49 (AAQ + AMX). Recurrent costs constituted 89% and 90% of the total direct costs of AAQ and AAQ + AMX, respectively. Deaths averted were 79.1 (AAQ) and 79.9 (AAQ + AMX), with DALYs averted being 2264.79 (AAQ) and 2284.57 (AAQ + AMX). The results show that cost per anaemia case averted were US$ 150.18 (AAQ) and US$ 227.49 (AAQ + AMX) and cost per death averted was US$ 2585.58 for AAQ and US$ 3272.20 for AAQ + AMX. Cost per DALY averted were US$ 90.25 (AAQ) and US$ 114.21 (AAQ + AMX).
Both AAQ and AAQ + AMX approaches were cost-effective, each averting one DALY at less than the standard US$ 150 threshold recommended by the World Health Organisation. However, AAQ was more cost-effective. Home management of under-five fevers in rural settings is cost-effective in reducing under-five mortality.
closed_qa
Does the epidemiologic paradox hold in the presence of risk factors for low birth weight infants among Mexican-born women in Colorado?
To test whether foreign-born status confers a protective effect against low birth weight (LBW) outcomes among Mexican-origin women in Colorado. Retrospective cohort study utilizing Colorado birth records from 1989-2004 for multivariate logistic regression analysis. The study population was 66,422 U.S.-born women of Mexican origin (USB) and 85,000 Mexican-born (MB) women with singleton births. Mexican-born women had 24.9% lower odds of LBW (OR 0.751 95% CI 0.782) than USB women. Mexican-born women had a higher prevalence of risk factors for LBW than their USB counterparts (anemia, cardiac disease, hypertension, inadequate prenatal care, less than high school education). After adjusting for these risk factors, MB women had 22.5% lower odds of having LBW infants than USB women (OR 0.775, 95% CI 0.73-0.81).
This study supports the epidemiologic paradox of LBW; despite higher prevalence of risk factors, foreign-born status confers an overall protective effect against low birth weight outcomes.
closed_qa
Racial disparities in intensity of care at the end-of-life: are trauma patients the same as the rest?
Medicare data show Blacks and Hispanics use more health care resources in the last six months of life than Whites. We sought to determine if similar differences exist in trauma patients who died following moderate to severe injury. We analyzed data from a prospective cohort study of 18 Level 1 and 51 non-trauma centers in 12 states to examine racial/ethnic variation in intensity of care and hospital costs. Blacks were more likely than Whites to receive critical care consultation RR=1.67 (95% CI, 1.22, 2.30), specialty assessments RR=1.44 (95% CI, 1.12, 1.86) and procedures RR=1.22 (95% CI, 1.00, 150). Hispanics were less likely than Whites to have withdrawal-of-care orders, RR=0.72 (95% CI, 0.53, 0.98).
Among patients who die after trauma, Blacks receive higher intensity of care and Hispanics were less likely to have withdrawal of care orders than others. This suggests racial disparities in patient preferences and provider treatment.
closed_qa
Is a positive nasal lysine-aspirin challenge test associated with a more severe phenotype of chronic rhinosinusitis and asthma?
Current guidelines recommend a greater use of aspirin challenge testing in the diagnosis of aspirin-intolerant rhinosinusitis and asthma, a disorder with high burden of illness and resistance to treatment. The indications for these tests and their clinical significance remain unclear. This study was designed to characterize the phenotype of patients with chronic rhinosinusitis with nasal polyposis (CRSwNP) with or without asthma undergoing a nasal lysine-aspirin (L-ASA) challenge to evaluate which factors strongly predict a positive test. Seventy-five patients with CRSwNP underwent nasal challenge with 16 mg (total) of L-ASA after 30 minutes of acclimatization and diluent challenge. A positive challenge was defined as a 25% drop in total nasal volume measured by acoustic rhinometry. Twenty-three (31%) participants gave a history of aspirin intolerance and 38 (51%) had a positive nasal L-ASA challenge. Upper airway measures (CT scan score, olfaction, polyp grading, peak nasal inspiratory flow, nasal symptoms, etc.) and lower airway measures (methacholine provocative concentration required to produce a 20% drop in forced expiratory volume in 1 second, effective special airway resistance, and spirometry) were not significantly worse in patients with a positive aspirin challenge. Test sensitivity was 48%, specificity was 52%, positive predictive value was 29%, and negative predictive value was 68%. A regression analysis identified forced expiratory flow at 25-75% (FEF(25-75)), history of aspirin intolerance, and duration of rhinosinusitis as significant predictors of a positive aspirin challenge.
A positive response to nasal L-ASA challenge is not associated with a more severe phenotype of CRSwNP with or without asthma. A history of aspirin intolerance, duration of rhinosinusitis, and FEF(25-75) predict a greater response to aspirin.
closed_qa
Work burden with remote monitoring of implantable cardioverter defibrillator: is it time for reimbursement policies?
The efficacy and accuracy, as well as patients' satisfaction, of device remote monitoring are well demonstrated. However, the workload of remote monitoring management has not been estimated and reimbursement schemes are currently unavailable in most European countries. This study evaluates the workload associated with remote monitoring systems. A total of 154 consecutive implantable cardioverter defibrillator patients (age 66±12 years; 86.5% men) with a remote monitoring system were enrolled. Data on the clinician's workload required for the management of the patients were analyzed. A total of 1744 transmissions were received during a mean follow-up of 15.3±12.4 months. Median number of transmissions per patient was 11.3. There were 993 event-free transmissions, whereas 638 transmissions regarded one or more events (113 missed transmissions, 141 atrial events, 132 ventricular episodes, 299 heart failure-related transmissions, 14 transmissions regarding lead malfunction and 164 transmissions related to other events). In 402 cases telephonic contact was necessary, whereas in 68 cases an in-clinic visit was necessary and in 23 of them an in-clinic visit was prompted by the manufacturer due to technical issues of the transmitter. During follow-up, 316 work hours were required to manage the enrolled patients. Each month, a total of 14.9 h were spent on the remote monitoring of 154 patients (9.7 h for 100 patients monthly) with approximately 1.1±0.15 h per year for each patient.
The clinician's work burden is high in patients with remote monitoring. In order to expand remote monitoring in all patients, reimbursement policies should be considered.
closed_qa
2011 Mid-America Orthopaedic Association Dallas B. Phemister Physician in Training Award: Can musculoskeletal tumors be diagnosed with ultrasound fusion-guided biopsy?
Percutaneous biopsy for musculoskeletal tumors commonly relies on imaging adjuncts including ultrasound (US), CT, or MRI. These modalities however have disadvantages (US) or are cumbersome, not universally available, and costly (CT and MRI). US fusion is a novel technique that fuses previously obtained CT or MRI data with real-time US, which allows biopsies to be performed in an US suite. It has proven useful in various body systems but musculoskeletal applications remain scarce. Our goal is to evaluate the fusion technology and determine its ability to diagnose musculoskeletal tumors.QUESTIONS/ We determined whether biopsies performed via US fusion compared with CT guidance provide equivalent diagnostic yield and accuracy and allow quicker biopsy scheduling and procedure times. Forty-seven patients were assigned to undergo either US fusion (with MR, n = 16 or CT, n = 15) or CT-guided biopsies (n = 16). We evaluated adequacy of the histologic specimen (diagnostic yield) and correlation with surgical pathology (diagnostic accuracy). We determined scheduling times and lengths of the biopsy. US fusion and CT-guided biopsy groups had comparable diagnostic yields (CT = 94%; US/MRI = 94%; US/CT = 93%) and accuracy (CT = 83%; US/MRI = 90%; US/CT = 100%). US fusion biopsies were faster to schedule and perform. All procedures were safe with minimal complications.
US fusion provides a high diagnostic yield and accuracy comparable to CT-guided biopsy while performed in the convenience of an US suite. This may have resulted in the observed faster scheduling and biopsy times.
closed_qa
Is serum alpha-fetoprotein useful for predicting recurrence and mortality specific to hepatocellular carcinoma after hepatectomy?
Serum alpha-fetoprotein (AFP) is frequently used to predict posthepatectomy outcomes in patients with hepatocellular carcinoma (HCC), but its predictive value is still not established. Therefore, we assessed the prognostic significance of AFP status. Of 525 patients undergoing curative hepatectomy for HCC, 290 had preoperative AFP levels of ≥20 ng/mL (AFP-positive group) and 235 had AFP levels of<20 ng/mL (AFP-negative group). We compared the 2 groups with respect to time-to-recurrence, using the inverse probability of treatment weighted (IPTW) for the entire cohort and propensity score matching, and the cumulative incidence of HCC-specific mortality using competing risks regression. During follow-up (median duration 64 months, range 2-137 months), HCC recurred in 54.9 % of the AFP-negative group and 52.4 % of the AFP-positive group; there was no death without recurrence. After IPTW adjustment, time-to-recurrence did not differ in the 2 groups (hazard ratio [HR] 0.86, 95 % confidence interval [95 % CI]0.66-1.12; P = 0.28). In a propensity-score matched cohort (152 pairs), time-to-recurrence data were similar to those obtained by IPTW adjustment (HR 0.91, 95 % CI 0.65-1.25; P = 0.55). There was no difference in recurrence pattern (site and stage) or treatment between the 2 groups even after propensity-score matching. The adjusted HR evaluating the impact of AFP positivity on the risk of HCC-specific mortality was 0.77 (95 % CI 0.54-1.08; P = 0.13) A multivariable competing risks analysis also failed to reveal a significant correlation between baseline AFP level and HCC-specific mortality in the AFP-positive group.
Preoperative AFP levels are not useful for predicting recurrence or survival endpoints following curative hepatectomy for HCC.
closed_qa
CCR7 and VEGF-C: molecular indicator of lymphatic metastatic recurrence in pN0 esophageal squamous cell carcinoma after Ivor-Lewis esophagectomy?
Esophageal squamous cell carcinoma (ESCC) is a malignant tumor with a high incidence of lymph node metastasis. This study was undertaken to investigate the expression of CCR7 and VEGF-C in pN0 ESCC and its relationship with lymphatic metastatic recurrence. The expression of CCR7 and VEGF-C was examined by RT-PCR and immunohistochemistry. The recurrence rates were calculated by the Kaplan-Meier method and their difference was determined by log rank analysis. Cox regression analysis was performed to determine the independent risk factors. In 99 patients, CCR7 mRNA expression was observed in 42 patients with a 3 year recurrence rate of 57.1%; VEGF-C mRNA expression was observed in 52 patients with a 3 year recurrence rate of 53.8%; and coexpression of CCR7 mRNA and VEGF-C mRNA was observed in 22 patients with a 3 year recurrence of 63.6%. Neither CCR7 mRNA nor VEGF-C mRNA expression was observed in 27 patients with a 3 year recurrence rate of 22.2%. The recurrence rates of patients with positive expression of CCR7 mRNA and/or VEGF-C mRNA were significantly higher than in patients without expression of both CCR7 mRNA and VEGF-C mRNA. We achieved better concordance between RT-PCR and immunohistochemistry detection of both markers. The Cox regression analysis showed tumor T classification, positive expression of CCR7/VEGF-C mRNA, and positive expression of CCR7/VEGF-C protein in tumor tissues to be independent risk factors for 3 year recurrence.
Patients with positive expression of CCR7 and/or VEGF-C have a higher recurrence rate than patients without expression of both CCR7 and VEGF-C. CCR7 and VEGF-C may become molecular indicators of disease in patients vulnerable to lymphatic metastatic recurrence.
closed_qa
Persistence of mitral regurgitation following ring annuloplasty: is the papillary muscle outside or inside the ring?
Ischemic mitral regurgitation (IMR) often persists, despite annular ring reduction. It has been hypothesized that persistent IMR following ring annuloplasty was related to a continued tethering of the mitral leaflets, as defined by the distance by which the papillary muscles (PMs) were displaced outside the mitral annular ring. Seven sheep (four acute, three chronic) with persistent mitral regurgitation (MR) following ring annuloplasty for IMR were studied using three-dimensional (3D) echocardiography to examine the mitral valve geometry. The three stages examined were: Stage 1, baseline; Stage 2, post myocardial infarction (via ligation of the obtuse marginal branches); and Stage 3, post undersized ring annuloplasty. The 3D echocardiography measurements included mitral annular area, tethering distance from the ischemic PM to the anterior annulus, and the outside displacement of the PM relative to ring PM displacement. Persistent moderate MR remained in these seven sheep following undersized ring annuloplasty (MR vena contracta change (pre versus post ring): 7.0 versus 5.8 +/- 2.4 mm, p = NS), despite a reduction in the mitral annular area of 50 +/- 18% (10.3 +/- 6.3 versus 4.7 +/- 1.3 cm2). Ring annuloplasty shifted the posterior annulus towards the anterior annulus, such that the infarcted PM became displaced outside the mitral annulus. The projected displacement distance of the PM outside versus inside the annular ring was 8.4 +/- 2.4 mm outside mitral annulus post ring versus 3.6 +/- 2.5 mm within mitral annulus pre ring, p<0001). The displacement distance from the infarcted PM to the mitral annulus restricted the ability of the posterior leaflet to move anteriorly, preventing effective coaptation. By multivariate analysis, this displacement distance was an important determinant of residual MR (p<0.02).
Persistent MR following ring annuloplasty for IMR relates to persistently abnormal leaflet tethering, with restricted posterior leaflet motion due to PM displacement outside of the mitral annulus.
closed_qa
Does thinking and doing the same thing amount to involved participation?
Participation as involvement in a situation includes two dimensions; doing the activity and the experience of involvement. The ICF-CY only measures doing using the capacity and performance qualifiers, a dimension measuring the experience is needed; a third qualifier. The experienced involvement of pupils in school activities is higher when thinking and doing coincided. By comparing self-reported experiences of involvement of children, data about what children were thinking and doing during activities were gathered from 21 children with and 19 without disabilities in inclusive classrooms. A relationship exists between an index of the subjective experience of involvement and whether children were thinking and doing the same things.
This index can be constructed using measures of concentration, control, involvement, and motivation. Choice is influential, as knowledge about why an activity is undertaken affects involvement. Additionally, increased subjective experience of involvement gives better psychological health and well-being.
closed_qa
Can simulation replace part of clinical time?
Education in simulated learning environments (SLEs) has grown rapidly across health care professions, yet no substantive randomised controlled trial (RCT) has investigated whether SLEs can, in part, substitute for traditional clinical education. Participants were physiotherapy students (RCT 1, n = 192; RCT 2, n = 178) from six Australian universities undertaking clinical education in an ambulatory care setting with patients with musculoskeletal disorders. A simulated learning programme was developed as a replica for clinical education in musculoskeletal practice to replace 1 week of a 4-week clinical education placement. Two SLE models were designed. Model 1 provided 1 week in the SLE, followed by 3 weeks in clinical immersion; Model 2 offered training in the SLE in parallel with clinical immersion during the first 2 weeks of the 4-week placement. Two single-blind, multicentre RCTs (RCT 1, Model 1; RCT 2, Model 2) were conducted using a non-inferiority design to determine if the clinical competencies of students part-educated in SLEs would be any worse than those of students educated fully in traditional clinical immersion. The RCTs were conducted simultaneously, but independently. Within each RCT, students were stratified on academic score and randomised to either the SLE group or the control ('Traditional') group, which undertook 4 weeks of traditional clinical immersion. The primary outcome measure was a blinded assessment of student competency conducted over two clinical examinations at week 4 using the Assessment of Physiotherapy Practice (APP) tool. Students' achievement of clinical competencies was no worse in the SLE groups than in the Traditional groups in either RCT (Margin [Δ] ≥ 0.4 difference on APP score; RCT 1: 95% CI - 0.07 to 0.17; RCT 2: 95% CI - 0.11 to 0.16).
These RCTs provide evidence that clinical education in an SLE can in part (25%) replace clinical time with real patients without compromising students' attainment of the professional competencies required to practise.
closed_qa
Top-down and bottom-up approaches to motor skill assessment of children: are child-report and parent-report perceptions predictive of children's performance-based assessment results?
BACKGROUND/ AIM: Therapists use different types of tests, scales, and instruments to assess children's motor skills, including those classified as being top-down and bottom-up. The aim of the study was to investigate the ability of measures of children's motor skill performance from the perspectives of children and parents (a type of top-down assessment) to predict children's performance-based motor ability test results (a type of bottom-up assessment). A convenience sample of 38 children and parents was recruited from Victoria, Australia. Motor skill performance was evaluated from a top-down perspective using the Physical Self-Description Questionnaire (PSDQ) and the Movement Assessment Battery for Children--Second Edition (MABC-2) Checklist to measure children's and parents' perspectives respectively. Motor skill performance was also evaluated from a bottom-up approach using the Bruininks-Oseretsky Test of Motor Proficiency--Second Edition (BOT-2). Data were analyzed using multiple linear regression analysis to determine whether the PSDQ or MABC-2 Checklist was predictive of the children's BOT-2 performance results. Two predictive relationships were identified based on parents' perspectives, where the total score of the MABC-2 Checklist was found to be a significant predictor of the BOT-2 Manual Coordination motor composite score, accounting for 8.35% of its variance, and the BOT-2 Strength and Agility motor composite score, accounting for 11.6% of its variance. No predictive relationships were identified between the children's self-report PSDQ perspectives and the BOT-2 performance scores.
Therapists are encouraged to utilize a combination of top-down and bottom-up approaches and purposefully to seek parents' and children's perspectives when evaluating children's motor skill performance.
closed_qa
Functional neuroimaging in craniopharyngioma: a useful tool to better understand hypothalamic obesity?
To use functional magnetic resonance imaging (fMRI) in craniopharyngioma (CP) patients to examine the hypothesis that hypothalamic damage due to CP and its treatment results in enhanced perception of food reward and/or impaired central satiety processing. Pre- and post-meal responses to visual food cues in brain regions of interest (ROI; bilateral nucleus accumbens, bilateral insula, and medial orbitofrontal cortex) were assessed in 4 CP patients versus 4 age- and weight-matched controls. Stimuli consisted of images of high- ('fattening') and low-calorie ('non-fattening') foods in blocks, alternating with non-food object blocks. After the first fMRI scan, subjects drank a high-calorie test meal to suppress appetite, then completed a second fMRI scan. Within each ROI, we calculated mean z-scores for activation by fattening as compared to non-fattening food images. Following the test meal, controls showed suppression of activation by food cues while CP patients showed trends towards higher activation.
These data, albeit in a small group of patients, support our hypothesis that perception of food cues may be altered in hypothalamic obesity (HO), especially after eating, i.e. in the satiated state. The fMRI approach is encouraging for performing future mechanistic studies of the brain response to food cues and satiety in patients with hypothalamic or other forms of childhood obesity.
closed_qa
Is peripheral neuron degeneration involved in multiple system atrophy?
Lower motor neuron lesions are not among the characteristic features of multiple system atrophy (MSA), although electromyography (EMG) and autopsy studies revealed peripheral neuron abnormalities in some cases of MSA. The aim of the study was to evaluate subclinical involvement of the peripheral neuron in MSA using EMG and electroneurography (ENG). 48 patients with clinically probable MSA (mean age 60.6 years; 67% males) were included in the study and divided into subgroups, with predominant cerebellar (MSA-C) and parkinsonian signs (MSA-P). ENG in ulnar, peroneal and sural nerves and EMG of the first interosseus dorsal and tibial anterior muscles were performed. Abnormal ENG in one nerve was recorded in 20.8% of patients, and in two nerves in another 20.8% of patients. The most frequent and significant findings were decreased compound motor action potential amplitudes in the ulnar nerve in the overall MSA group as well as in the MSA-P type as compared to controls. Abnormalities suggesting reinnervation was observed in 43 of 96 examined muscles (44.7%). In individual cases, neurogenic features were recorded in one muscle in 31.2% of patients and in two muscles in 29.1% of patients.
Subclinical axonopathy in MSA is not frequent and is more pronounced in MSA with predominant parkinsonian signs. In MSA, neurogenic EMG abnormalities in muscles are more frequent than peripheral nerve lesions and as evidenced by increased motor unit potential amplitudes, could be considered a sign of anterior horn cell involvement and a hallmark of the "continuum" of neurodegeneration in MSA.
closed_qa
Cardiac findings in routine fetal autopsies: more than meets the eye?
The main objective of this study was to evaluate the spectrum of cardiac anomalies found in routinely performed fetal autopsies and to establish the correlation between prenatal and postmortem diagnosis. A retrospective study of fetal autopsies was performed. Cases with cardiac anomalies were analyzed. Seven main categories were established and each case was assigned to a single group. Cardiac defects were also classified as isolated or with associated anomalies. In the cases with prenatal diagnosis, we performed a correlation between prenatal and postmortem findings. Abnormal cardiac findings were identified in 99 fetuses (13.6%). The two most common categories were septal defects and complex anomalies, each occurring in 21 fetuses (21.2%). Sixty-seven (67.7%) had associated anomalies. Septal anomalies were more frequent in cases with associated anomalies (p=0.012). Prenatal diagnosis had been performed in 50 cases. There was complete agreement between prenatal and postmortem diagnosis in 36 cases (72%), and major agreement with additional information in ten (20%). When the echocardiogram was not performed by a specialist, the number of cases classified with complete disagreement was higher (33.3% vs 2.4%) (p=0.002).
The high prevalence of cardiac defects in lost pregnancies, some of them lacking prenatal diagnosis, highlights the importance of examining the heart in all cases.
closed_qa
Radical resection of gallbladder cancer: could it be robotic?
The only potentially curative option for patients with gallbladder cancer is radical resection. This is the first report that describes the successful application of a minimally invasive, robot-assisted radical resection, including lymphadenectomy, in five gallbladder cancer patients. Medical records of patients who underwent radical resection of gallbladder cancer via the da Vinci robotic surgical system in the Hepato-Bilio-Pancreatic Surgical Department of the Shanghai Ruijin Hospital, China, between March 2010 and July 2011 were reviewed and analyzed. Robot-assisted radical resection was successful in all five patients. The mean number of excised lymph nodes was 9 (range = 3-11), mean operative time was 200 min (range = 120-300 min), mean intraoperative blood loss was 210 ml (range = 50-400 ml), and mean length of hospital stay was 7.4 days (range = 7-8 days). All patients were discharged with no reported complications. Mean postoperative follow-up was 11 months (range = 1-17 months). One patient died due to tumor recurrence 10 months postsurgically, but there was no recurrence in the remaining four patients during the follow-up period.
Robot-assisted radical resection for gallbladder cancer is both feasible and safe. Compared to laparoscopic surgery, the robotic surgery system is better suited for subtle dissection in a narrow, deep space. This is advantageous for both the removal of lymph nodes near the pancreas and hepatoduodenal ligament and the skeletonization of the hepatoduodenal ligament, the hepatic artery, and the celiac axis. The long-term outcome and direct comparisons to laparotomy in a larger patient cohort are needed to provide more clinical data supporting the superiority of this approach.
closed_qa
Can early laparoscopic cholecystectomy be the optimal management of cholecystitis with gallbladder perforation?
Gallbladder perforation is a rare but serious complication of cholecystitis. It was usually managed by percutaneous gallbladder drainage (PTGBD) followed by elective cholecystectomy. However, evidences are emerging that early laparoscopic cholecystectomy (LC) is still feasible under these conditions. We hypothesized that early LC may have comparable surgical results as to those of PTGBD + elective LC. From January 2005 to October 2011, patients admitted to China Medical University Hospital with a diagnosis of perforated cholecystitis were retrospectively reviewed. The diagnosis of gallbladder perforation was made by image and/or intraoperative findings. Those patients who had unstable hemodynamics that were not fitted for general anesthesia or those who had concomitant major operations were excluded. Patients were divided into three groups: early open cholecystectomy (group 1), early LC (group 2), and PTGBD followed by elective LC (group 3). The demographic features, surgical results, and patient outcome were analyzed and compared between groups. A total of 74 patients were included. All patients had similar demographic features except that patients in group 2 were younger (62 vs. 72 and 73.5 years) compared with group 1 and group 3 (p = 0.016). There were no differences in terms of operative time, blood loss, conversion, and complication rate between three groups. The length of hospital stay (LOS) was significant shorter in group 2 patients compared with that of groups 1 and 3.
Although PTGBD followed by elective LC was still the mainstay for the treatment of gallbladder perforation, early LC had comparable surgical outcomes as that of PTGBD + LC but with a significantly shorter LOS. Early LC should be considered the optimal treatment for gallbladder perforation, and PTGBD + LC can be preserved for those who carried a high risk of operation.
closed_qa
Does fellow participation in laparoscopic Roux-en-Y gastric bypass affect perioperative outcomes?
Laparoscopic Roux-en-Y gastric bypass (LRYGB) requires specialized training commonly acquired during a fellowship. We hypothesized that fellows affect patient outcomes and this effect varies during training. We included all LRYGB from the 2005 to 2009 American College of Surgeons-National Surgical Quality Improvement Program database. Cases without trainees (attending) were compared to those with trainees of ≥6 years (fellow). Outcomes were pulmonary, infectious, and wound complications and deep venous thrombosis (DVT). Multivariable regression controlled for age, BMI, and comorbidities. Of the 18,333 LRYGB performed, 4,349 (24%) were fellow cases. Fellow patients had a higher BMI (46.1 vs. 45.7, p<0.001) and fewer comorbidities. Mortality was 0.2 and 0.1% and overall morbidity was 4.8 and 6.0% for attending and fellow groups, respectively. On adjusted analysis, mortality was similar, but fellow cases had 30% more morbidity (p = 0.001). Specifically, fellows increased the odds of superficial surgical site infections (SSSIs) [odds ratio (OR) = 1.4, p = 0.01], urinary infections (UTIs) (OR = 1.7, p = 0.002), and sepsis (OR = 1.5, p = 0.05). During the first 6 months, fellows increased the odds of DVT (OR = 4.7, p = 0.01), SSIs (OR = 1.5, p = 0.001), UTIs (OR = 1.8, p = 0.004), and sepsis (OR = 1.9, p = 0.008). By the second half of training, fellow cases demonstrated outcomes equivalent to attending cases.
Involving fellows in LRYGB may increase DVT, SSIs, UTIs, and sepsis, especially early in training. By completion of their training, cases involving fellows exhibited outcomes similar to cases without trainees. This supports both the need for fellowship training in bariatric surgery and the success of training to optimize patient outcomes.
closed_qa
Is laparoscopic surgery for recurrent Crohn's disease beneficial in patients with previous primary resection through midline laparotomy?
Patients undergoing abdominal surgery for Crohn's disease are predisposed to recurrence requiring reoperation. The effectiveness of laparoscopic versus open resection in patients with previous intestinal resection for Crohn's through midline laparotomy is controversial. Patients with previous open resection for intestinal Crohn's disease undergoing elective laparoscopic surgery for recurrent bowel disease from 1997 to 2011 were case-matched with open counterparts based on age (±5 years), gender, body mass index (±2 kg/m(2)), American Society of Anesthesiologists (ASA) score, surgical procedure, and year of surgery (±3 years). Groups were compared using Chi-square or Fisher exact tests for categorical and the Wilcoxon rank-sum test for quantitative data. 26 patients undergoing laparoscopic ileocolectomy (n = 14), proctocolectomy (n = 5), small bowel resection (n = 4), abdominoperineal resection (n = 1), extended right colectomy (n = 1), and strictureplasty (n = 1) were well matched to 26 patients undergoing open surgery. The number of previous operations, disease phenotypes, steroid use, and comorbidities were comparable in the two groups. There were no deaths, and three patients (12%) required conversion because of adhesions. Laparoscopic and open groups had statistically similar operating times (169 versus 158 min, p = 0.94), estimated blood loss (222 versus 427 ml, p = 0.32), overall morbidity (39 versus 69%, p = 0.051), reoperation rates (8 versus 0%, p = 0.5), postoperative return of bowel function (3.5 ± 1.4 versus 3.9 ± 1.7 days, p = 0.3), mean length of hospital stay (6.4 ± 6.2 versus 6.9 ± 3.5 days, p = 0.12), and readmission rates (8 versus 12%, p = 0.64). Wound infection rate was decreased after laparoscopic surgery (0 versus 27%, p = 0.01).
Surgery for recurrent Crohn's disease in patients with previous primary resection through laparotomy can be frequently and safely completed laparoscopically. Wound infection rates are reduced, but the recovery advantages of a minimally invasive approach are not maintained when compared with open surgery. The decision to operate laparoscopically should therefore be carefully calibrated.
closed_qa
Chemotherapy in patients with advanced pancreatic cancer: too close to death?
We evaluated the attitude in using chemotherapy near the end of life in advanced pancreatic adenocarcinoma (PAC). Clinical and laboratory parameters recorded at last chemotherapy administration were analyzed, in order to identify risk factors for imminent death. Retrospective analysis of patients who underwent at least one line of palliative chemotherapy was made. Data concerning chemotherapy (regimens, lines, and date of last administration) were collected. Clinical and laboratory factors recorded at last chemotherapy administration were: performance status, presence of ascites, hemoglobin, white blood cell (WBC), platelets, total bilirubin, albumin, LDH, C-reactive protein (C-rp), and Ca 19.9. We analyzed 231 patients: males/females, 53/47 %; metastatic/locally advanced disease, 80/20 %; and median age, 66 years (range 32-85). All patients died due to disease progression. Median overall survival was 6.1 months (95 % CI 5.1-7.2). At the last chemotherapy delivery, performance status was 0-1 in 37 % and 2 in 63 %. Fifty-nine percent of patients received one chemotherapy line, while 32, 8, and 1 % had second-, third-, and fourth line, respectively. The interval between last chemotherapy administration and death was<4 weeks in 24 %, ≥4-12 in 47 %, and>12 in 29 %. Median survival from last chemotherapy to death was 7.5 weeks (95 % CI 6.7-8.4). In a univariate analysis, ascites, elevated WBC, bilirubin, LDH, C-rp and Ca 19.9, and reduced albumin were found to predict shorter survival; however, none of them remained significant in a multivariate analysis.
A significant proportion of patients with advanced PAC received chemotherapy within the last month of life. The clinical and laboratory parameters recorded at last chemotherapy delivery did not predict shorter survival.
closed_qa
Do immigrants from Turkey, Pakistan and Ex-Yugoslavia with newly diagnosed type 2 diabetes initiate recommended statin therapy to the same extent as Danish-born residents?
To explore whether newly diagnosed type 2 diabetes patients without previous cardiovascular disease (CVD) initiate preventive statin therapy regardless of ethnic background. Using nationwide individual-level registers, we followed a cohort of Danish-born residents and immigrants from Turkey, Pakistan and Ex-Yugoslavia, all without previous diabetes or CVD, during the period 2000-2008 for first dispensing of oral glucose-lowering medication (GLM), first dispensing of statins and register-markers of CVD (N = 3,764,620). Logistic regression analyses were used to test whether the odds ratios (ORs) of early statin therapy initiation (within 180 days after first GLM dispensing) are the same regardless of ethnic background. While age and gender were included as confounders in the basic model, income was included in the second model as a potential mediating variable. Compared to native Danes, the ORs for early statin therapy were 0.68 (95 % confidence interval 0.50-0.92], 0.67 (0.56-0.81) and 0.56 (0.44-0.71) for Ex-Yugoslavians, Turks and Pakistanis, respectively. The differences remained largely unchanged after adjusting for income and tended to be accentuated when the threshold period was extended. The ORs of women initiating therapy (compared to native Danes) were 0.56 (0.35-0.90), 0.60 (0.46-0.78) and 0.48 (0.32-0.72) for Ex-Yugoslavians, Turks and Pakistanis, respectively, and those for men were 0.78 (0.52-1.17), 0.74 (0.58-0.95) and 0.60 (0.44-0.83), respectively.
Immigrants from Turkey, Pakistan and Ex-Yugoslavia with type 2 diabetes were less likely to initiate statin therapy than Danish-born residents-despite a similar or even higher risk of CVD. The treatment inequities associated with ethnicity were more pronounced in women than men.
closed_qa
Does health information exchange reduce unnecessary neuroimaging and improve quality of headache care in the emergency department?
Health information exchange (HIE) is advocated as an approach to reduce unnecessary testing and improve quality of emergency department (ED) care, but little evidence supports its use. Headache is a specific condition for which HIE has theoretical benefits. To determine whether health information exchange (HIE) reduces potentially unnecessary neuroimaging, increases adherence with evidence-based guidelines, and decreases costs in the emergency department (ED) evaluation of headache. Longitudinal data analysis All repeat patient-visits (N = 2,102) by all 1,252 adults presenting with headache to a Memphis metropolitan area ED two or more times between August 1, 2007 and July 31, 2009. Use of a regional HIE connecting the 15 major adult hospitals and two regional clinic systems by authorized ED personnel to access the patient's record during the time period in which the patient was being seen in the ED. Diagnostic neuroimaging (CT, CT angiography, MRI or MRI angiography), evidence-based guideline adherence, and total patient-visit estimated cost. HIE data were accessed for 21.8 % of ED patient-visits for headache. 69.8 % received neuroimaging. HIE was associated with decreased odds of diagnostic neuroimaging (odds ratio [OR] 0.38, confidence interval [CI]0.29-0.50) and increased adherence with evidence-based guidelines (OR 1.33, CI 1.02-1.73). Administrative/nursing staff HIE use (OR 0.24, CI 0.17-0.34) was also associated with decreased neuroimaging after adjustment for confounding factors. Overall HIE use was not associated with significant changes in costs.
HIE is associated with decreased diagnostic imaging and increased evidence-based guideline adherence in the emergency evaluation of headache, but was not associated with improvements in overall costs. Controlled trials are needed to test whether specific HIE enhancements to increase HIE use can further reduce potentially unnecessary diagnostic imaging and improve adherence with guidelines while decreasing costs of care.
closed_qa
Do children eat less at meals when allowed to serve themselves?
The effect of self-serving on young children's energy intake is not well understood. The objective was to examine individual differences in the effects of plated and self-served entrée portions on children's energy intake. Two within-subjects experiments were used to examine ad libitum intake at meals in 63 children aged 3-5 y when 400 g of a pasta entrée was either plated or available for children to self-serve. Child age, sex, BMI, and responsiveness to increasing portion size (defined as individual slope estimates relating ad libitum intake of the entrée across a range of entrée portions) were evaluated as predictors of self-served portions. Children's entrée and meal intakes did not differ between the self-served and plated conditions for the total sample or by child weight status. However, larger self-served entrée portions were associated with greater entrée and meal intakes. Children who served themselves larger entrée portions tended to be overweight and more responsive to portion size (ie, greater increases in entrée intake as plated portion size increased). Last, self-served portion predicted both entrée and meal intake over and above BMI z score and responsiveness to portion.
Contrary to our hypothesis, relative to plated portions, allowing children to self-serve the entrée portion did not reduce energy intake. Children who were more responsive to portion-size effects were likely to self-serve and eat larger entrée portions. Self-serving is not a one-size-fits-all approach; some children may need guidance and rules to learn how to self-select appropriate portion sizes.
closed_qa
Does levator trauma 'heal'?
To evaluate if pregnancy- and delivery-related changes to levator morphology and distensibility regress with time. 488 nulliparous pregnant women, recruited between 36 and 38 weeks' gestation, were invited for assessment at 3-6 months and again at 2-3 years postpartum. All underwent an interview and four-dimensional translabial ultrasound examination. Hiatal morphometry and bladder neck descent (BND) were determined and compared between the two postpartum visits. 367 participants returned for assessment at 4.1 (interquartile range (IQR), 3.7-5.0) months and 161 returned at 2.6 (IQR, 2.0-3.1) years, allowing a groupwise comparison. There was no significant difference in hiatal area (22 vs 22 cm(2), P = 0.95) or BND on Valsalva maneuver (26.3 vs 25.5 mm, P = 0.49). Pairwise comparison in women who had attended both postpartum appointments without second births (n = 77), separately for those who had a cesarean section (n = 24) and those who had a vaginal delivery (n = 53) originally, showed no significant changes, except a reduction in BND (31.2 vs 28.3 mm, P = 0.025) in those who had delivered vaginally. Two women out of 12 diagnosed with a levator avulsion at 3-6 months showed obvious anatomical improvement on translabial ultrasound at 2-3 years.
We found no evidence of regression or healing of pregnancy- and delivery-related changes to levator distensibility on comparing imaging data obtained at 3-6 months and 2-3 years postpartum. However, we documented anatomical improvement on translabial ultrasound at the second postpartum visit in two women diagnosed with levator avulsion at 3-6 months postpartum.
closed_qa
Should LSIL-H be a distinct cytology category?
The 2001 Bethesda System for gynecologic cervical cytology reporting classifies squamous intraepithelial lesions into low-grade (LSIL) and high-grade (HSIL) lesions. An intermediate term, "low-grade squamous intraepithelial lesion, cannot exclude high-grade squamous intraepithelial lesion (LSIL-H)," has been used in a small percentage of LSIL cases. To the authors' knowledge, little is known regarding the human papillomavirus (HPV) status in patients with LSIL-H. A total of 808 SurePath specimens obtained between December 2009 and April 2011 were tested for 40 HPV genotypes using DNA microarray, followed by a confirmatory DNA sequencing assay. The infection rate for high-risk HPV in women with LSIL-H (92%) was strikingly close to that for women with HSIL (91%), which was higher than that for those with LSIL (74%); atypical squamous cells, cannot rule out high-grade lesion (ASC-H) (78%); or LSIL and ASC-H combined (74%). HPV type 16, the most common carcinogenic HPV genotype, was detected in 36% of women with LSIL-H, which was significantly higher than that in women with LSIL and ASC-H combined (13.8%), but less than that in women with HSIL (44.6%). Patients with LSIL-H and HSIL had similar infection rates for low-risk/intermediate-risk HPV genotypes, which were lower than those in LSIL or LSIL and ASC-H combined.
Women found to have LSIL-H on a Papanicolaou test appear to have a unique HPV distribution pattern that clearly differs from LSIL and is comparable to that for HSIL, suggesting an increased risk of high-grade lesions over that of women with LSIL. Recognizing LSIL-H as an independent diagnostic category may help in the early identification of the high-risk subgroup that may require a management algorithm comparable to that for patients with HSIL. Cancer (Cancer Cytopathol) 2012. © 2012 American Cancer Society.
closed_qa
Retinopathy of prematurity: are we missing any infant with retinopathy of prematurity?
To evaluate incidence of retinopathy of prematurity (ROP) and validity of current ROP screening criteria adopted in the Armed Forces Hospitals. A retrospective study of ROP indices was carried out in infants admitted to the neonatal intensive care unit in Khamis Mushait, Saudi Arabia, from January 2009 to December 2011. All infants who had birth weight (BW) of 1500 g or less and/or gestational age (GA) of 32 weeks or less had their data reviewed for ROP diagnosis and related indices. Some heavier or older infants were also included in the study. 386 infants were screened and ROP was diagnosed in 90 infants (23.31%). 25 infants (6.47%) had type 1 or worse ROP and were treated with cryopexy or laser. For patients having ROP, the mean GA was 27.8 ± 2 weeks and mean BW was 961.8 ± 237.4 g. Infants who needed treatment for ROP had a mean BW of 828.8 ± 192.8 g and mean GA of 27 ± 1.8 weeks. All infants who developed ROP fulfilled both or one criterion of ROP screening. No infant greater than 1350 g BW or more than 31 weeks GA was treated for ROP.
Current ROP screening criteria of BW of 1500 g or less and/or GA of 32 weeks or less seems reasonable in our set up as no infant having ROP was missed by using these criteria. We do not recommend lowering or using only one index for ROP screening.
closed_qa
Are physicians underestimating the challenges of hypertension management?
The Supporting Hypertension Awareness and Research Europe-wide (SHARE) survey aimed to qualify the key challenges that physicians face when trying to get patients to recommended blood pressure (BP) goals. The survey was open to physicians involved in the treatment of hypertension, was anonymous, and included 45 questions covering: physicians' demographic information, familiarity with BP treatment guidelines, views on the BP targets recommended by the 2007 European Society of Hypertension and European Society of Cardiology (ESH-ESC) guidelines, and perceptions on the proportion of 'challenging patients' in hypertension management (defined as patients not achieving the BP goal, where the BP goal is at least<140/90 mm Hg, and<130/80 mm Hg for patients with co-morbidities or high CV risk). Physicians significantly underestimated the proportions of their 'challenging patients' with hypertension compared with their perceptions of the proportions achieving 2007 ESH-ESC BP targets (p<0.0001). The majority of cardiologists (75.5%) and general/family practitioners (GPs) (81.3%) as well as internists (59.3%) (p<0.05 for cardiologists and GPs vs internists) felt that it was a challenge to get their patients to target BP, stating that only 43.2%, 57.4% and 38.2% of their patients, respectively, achieved these targets in practice (p<0.05 for GPs vs cardiologists and internists).
Physicians may underestimate the proportion of 'challenging patients' with hypertension and there is a need to improve their BP control. Increasing physicians' awareness about the risks of uncontrolled BP and improving compliance are two possible ways to improve management of hypertension.
closed_qa
The pattern, clinical characteristics and outcome of ESRD in Ile-Ife, Nigeria: is there a change in trend?
The prevalence of chronic renal failure and End Stage Renal Disease (ESRD) has remained high worldwide and the epidemiology has changed significantly in the last decade in industrialised countries. While there have been significant improvements in these patient's outcomes in developed countries, their state and survival is still appalling in developing countries. To determine the clinical pattern, presentation and management outcomes in our ESRD population over a 19-year period (1989-2007). Seven hundred and sixty patients' records were reviewed. Data on major causes, clinical presentation, management and survival were retrieved and collated. Data was analysed using SPSS package version 16. Their ages ranged between 15-90 years (mean ± SD; 39.9±1.67years) with male preponderance (70.3%). Major presenting complaints were body swelling and uraemic symptoms in most studied patients. The predisposing conditions included chronic glomerulonephritis, hypertension, obstructive uropathy and diabetes mellitus. Renal replacement therapy offered included HD in 556(73.2%), Continous Ambulatory Peritoneal Dialysis (CAPD) in only 9(1.2%) patients and renal transplantation in only 7(0.9%). Only 38(6.8%) survived on HD for longer than three months while 7(77.8%) CAPD patients and all transplanted patients survived for between six months and four years (p<0.00001). Median duration of survival after diagnosis for all the patients was 2 weeks (range 0-50 months).
End stage renal disease is still prevalent with chronic glomerulonephritis and hypertension being the common causes. Prognosis is still grave hence subsidized renal replacement therapy and preventive nephrology should be targeted in such underserved populations.
closed_qa
Does transrectal color Doppler ultrasound improve the diagnosis of prostate cancer?
To prospectively evaluate whether TRUS guided biopsy associated with Color Doppler (CD) imaging improves the detection of PCa. From January 2008 to December 2010, 144 subjects, with an increased PSA value or with a suspect digital rectal examination, were enrolled. Transrectal grey-scale Ultrasound (US) and CD examination were performed in all patients. CD US was considered positive or negative on the basis of the presence or absence of vascular abnormality. Prostate biopsy was performed immediately after grey-scale and Doppler evaluation, with a mean of 10 core-biopsy for each patient as well as a selective biopsy of all US abnormal areas (hypoechoic lesion or CD abnormality areas). PCa has been detected in 71 (49.3%) patients. 58 of the 71 patients had a hypoechoic area at US scan and 27 had a CD abnormality. The PSA value was<4 ng/ml in 11 patients (Group 1), in 63 patients PSA ranged between 4 and 10 ng/ml (Group 2) and in 70 patients PSA was greater than 10 ng/ml (Group 3). The detection rate was 36.7, 36.5 and 62.8% respectively. In Group 1 we detected 5 hypoechoic areas and 4 CD abnormal areas. Moreover 6 of 11 patients had a positive DRE. In the Group 2, 20 patients were positive to DRE; we visualized 21 hypoechoic areas and 7 CD abnormality. In the Group 3, 38 patients had a positive DRE, with 32 hypoechoic areas and 16 CD abnormalities found. We obtained 1537 total bioptic cores, 1440 randomly from peripheral gland, 70 from hypoechoic areas and 27 from abnormal CD flow areas. The detection rate was 17.1, 65.7 and 22.2% respectively.
CD US showed to be a complement to grey-scale imaging of prostate unless insufficiently sensitive to replace the standard systematic 8-12 core random peripherally biopsy. Furthermore it should be associated routinely to TRUS to easily focus suspect areas.
closed_qa
Paradox of schizophrenia genetics: is a paradigm shift occurring?
Genetic research of schizophrenia (SCZ) based on the nuclear genome model (NGM) has been one of the most active areas in psychiatry for the past two decades. Although this effort is ongoing, the current situation of the molecular genetics of SCZ seems disappointing or rather perplexing. Furthermore, a prominent discrepancy between persistence of the disease at a relatively high prevalence and a low reproductive fitness of patients creates a paradox. Heterozygote advantage works to sustain the frequency of a putative susceptibility gene in the mitochondrial genome model (MGM) but not in the NGM. We deduced a criterion that every nuclear susceptibility gene for SCZ should fulfill for the persistence of the disease under general assumptions of the multifactorial threshold model. SCZ-associated variants listed in the top 45 in the SZGene Database (the version of the 23rd December, 2011) were selected, and the distribution of the genes that could meet or do not meet the criterion was surveyed. 19 SCZ-associated variants that do not meet the criterion are located outside the regions where the SCZ-associated variants that could meet the criterion are located. Since a SCZ-associated variant that does not meet the criterion cannot be a susceptibility gene, but instead must be a protective gene, it should be linked to a susceptibility gene in the NGM, which is contrary to these results. On the other hand, every protective gene on any chromosome can be associated with SCZ in the MGM. Based on the MGM we propose a new hypothesis that assumes brain-specific antioxidant defenses in which trans-synaptic activations of dopamine- and N-methyl-d-aspartate-receptors are involved. Most of the ten predictions of this hypothesis seem to accord with the major epidemiological facts and the results of association studies to date.
The central paradox of SCZ genetics and the results of association studies to date argue against the NGM, and in its place the MGM is emerging as a viable option to account for genomic and pathophysiological research findings involving SCZ.
closed_qa
Decreased gray matter diffusivity: a potential early Alzheimer's disease biomarker?
Gray matter atrophy, an important biomarker for early Alzheimer's disease, might be due to white matter changes within gray matter. Twenty older participants with significant memory decline over a 12-year period (T12) were matched to 20 nondeclining participants. All participants were magnetic resonance imaging scanned at T12. Cortical thickness and diffusion tensor imaging analyses were performed. Lower cortical thickness values were associated with lower diffusion values in frontal and parietal gray matter areas. This association was only present in the memory decline group. The cortical thickness-diffusion tensor imaging correlations showed significant group differences in the posterior cingulate gyrus, precuneus, and superior frontal gyrus.
Decreased gray matter diffusivity in the posterior cingulate/precuneus area might be a disease-specific process and a potential new biomarker for early Alzheimer's disease. Future studies should validate its potential as a biomarker and focus on cellular changes underlying diffusivity changes in gray matter.
closed_qa
Can virtual reality simulation be used for advanced bariatric surgical training?
Laparoscopic bariatric surgery is a safe and effective way of treating morbid obesity. However, the operations are technically challenging and training opportunities for junior surgeons are limited. This study aims to assess whether virtual reality (VR) simulation is an effective adjunct for training and assessment of laparoscopic bariatric technical skills. Twenty bariatric surgeons of varying experience (Five experienced, five intermediate, and ten novice) were recruited to perform a jejuno-jejunostomy on both cadaveric tissue and on the bariatric module of the Lapmentor VR simulator (Simbionix Corporation, Cleveland, OH). Surgical performance was assessed using validated global rating scales (GRS) and procedure specific video rating scales (PSRS). Subjects were also questioned about the appropriateness of VR as a training tool for surgeons. Construct validity of the VR bariatric module was demonstrated with a significant difference in performance between novice and experienced surgeons on the VR jejuno-jejunostomy module GRS (median 11-15.5; P = .017) and PSRS (median 11-13; P = .003). Content validity was demonstrated with surgeons describing the VR bariatric module as useful and appropriate for training (mean Likert score 4.45/7) and they would highly recommend VR simulation to others for bariatric training (mean Likert score 5/7). Face and concurrent validity were not established.
This study shows that the bariatric module on a VR simulator demonstrates construct and content validity. VR simulation appears to be an effective method for training of advanced bariatric technical skills for surgeons at the start of their bariatric training. However, assessment of technical skills should still take place on cadaveric tissue.
closed_qa
Can skills coaches be used to assess resident performance in the skills laboratory?
The purpose of this study was to compare faculty ratings between live versus video-recorded resident performances and faculty versus skills coaches' ratings of video-recorded resident performances. PGY1 residents were observed, video-recorded, and rated during a Verification of Proficiency examination on 4 stations (ie, suturing, laparotomy, central line, and cricothyroidotomy). One surgeon and 2 trained skills coaches independently rated each video-recorded performance (N = 25). The chi-square test was used to compare checklist ratings. Analysis of variance was used to compare global ratings. Intraclass correlations were used to evaluate inter-rater agreement. There were no statistical differences in faculty checklist ratings for live versus video-recorded performances (P>.05), and we found a nearly perfect interrater agreement, intraclass correlation coefficient (ICC) = 0.99 (P<.001). When comparing faculty versus skills coaches' ratings on video-recorded performances, we found no differences for the global or checklist ratings. Inter-rater agreement was moderately high for the global ratings, ICC = 0.71 (P<. 0.01, 95% confidence interval 0.23-0.96), and nearly perfect for the checklist ratings, ICC = 0.99 (P<.001, 95% confidence interval 0.94-1.00).
When assessing residents' performances, use of video-recorded performance ratings and skills coaches may be viable alternatives to live ratings performed by surgical faculty.
closed_qa
Is a closed bladder neck on preoperative videourodynamic studies an important factor for continence following augmentation ileocystoplasty in myelodysplastic patients?
The aim of this study is to evaluate the importance of a closed bladder neck during videourodynamic (VUDE) studies in relation to urinary continence following augmentation ileocystoplasty in myelodysplastic patients. We retrospectively reviewed the records of 24 myelodysplastic patients who underwent augmentation ileocystoplasty, using a standard technique. All patients had a closed bladder neck during preoperative VUDE studies. Their charts, imaging studies and VUDE data before and after surgery were analyzed. The mean follow-up after augmentation ileocystoplasty was 8.4 years. The overall incidence of urinary incontinence following the augmentation ileocystoplasty was 12.5%. Continence was achieved in 21 of 24 (87.5%) patients without additional outlet procedures. No significant upper tract changes developed. A clinically apparent tethered cord significantly hindered the achievement of continence. No significant correlation was found between the other videourodynamic parameters and obtaining continence.
Our study provides evidence that a coexisting cord tethering in this myelodysplastic group can affect bladder neck morphology and function, and subsequent continence.
closed_qa
Is ultrasound follow-up necessary in humero-axillary prosthetic arteriovenous fistulas for haemodialysis?
Retrospective study. Between January 2005 and December 2009, 108 Hax-AVF were implanted. From June 2007 a T-AVF was established. A preoperative duplex was performed and a follow-up control carried out a month after the intervention and subsequently every 3 months. An analysis was made of the permeability of 57 Hax-AVF carried out between June 2007 and December 2009 (T-AVF Group), in comparison to 51 interventions performed during the previous 30 months (Control Group). No differences in the permeability achieved were found at 12 and 24 months, with a secondary permeability at 12 months of 49% in the T-AVF Group and 52% in the Control Group. The percentage of patients needing to be re-operated was inferior in the T-AVF Group (35%) than in the Control Group (67%) p=0.02. The re-operation per patient average was lower in the T-AVF Group than in the Control Group (0.49 vs. 1.18 p=0.01). The patients of the TAVF Group underwent a lesser number of re-operations for obstruction as opposed to the Control Group (0.42 vs 1.04 p=0.01).
In our experience, the intensive follow-up controls did not improve the permeability of the Hax-AVF, although re-operations due to obstruction did diminish. The follow-up of these access fistulas should be clinical based on hemodialysis data, leaving ultrasonographic evaluation for those cases where a malfunction is suspected.
closed_qa
Does aggressive phototherapy increase mortality while decreasing profound impairment among the smallest and sickest newborns?
Aggressive phototherapy (AgPT) is widely used and assumed to be safe and effective for even the most immature infants. We assessed whether the benefits and hazards for the smallest and sickest infants differed from those for other extremely low-birth-weight (ELBW; ≤ 1000 g) infants in our Neonatal Research Network trial, the only large trial of AgPT. ELBW infants (n=1974) were randomized to AgPT or conservative phototherapy at age 12 to 36 h. The effect of AgPT on outcomes (death, impairment, profound impairment, death or impairment (primary outcome), and death or profound impairment) at 18 to 22 months of corrected age was related to BW stratum (501 to 750 g; 751 to 1000 g) and baseline severity of illness using multilevel regression equations. The probability of benefit and of harm was directly assessed with Bayesian analyses. Baseline illness severity was well characterized using mechanical ventilation and FiO(2) at 24 h age. Among mechanically ventilated infants ≤ 750 g BW (n=684), a reduction in impairment and in profound impairment was offset by higher mortality (P for interaction<0.05) with no significant effect on composite outcomes. Conservative Bayesian analyses of this subgroup identified a 99% (posterior) probability that AgPT increased mortality, a 97% probability that AgPT reduced impairment, and a 99% probability that AgPT reduced profound impairment.
Findings from the only large trial of AgPT suggest that AgPT may increase mortality while reducing impairment and profound impairment among the smallest and sickest infants. New approaches to reduce their serum bilirubin need development and rigorous testing.
closed_qa
Diagnostic implications of informant disagreement about rage outbursts: bipolar disorder or another condition?
Modest agreement between parent- and teacher-reports of child behavior is a common finding. This study examines diagnoses made when significant disparity occurs in parent- and teacher-reports of rage behaviors. Parents and teachers of 911 5-18 year-olds referred for psychiatric outpatient services completed rating scales and received a psychiatric evaluation blind to parent- and teacher-ratings. Children with rage outbursts (n=431, 47.2%) were assessed for diagnosis, family history, and clinical variables. Children were 12.0 (3.6) years; 26.5% were female. Bipolar disorder was rare (11.2%) in this sample; however, in children with parent- and teacher-reported rages, severe mood dysregulation was the most common condition (54.4%). In parent only reported rages, anxiety disorders were most common (40.6%) diagnoses, and in teacher only reported rages, learning/language disorders were the most common (46.0%) diagnoses.
The context in which a rage outburst occurs may impact the diagnosis; however, diagnosis alone does not explain this difficult and impairing behavior.
closed_qa
Does the new standard for eugenol designed to protect against contact sensitization protect those sensitized from elicitation of the reaction?
Potential fragrance allergens used in daily products should have a concentration limited to levels that are at, or below, acceptable exposure levels based on the quantitative risk assessment for the induction of dermal sensitization. To date, there are insufficient data to discern any quantitative relationship between induction and elicitation concentrations for fragrance ingredients that have a potential for dermal sensitization. When available, these data should be used to confirm the effectiveness of quantitative risk assessment-based risk management procedures. In this study, the relationship between the allergen concentration and the time to elicit allergic contact dermatitis in eugenol-sensitized patients was studied. The products used to elicit allergic contact dermatitis had a concentration of eugenol that was equal to, or below, the International Fragrance Association standard. Volunteers with and without known sensitization to eugenol were patch tested with various concentrations of eugenol (dilution series) and also underwent repeated open application tests (ROATs). This study model has previously been successfully used with stronger sensitizers. In this study, allergic contact dermatitis, as evidenced by a positive ROAT, could not be elicited by any of the concentrations studied, including in those patients where the patch tests were positive.
When tested in a 3-week ROAT at, or below, its current International Fragrance Association Standard, eugenol did not induce reactions even in those known to be sensitized. Whether this represents a false-negative result for a weak allergen is unknown.
closed_qa
Is quaternium-15 a formaldehyde releaser?
The question of whether quaternium-15 is a formaldehyde releaser is controversial. Understanding this relationship is critical because of the widespread use of quaternium-15 and the need to properly advise formaldehyde-allergic individuals. This study aimed to look for an association between allergy to quaternium-15 and formaldehyde by correlating reactions to both and to correlate the intensity of positive patch test results to formaldehyde with reactivity to quaternium-15. This is a retrospective analysis of 1905 patients who underwent patch testing for allergic contact dermatitis. Associations were analyzed by χ testing. Of all patients, 9.5% reacted to quaternium-15, 7.2% reacted to formaldehyde, and 5.4% reacted to both (P<0.001). Of 137, 86 had strong (2 or 3+) and 51 had weak (1+ or +/-) formaldehyde reactions; there was no relationship between the severity of formaldehyde reactivity and responsiveness to quaternium-15 (P = 0.229). Sex analysis did not change these findings. This study is limited by its retrospective analysis and small sample size.
A statistically significant relationship exists between reactivity to quaternium-15 and formaldehyde; however, the severity of the formaldehyde reaction does not predict reactivity to quaternium-15. Despite coreactivity with formaldehyde, quaternium-15 may not be a significant formaldehyde releaser. The coreactivity between quaternium-15 and formaldehyde requires further studies.
closed_qa
Erythema papulosa semicircularis recidivans: a new reactive dermatitis?
Nine patients with recrudescent and centrifugally expanded papuloerythematous eruptions were observed in our outpatient department during the recent 8 years. The patients were all young and middle-aged men presenting with characteristic skin lesions and relapsing each year in the warm humid season. Such an observation has not yet been described in the English literature. The objective of this study was to describe the clinical characteristics and explore the possible etiologies of our cases. In addition, differential diagnosis with other common figurate erythemas was also reviewed and discussed. In 5 of the patients, skin-prick testing was performed with common airborne and food allergens, and skin patch testing with the Chinese baseline series of contact allergens was performed, along with histopathologic examinations. The skin findings in our patients were conspicuous in their papular characteristics of the borders, semicircular arrangement, male predominance, and yearly crescendo of recurrence in warm seasons. Histopathologic examination showed superficial perivascular dermatitis, whereas skin-prick testing and patch testing showed negative results.
Although the etiology of our cases remains unclear at present, the clinical characteristics make them distinct from other well-described figurate erythemas characterized by annular erythematous lesions. We propose the term erythema papulosa semicircularis recidivans to depict a special form of recurrent papuloerythematous figurate erythema of unknown etiology.
closed_qa
Is there any relationship between pulmonary function tests and post-transplant complications of allogeneic hematopoetic stem cell transplantation?
One hundred and fifty one patients who had undergone allogeneic HSCT between years 2003 through 2008, and had the records of PFTs prior to and at 1, 3, 6, 9 and 12 months after transplantation were included in the study. Prospectively collected data of these patients were analysed retrospectively. Although no significant difference was identified in other PFT parameters, a significant decrease in DLCOadj was determined after 1st and 3rd months of HSCT. A significant correlation was found between pretransplant DLCOadj value<%70 and sinusoidal obstruction syndrome (SOS) (P=0.001, r=0.323), but in multivariate analysis pretransplant DLCOadj was not an independent predictor of SOS; only total body irradiation (TBI) (OR: 3.673, %95 CI: 0.880-15.804), the day of platelet engraftment (OR=1.093, %95 CI: 1.029-1.161) and serum ferritin (OR=1.001, %95 CI: 1.000-1.001) were significant. Advancing age and serum ferritin levels>600 ng/mL were the independent risk factors for pretransplant DLCOadj<%70 (OR: 0.970, %95 CI: 0.941-0.999 and OR: 2.355, %95 CI: 1.058-5.241 respectively).
Although a significant correlation exists between pretransplant DLCOadj values and post-transplant SOS development, pretransplant DLCOadj was not an independent predictor of SOS. Increased serum ferritin levels were common both for pretransplant DLCO decrease and post-transplant SOS development. Iron induced endothelial damage may be the common pathophysiologic mechanism causing lung and liver vulnerability, and DLCOadj may be a non-invasive method of demonstrating this vulnerability.
closed_qa
Is the carotid intima-media thickness really a good surrogate marker of atherosclerosis?
To evaluate the relationship between carotid IMT and atherosclerosis, postmortem specimens of the distal segments of the left common carotid artery (CCA) from 133 Korean men aged from 20 to 78 years were used for histopathology and computer-assisted morphometry. Blood lipids and atherosclerosis-associated collagen and elastin were quantitatively analyzed. Correlation coefficients of IMT were smaller than those of intima thickness but IMT was well associated with age (r= 0.55, p<0.00001), atherosclerosis score (or grade, AS, r= 0.73, p<0.00001), plaque area (PA, r= 0.72, p<0.00001), total cholesterol (TC, r= 0.69, p<0.00001), low-density lipoprotein cholesterol (LDL-c, r= 0.72, p<0.00001) and triglyceride (TG, r= 0.38, p<0.001). Coronary artery stenosis (CAS) and coronary calcification were also well associated with age (p<0.00001), IMT (p<0.005) and PA (p<0.00001). When IMT was thicker than 1 mm, the possibility of carotid atherosclerosis accompanied with CAS and coronary calcification, TC, LDL-c and TG was much higher (CAS with coronary calcification,p<0.005; TC, p<0.00001; LDL-c, p<0.00005; TG, p<0.00001). Collagen tended to increase while elastin tended to decrease as AS increased (p<0.005); collagen increased and elastin decreased (p<0.00001) when comparing plaque to the plaque-free area in the same segment.
These results support that the carotid IMT in association with TC, LDL-c and TG can be used as a good surrogate marker of atherosclerosis and predictor of coronary heart disease. Plaque formation may influence significant quantitative changes in collagen and elastin.
closed_qa
Is Doppler ultrasonography essential for hemorrhoidal artery ligation?
Doppler ultrasonography enables accurate identification of the terminal branches of the superior rectal artery prior to hemorrhoidal artery ligation (HAL). However, since the positions of these branches have been found to be relatively constant, the question arises as to the necessity of ultrasonography for their identification. The aim of the current study was to examine the positions of all arteries identified and ligated during the HAL procedure. We recorded the position of all arteries located and ligated in 135 consecutive patients who underwent the HAL procedure during the years 2003 to 2006. In all patients, 6-8 terminal arterial branches were located above the dentate line. In 102 (76 %) patients, terminal branches were located in all 6 of the odd-numbered clock positions around the anus (1, 3, 5, 7, 9, and 11 o'clock in the lithotomy position). If we had ligated arteries only at these odd-numbered clock positions, without using Doppler ultrasonography, we would have located all the arteries in 96 (71 %) of our patients.
The number and location of arterial branches of the superior rectal artery are relatively constant. Nevertheless, if, Doppler ultrasonography had not been performed and, ligation in the HAL procedure had been at the odd-numbered clock positions only, then at least one artery would have been missed in 29 % of our patients.
closed_qa
Could chemical enhancement of gold nanoparticle penetration be extrapolated from established approaches for drug permeation?
Investigations on chemical enhancement of skin penetration of gold nanoparticles are considered crucial to have a deeper insight into the main barrier of particle penetration. In this study, penetration of gold nanoparticles in the presence of several chemical enhancers - urea, sodium lauryl sulphate, polysorbate 80 and dimethyl sulfoxide (DMSO) - through human skin was studied. Among the tested chemical enhancers, DMSO could induce the penetration of hydrophilic (citrate-stabilized) gold colloid of no intrinsic penetration ability in a concentration-dependent manner. Pretreatment of the skin with DMSO however reduced the penetration of hydrophobic (cetrimide-coated) gold nanoparticles as a result of aggregation in the top layers of the stratum corneum limiting penetration into the deeper skin layers. In addition, nanoparticles-vehicle interaction and the stability of the nanoparticles in the applied vehicle were found important determinants of skin penetration.
Our results demonstrate that the already established approaches for chemical permeation enhancement of drug molecules and their postulated mechanisms could be used as preliminary guidelines for enhancing the penetration of nanoparticles. At this size range of 15 nm, intercellular lipids provide the main barrier to particle penetration through the stratum corneum.
closed_qa
Tetrasomy 15q26: a distinct syndrome or Shprintzen-Goldberg syndrome phenocopy?
The aim of this study was to characterize the clinical phenotype of patients with tetrasomy of the distal 15q chromosome in the form of a neocentric marker chromosome and to evaluate whether the phenotype represents a new clinical syndrome or is a phenocopy of Shprintzen-Goldberg syndrome. We carried out comprehensive clinical evaluation of four patients who were identified with a supernumerary marker chromosome. The marker chromosome was characterized by G-banding, fluorescence in situ hybridization, single nucleotide polymorphism oligonucleotide microarray analysis, and immunofluorescence with antibodies to centromere protein C. The marker chromosomes were categorized as being neocentric with all showing tetrasomy for regions distal to 15q25 and the common region of overlap being 15q26→qter.
Tetrasomy of 15q26 likely results in a distinct syndrome as the patients with tetrasomy 15q26 share a strikingly more consistent phenotype than do the patients with Shprintzen-Goldberg syndrome, who show remarkable clinical variation.
closed_qa
ADHD predominantly inattentive subtype with high sluggish cognitive tempo: a new clinical entity?
The authors investigate differences in the neuropsychological and behavioral profiles of two groups of children with ADHD, one with predominantly inattentive subtype of ADHD (PI) and high sluggish cognitive tempo (SCT; n = 19) and another formed by the rest of the sample (children with ADHD combined subtype and children with PI and low SCT scores; n = 68). Instruments included Wechsler Intelligence Scale for Children and subtests from Developmental Neuropsychological Assessment, Conners' Continuous Performance Test, Behavior Rating Inventory of Executive Function, and Achenbach's Child Behavior Checklist for ages 6 to 18. PI with high SCT had fewer problems with sustained attention, and more internalizing problems, anxiety/depression, and withdrawn/depressed behavior, and more executive problems with self-monitoring than the rest of the ADHD sample.
This study supports revising subtype's criteria and further studying the hypothesis that ADHD with high SCT constitutes a separate clinical entity.
closed_qa
The production of social discourse on Hansen' disease and health education materials in Brazil: a skin patch as something harmless or a serious disease?
Hansen's disease is endemic in Brazil and government control programmes promote publicity campaigns to increase the detection of new cases through the production and distribution of educative material. This study analyses a set of 276 educational materials produced by governmental and non-governmental organisations that work to control Hansen's disease in Brazil. It describes the content of the materials and the way the issues were approached. It is a qualitative study that adopts the theoretical and methodological framework of the semiology of social discourse. Analysis reveals that the relations between the enunciator and recipient of the materials are asymmetrical as a result of the technical and educational language employed. Biomedical information forms the basis for social representations an practices of Hansen's disease, as opposed to historical collective knowledge of 'leprosy'. The prioritised topics are: signs and symptoms of the disease, treatment stigma, cure and surveillance.
The institutionalisation of public education on Hansen's disease in Brazil was not limited simply to the change of terminology from 'leprosy' to 'Hansen's disease,' but was shaped also by new educational practices. It is recommended that the evaluation and production of new materials be incorporated into the set of activities already carried out in health centres so as to expand the discussion on content, language and the best way to address the disease in the materials.
closed_qa
Mini-sternotomy aortic valve replacement: is it safe and effective?
Mini-sternotomy aortic valve replacement (MSAVR) has been increasingly performed at the authors' institution since October 2003. The study aim was to compare results obtained with MSAVR to those following AVR with conventional sternotomy (SAVR). Between 1998 and 2008, a total of 143 consecutive patients (mean age: 67 +/- 12.5 years) underwent AVR at the authors' institution. Of these patients, 82 underwent SAVR, and 61 underwent MSAVR performed through a reversed-L-shaped median sternotomy with a transverse limb at the right fourth intercostal space. Ascending aortic and right atrial cannulation through the mini-sternotomy were employed for cardiopulmonary bypass (CPB). Typically, the MSAVR patients were slightly younger than SAVR patients (mean age: 67 +/- 16 years and 70 +/- 15 years, respectively; p = 0.037), had a lower incidence of diabetes (3% versus 18%, p = 0.008), and a slightly higher left ventricular ejection fraction (74.5 +/- 12% versus 71 +/- 12%, p = 0.019). There were no other inter-group preoperative differences. As expected, MSAVR required a slightly longer aortic cross-clamp time (49 +/- 19 min) compared to SAVR (44.5 +/- 16 min; p = 0.019), and longer CPB times (77 +/- 31 min versus 60 +/- 26 min; p<0.0001), though the overall operating times were similar (p = 0.38). Postoperatively, MSAVR patients were extubated at 3 +/- 5 h, similar to SAVR patients (4 +/- 5 h) (p = 0.13). The median intensive therapy unit stay was 1 +/- 1 days in both groups. The median hospital stay was comparable between groups (MSAVR, 7 +/- 5 days; SAVR, 8 +/- 4 days; p = 0.48). The MSAVR patients had a higher incidence of delayed pericardial effusions requiring pericardiocentesis (n = 4; p = 0.031), but this did not affect survival. The 30-day mortality was similar in both groups (MSAVR group, n = 1 (1.6%); SAVR group, n = 3 (3.7%); p = 0.64). At five years after surgery, freedom from cardiac-related death was 96 +/- 2.6% in MSAVR patients, and 89 +/- 4.9% in SAVR patients (p = 0.32).
Mini-sternotomy AVR is technically challenging with longer CPB and aortic cross-clamp times. However, with increasing surgical experience, it offers results comparable to those achieved with conventional AVR, and with acceptable cosmetic results.
closed_qa
Is posttraumatic cerebrospinal fluid fistula a predictor of posttraumatic meningitis?
Various factors have been reported in literature to be associated with the development of posttraumatic meningitis. There is a paucity of data regarding skull fractures and facial fractures leading to CSF leaks and their association with the development of meningitis. The primary objective of this study was to analyze the US Nationwide Inpatient Sample (NIS) database to elucidate the factors associated with the development of posttraumatic meningitis. A secondary goal was to analyze the overall hospitalization cost related to posttraumatic meningitis and factors associated with that cost. The NIS database was analyzed to identify patients admitted to hospitals with a diagnosis of head injury from 2005 through 2009. This data set was analyzed to assess the relationship of various clinical parameters that may affect the development of posttraumatic meningitis using binary logistic regression models. Additionally, the overall hospitalization cost for the head injury patients who did not undergo any neurosurgical intervention was further categorized into quartile groups, and a regression model was created to analyze various factors responsible for escalating the overall cost of the hospital stay. A total of 382,267 inpatient admissions for head injury were analyzed for the 2005-2009 period. Meningitis was reported in 0.2% of these cases (708 cases). Closed skull base fractures, open skull base fractures, cranial vault fractures, and maxillofacial fractures were reported in 20,524 (5.4%), 1089 (0.3%), 5064 (1.3%), and 88,649 (23.2%) patients, respectively. Among these patients with fractures, meningitis was noted in 0.17%, 0.18%, 0.05%, and 0.10% admissions, respectively. Cerebrospinal fluid rhinorrhea was reported in 453 head injury patients (0.1%) and CSF otorrhea in 582 (0.2%). Of the patients reported to have CSF rhinorrhea, 35 (7.7%) developed meningitis, whereas in the cohort with CSF otorrhea, 15 patients (2.6%) developed meningitis. Cerebrospinal fluid rhinorrhea (p<0.001, OR 22.8, 95% CI 15.6-33.3), CSF otorrhea (p<0.001, OR 9.2, 95% CI 5.2-16.09), and major neurosurgical procedures (p<0.001, OR 5.6, 95% CI 4.8-6.5) were independent predictors of meningitis. Further, CSF rhinorrhea (p<0.001, OR 2.0, 95% CI 1.6-2.7), CSF otorrhea (p<0.001, OR 2.3, 95% CI 1.9-2.7), and posttraumatic meningitis (p<0.001, OR 3.1, 95% CI 2.5-3.8) were independent factors responsible for escalating the cost of head injury in cases not requiring any major neurosurgical intervention.
Cerebrospinal fluid rhinorrhea and CSF otorrhea are independent predictors of posttraumatic meningitis. Furthermore, meningitis and CSF fistulas may independently lead to significantly increased cost of hospitalization in head injury patients not undergoing any major neurosurgical intervention.
closed_qa
Myoinositol: does it improve sperm mitochondrial function and sperm motility?
To evaluate whether an improvement in mitochondrial membrane potential was associated with sperm motility amelioration and greater sperm recovery after the swim-up procedure. A second purpose was to evaluate the effects of myoinositol (MYO) on sperm apoptosis, quality of chromatin compaction, and DNA integrity. Spermatozoa from 20 normozoospermic men and 20 patients with oligo-astheno-teratozoospermia were incubated in vitro with 2 mg/mL of MYO or phosphate-buffered saline as a control for 2 hours. After this incubation period, sperm motility was evaluated. Flow cytometry was used to analyze the mitochondrial membrane potential, phosphatidylserine externalization, chromatin compactness, and DNA fragmentation. We also evaluated the total number of motile spermatozoa recovered after swim-up after incubation with MYO or phosphate-buffered saline. MYO significantly increased the percentage of spermatozoa with progressive motility in both normozoospermic men and patients with oligo-astheno-teratozoospermia. Motility improvement in the latter group was associated with a significant increase in the percentage of spermatozoa with high mitochondrial membrane potential. MYO had no effects on mitochondrial function in spermatozoa from normozoospermic men. Sperm phosphatidylserine externalization, chromatin compactness, and DNA fragmentation were unaffected by MYO in both groups. After incubation with MYO, the total number of spermatozoa recovered after swim-up had improved significantly in both groups.
These data show that MYO increases sperm motility and the number of spermatozoa retrieved after swim-up in both normozoospermic men and patients with abnormal sperm parameters. In patients with oligo-astheno-teratozoospermia, the improvement in these parameters was associated with improved sperm mitochondrial function. These findings support the use of MYO in both in vivo- and in vitro-assisted reproductive techniques.
closed_qa
Can conventional magnetic resonance imaging, prostate needle biopsy, or their combination predict the laterality of clinically localized prostate cancer?
To evaluate conventional magnetic resonance imaging (MRI), prostate needle biopsy (PBx), and the combination of both tests in predicting the laterality of final specimen pathology after radical prostatectomy. A total of 574 radical prostatectomy cases that had PBx with at least 12 cores and preoperative prostate MRI with pelvic coil were included. We analyzed the clinicopathologic data with laterality based on PBx and MRI. Unilateral disease in combination was defined as unilateral cancer in PBx and at the same time MRI findings of undetectable or ipsilateral disease. Cohen's kappa (κ) was used to measure agreement between the laterality data. There were a total of 316 (55.1%) unilateral cancers detected by PBx, whereas there were 139 (24.2%) cases in the final specimen pathology. MRI resulted in 119 (20.7%) undetectable and 205 (35.7%) unilateral cancers. Cancer laterality based on final specimen pathology had only fair agreements with PBx (κ = 0.286), MRI (κ = 0.200), and their combination (k = .291). The positive predictive values to predict pathologic concurrent unilaterality were only 30.4% (96/316), 25.9% (53/205), and 34.8% (72/207), respectively. These trends were similar in low-risk cases.
Preoperative PBx, MRI, and the combination of both methods had only a fair correlation with the laterality of prostate cancer (PC), even in low-risk cases. Approximately two thirds of cases diagnosed as unilateral disease by contemporary PBx, MRI, or their combination were not concurrent unilateral disease in final pathology. This should be recognized when planning nerve-sparing surgery and potentially for candidate selection for focal therapy to treat PC.
closed_qa
Is co-morbid depression adequately treated in patients repeatedly referred to specialist medical services with symptoms of a medical condition?
Patients with a medical condition and co-morbid depression have more symptoms and use more medical services. We aimed to determine the prevalence of depression and the adequacy of its treatment in patients who had been repeatedly referred from primary to specialist medical care for the assessment of a medical condition. All patients who had at least three referrals to medical and surgical specialists for an assessment of symptoms attributed to a medical condition, over a five year period from five primary care practices in Edinburgh, UK were identified using a referral database and review of records. Participants were sent a questionnaire which included the PHQ-9 depression scale and additional questions about depression during the preceding 5years. Details of treatment for depression were obtained from primary care records. Questionnaires were sent to 230 patients and returned by 162 (70.4%). Forty-one (25.3%) had a PHQ-9 score of 10 or more and hence probable current depressive disorder. An additional 36 (22.2%) reported depression in the previous 5years. Only eight (19.5%) of those reporting current depression and 20 (26%) of the 77 patients reporting previous depression had received minimally adequate treatment for it.
Whilst we know that patients with medical conditions are often depressed and that such co-morbid depression is often undertreated, we have found that it is undertreated even in patients repeatedly referred to medical specialists. Better assessment and management of depression in such patients could both improve patients' quality of life and reduce the cost of care.
closed_qa
Are therapeutics decisions homogeneous in multidisciplinary onco-urology staff meeting?
One of the priorities of the "Plan against the Cancer" in France is to ensure the discussion of all cancer cases in a multidisciplinary meeting staff (RCP). The multidisciplinary collaboration is proposed to guarantee a discussion between specialists in every cases, particularly in complex cases. The aim of this study was to compare the therapeutic decision taken in four RCP in Paris Île-de-France academic centres for three identical cases. Three cases of urological oncology (prostate cancer [PCa], renal cell carcinoma [RCC]and bladder tumour) were selected by a single urologist, not involved in further discussion. These cases were blindly presented in four academic urology department from Paris: Pitié-Salpêtrière Hospital, Mondor Hospital, the Georges-Pompidou European Hospital and Foch Hospital. The four centres met the criteria of quality of RCP in terms of multidisciplinarity, frequency and standardization. The therapeutic suggestions were similar in the RCC cases, there were differences in the surgical approaches and preoperative work-up in the PCa case and, lastly, the proposals were different for the bladder cancer case.
The decisions relies on clinical data and preoperative work-up but also on the experience and habits of the centre of excellence. For complex cases that does not fit with current guidelines, the panel discussion can lead to different therapeutic options from a centre to another and is largely influenced by the local organisation of the RCP.
closed_qa
Does ED crowding decrease the number of procedures a physician in training performs?
The aim of the study was to determine whether the number of procedures performed by residents and medical students in the emergency department (ED) is affected by ED crowding. In this single-center, prospective, observational study, standardized data collection forms were completed by both trainees and supervising emergency physicians (EPs) at the end of each ED shift from August 2009 to March 2010. Shifts with no trainees were excluded. All procedures that were offered to a trainee were recorded as well as the number of potential ED procedures that were, instead, referred to a consulting service. Emergency department crowding was measured in 2 ways: ED length of stay (LOS) and the EP's assessment of crowding during the shift. Poisson regression was used to assess the adjusted effect of ED crowding on the number of trainee procedures performed as well as on the number of procedures given away. There were 804 procedures performed by 113 trainees during 647 trainee shifts. Medical students comprised 51% of trainees. Median number of procedures performed per shift was 1.0 (Fine interquartile range, 0-2.0). Emergency department crowding was not associated with the adjusted number of procedures trainees performed using either the EP's assessment of crowding (P = .52) or ED LOS (P = .84). Emergency department crowding was associated with an adjusted 256% increase in the mean number of procedures given away (P = .02) when measured using physician assessment but was not associated with crowding when assessed using ED LOS (P = .06).
Crowding was not significantly associated with the number of procedures availed to ED trainees. In patients being considered for admission, however, when the managing EP felt that it was crowded, there was an association with giving procedures to consulting services.
closed_qa