Refine
Year of publication
Document Type
- Article (394) (remove)
Language
- English (394) (remove)
Keywords
- Nutritional footprint (5)
- Danube (4)
- Lemnaceae (4)
- Sustainability (4)
- Sustainable nutrition (4)
- selenium (4)
- Amino acid (3)
- Electroporation (3)
- Energy (3)
- Hay transfer (3)
Institute
- Fakultät WiSo (206)
- Fakultät AuL (146)
- Institut für Duale Studiengänge (17)
- Fakultät IuI (13)
- Institut für Management und Technik (10)
- Fakultät MKT (1)
- Institut für Kommunikationsmanagement (1)
- Präsidium Hochschule Osnabrück (1)
Objectives: To measure and assess the economic impact of adherence to a single quality indicator (QI) regarding weaning from invasive ventilation.
Design: Retrospective observational single-centre study, based on electronic medical and administrative records.
Setting: Intensive care unit (ICU) of a German university hospital, reference centre for acute respiratory distress syndrome.
Participants: Records of 3063 consecutive mechanically ventilated patients admitted to the ICU between 2012 and 2017 were extracted, of whom 583 were eligible adults for further analysis. Patients’ weaning protocols were evaluated for daily adherence to quality standards until ICU discharge. Patients with <65% compliance were assigned to the low adherence group (LAG), patients with ≥65% to the high adherence group (HAG).
Primary and secondary outcome measures: Economic healthcare costs, clinical outcomes and patients’ characteristics.
Results: The LAG consisted of 378 patients with a median negative economic results of −€3969, HAG of 205 (−€1030), respectively (p<0.001). Median duration of ventilation was 476 (248; 769) hours in the LAG and 389 (247; 608) hours in the HAG (p<0.001). Length of stay (LOS) in the LAG on ICU was 21 (12; 35) days and 16 (11; 25) days in the HAG (p<0.001). LOS in the hospital was 36 (22; 61) days in the LAG, and within the HAG, respectively, 26 (18; 48) days (p=0.001).
Conclusions: High adherence to this single QI is associated with better clinical outcome and improved economic returns. Therefore, the results support the adherence to QI. However, the examined QI does not influence economic outcome as the decisive factor.
Background
Lay family caregivers of patients receiving palliative care often confront stressful situations in the care of their loved ones. This is particularly true for families in the home-based palliative care settings, where the family caregivers are responsible for a substantial amount of the patient’s care. Yet, to our knowledge, no study to date has examined the family caregivers’ exposure to critical events and distress with home-based palliative care has been reported from Germany. Therefore, we attempt to assess family caregiver exposure to the dying patient’s critical health events and relate that to the caregiver’s own psychological distress to examine associations with general health within a home-based palliative care situation in Germany.
Methods
A cross-sectional study was conducted among 106 family caregivers with home-based palliative care in the Federal State of North Rhine Westphalia, Germany. We administered the Stressful Caregiving Adult Reactions to Experiences of Dying (SCARED) Scale. Descriptive statistics and linear regression models relating general health (SF-36) were used to analyze the data.
Results
The frequency of the caregiver’s exposure, or witness of, critical health events of the patient ranged from 95.2% “pain/discomfort” to 20.8% “family caregiver thought patient was dead”. The highest distress scores assessing fear and helpfulness were associated with “family caregiver felt patient had enough’” and “family caregiver thought patient was dead”. Linear regression analyses revealed significant inverse associations between SCARED critical health event exposure frequency (beta = .408, p = .025) and total score (beta = .377, p = .007) with general health in family caregivers.
Conclusions
Family caregivers with home-based palliative care in Germany frequently experience exposure to a large number of critical health events in caring for their family members who are terminally ill. These exposures are associated with the family caregiver’s degree of fear and helplessness and are associated with their worse general health. Thus the SCARED Scale, which is brief and easy to administer, appears able to identify these potentially upsetting critical health events among family caregivers of palliative care patients receiving care at home. Because it identified commonly encountered critical events in these patients and related them to adverse general health of family caregivers, the SCARED may add to clinically useful screens to identify family caregivers who may be struggling.
Aim: The aim of this study was to give an overview of family caregiver satisfaction within the home palliative care situation in North Rhine-Westphalia, Germany.
Design: A cross-sectional study.
Methods: An anonymous questionnaire, with seven validated scales, and comprising of 71 items, was used. The items investigated perceived needs and burdens of families within a home-based palliative care situation.
The satisfaction of the family caregivers with the services delivered by palliative care teams was measured by the FAMCARE-2 Scale. Descriptive statistics and analysis of covariance (ANCOVA) were performed.
Results: A convenience sample of 106 family members agreed to participate in the study. Overall, we found high satisfaction within our sample. There was high satisfaction with how the services respected the dignity of families, and how they provided comfort to patients. Satisfaction was lower with regard to information about patients.
Conclusion: High or low satisfaction with palliative care, tells us little about the quality of services. The high satisfaction within this study could be interpreted as a sign that palliative care was important to families at the time of availability. Regular and continuous assessment can serve to inform the continuous quality of care provision for patients and their families.
Background
Forced migration significantly endangers health. Women face numerous health risks, including sexual violence, lack of contraception, sexually transmitted disease, and adverse perinatal outcomes. Therefore, sexual and reproductive healthcare is a significant aspect of women asylum seekers’ health.
Even when healthcare costs of asylum seekers are covered by the government, there may be strong barriers to healthcare access and specific needs may be addressed inadequately. The study’s objectives were a) to assess the accommodation and healthcare services provided to women asylum seekers in standard and specialised health care, b) to assess the organisation of healthcare provision and how it addresses the sexual and reproductive healthcare needs of women asylum seekers.
Methods
The study utilised a multi-method approach, comprising a less-dominant quantitative component and dominant qualitative component. The quantitative component assessed accommodation conditions for women in eight asylum centres using a survey. The qualitative component assessed healthcare provision on-site, using semi-structured interviews with health and social care professionals (n = 9). Asylum centres were selected to cover a wide range of characteristics. Interview analysis was guided by thematic analysis.
Results
The accommodation in the asylum centres provided gender-separate rooms and sanitary infrastructure. Two models of healthcare were identified, which differed in the services they provided and in their organisation: 1) a standard healthcare model characterised by a lack of coordination between healthcare providers, unavailability of essential services such as interpreters, and fragmented healthcare, and 2) a specialised healthcare model specifically tailored to the needs of asylum-seekers. Its organisation is characterised by a network of closely collaborating health professionals. It provided essential services not present in the standard model. We recommend the specialised healthcare model as a guideline for best practise.
Conclusions
The standard, non-specialised healthcare model used in some regions in Switzerland does not fully meet the healthcare needs of women asylum seekers. Specialised healthcare services used in other regions, which include translation services as well as gender and culturally sensitive care, are better suited to address these needs. More widespread use of this model would contribute significantly toward protecting the sexual and reproductive integrity and health of women asylum seekers.
This review aimed to synthesize the current evidence on the effectiveness of invasive treatments for complex regional pain syndrome in children and adolescents. Studies on children and adolescents with complex regional pain syndrome that evaluated the effects of invasive treatment were identified in PubMed (search March 2013). Thirty-six studies met the inclusion criteria. Articles reported on a total of 173 children and adolescents with complex regional pain syndrome. Generally, many studies lack methodological quality. The invasive treatments applied most often were singular sympathetic blocks, followed by epidural catheters and continuous sympathetic blocks. Rarely, spinal cord stimulation and pain-directed surgeries were reported. An individual patient frequently received more than one invasive procedure. Concerning outcome, for approximately all patients, an improvement in pain and functional disability was reported. However, these outcomes were seldom assessed with validated tools. In conclusion, the evidence level for invasive therapies in the treatment of complex regional pain syndrome in children and adolescents is weak.
Currently, the treatment of musicians is an interprofessional approach. Playing-related health complaints may impact the performance of a musician. In Germany, a medical consulting hour for musicians exists, but those for athletes in sports medicine are not so common. The diagnosing and treatment procedure within the physiotherapy consultation for musicians follows a specific concept-b and requires knowledge of instruments and musician-specific complaints. Based on the consulting hour in a clinic in Osnabrueck, 614 case reports were part of this sample, of which 558 data sets were complete. The focus of the analysis is the instrument and the primary complaint. Also, the type of therapy is characterized, and the amount is calculated. Primary complaints of musicians, in general, are found most frequently in the spine and upper extremity. Musician complaints are different between instruments. Instrumentalists have a significantly higher chance to suffer from a primary complaint in the area of the upper extremity. Furthermore, the groups without an instrument (e.g., singing or dancing) are developing complaints in the anatomical area which they primarily use. Therefore, these types of therapy were used: physiotherapy, manual therapy, and osteopathy with an average of 5.9 treatment units. This study underpinned the importance of musician-specific physiotherapy as a profession to treat musicians. Also, an interdisciplinary approach is necessary to treat all aspects of complaints.
Introduction Postoperative delirium (POD) is seen in approximately 15% of elderly patients and is related to poorer outcomes. In 2017, the Federal Joint Committee (Gemeinsamer Bundesausschuss) introduced a ‘quality contract’ (QC) as a new instrument to improve healthcare in Germany. One of the four areas for improvement of in-patient care is the ‘Prevention of POD in the care of elderly patients’ (QC-POD), as a means to reduce the risk of developing POD and its complications.
The Institute for Quality Assurance and Transparency in Health Care identified gaps in the in-patient care of elderly patients related to the prevention, screening and treatment of POD, as required by consensus-based and evidence-based delirium guidelines. This paper introduces the QC-POD protocol, which aims to implement these guidelines into the clinical routine. There is an urgent need for well-structured, standardised and interdisciplinary pathways that enable the reliable screening and treatment of POD. Along with effective preventive measures, these concepts have a considerable potential to improve the care of elderly patients.
Methods and analysis The QC-POD study is a non-randomised, pre–post, monocentric, prospective trial with an interventional concept following a baseline control period. The QC-POD trial was initiated on 1 April 2020 between Charité-Universitätsmedizin Berlin and the German health insurance company BARMER and will end on 30 June 2023. Inclusion criteria: patients 70 years of age or older that are scheduled for a surgical procedure requiring anaesthesia and insurance with the QC partner (BARMER). Exclusion criteria included patients with a language barrier, moribund patients and those unwilling or unable to provide informed consent. The QC-POD protocol provides perioperative intervention at least two times per day, with delirium screening and non-pharmacological preventive measures.
Ethics and dissemination This protocol was approved by the ethics committee of the Charité-Universitätsmedizin, Berlin, Germany (EA1/054/20). The results will be published in a peer-reviewed scientific journal and presented at national and international conferences.
A brief questionnaire for measuring alarm fatigue in nurses and physicians in intensive care units
(2023)
When exposed to hundreds of medical device alarms per day, intensive care unit (ICU) staff can develop “alarm fatigue” (i.e., desensitisation to alarms). However, no standardised way of quantifying alarm fatigue exists. We aimed to develop a brief questionnaire for measuring alarm fatigue in nurses and physicians. After developing a list of initial items based on a literature review, we conducted 15 cognitive interviews with the target group (13 nurses and two physicians) to ensure that the items are face valid and comprehensible. We then asked 32 experts on alarm fatigue to judge whether the items are suited for measuring alarm fatigue. The resulting 27 items were sent to nurses and physicians from 15 ICUs of a large German hospital. We used exploratory factor analysis to further reduce the number of items and to identify scales. A total of 585 submissions from 707 participants could be analysed (of which 14% were physicians and 64% were nurses). The simple structure of a two-factor model was achieved within three rounds. The final questionnaire (called Charité Alarm Fatigue Questionnaire; CAFQa) consists of nine items along two scales (i.e., the “alarm stress scale” and the “alarm coping scale”). The CAFQa is a brief questionnaire that allows clinical alarm researchers to quantify the alarm fatigue of nurses and physicians. It should not take more than five minutes to administer.
The present study investigates consumers’ acceptance of Se-biofortified apples, as well as Se health and nutrition claims that have been approved by the European Commission. Despite indications that such statements are more likely to be accepted if the carrier product itself has a healthy image, unprocessed fruits biofortified with Se have not been investigated in this context yet. Apples as the most frequently-consumed type of fresh fruit in Germany may offer the potential to improve the Se status of consumers when the produce is enriched with Se. Therefore, an online survey of 356 German adults was conducted to analyze different aspects that could affect the perception of and preference for the proposed product concept by consumers. The findings indicate a moderate acceptance of Se-biofortified apples, as well as of Se health and nutrition claims among the participants. Additional information about beneficial health effects of Se had a significant impact on consumer acceptance. People who regularly eat convenience food and prefer to buy apples at supermarkets were particularly attracted by the product idea. In conclusion, the results of the study indicate good prospects for a successful introduction of Se-rich apples in the German food market, if the produce is advertised with approved health and nutrition claims.
This study reported the impact of electron beam (e-beam) treatment on microbiota and mycotoxins naturally present in red pepper powder and physicochemical quality changes. Treatment at 6 kGy indicated significant (p < 0.05) decontamination of yeasts and molds by 3.0 and 4.4 log CFU/g, respectively. A reduction of 4.5 log CFU/g of total plate counts (TPC) was observed at 10 kGy for 23 s. Fungal inactivation followed first-order kinetics while TPC better fitted with Gompertz function (R2 = 0.9912). E-beam treatment was not efficient for the degradation of aflatoxins but indirectly controlled their production by inactivation of mycotoxigenic molds. Indeed, reduction of 25% ochratoxin A was recorded at 30 kGy retaining >85% of total phenols, carotenoids and antioxidants activity. Moreover, treatment impact on total color difference (ΔE*) indicated ‘slight differences’. Overall, e-beam treatments up to 10 kGy were efficient in decontaminating the natural microbiota without detrimental effects on the physicochemical qualities of red pepper powder.
The impact of pulsed light (PL) treatment on naturally occurring microorganisms, mycotoxins, and on physicochemical properties in red pepper powder was investigated. Powder samples were exposed to different PL treatments up to 61 pulses, with fluence ranging from 1.0 to 9.1 J/cm2. The highest fluence applied (9.1 J/cm2, 61 pulses, 20 s) resulted in 2.7, 3.1, and 4.1 log CFU/g reduction of yeasts, molds, and total plate counts (TPC), where initial microbial loads were 4.6, 5.5, and 6.5 log CFU/g, respectively. At the same fluence intensity, a maximum reduction of 67.2, 50.9, and 36.9% of aflatoxin B1 (AFB1), total aflatoxins (AF), and ochratoxin A (OTA) were detected, respectively. Proportional increase in temperature of the samples was observed from the absorbed PL energy, reaching maximum of 59.8°C. The inactivation of investigated microorganisms and mycotoxins followed first-order kinetics (R2 > 0.95). The fluence intensity at 6.9 and 9.1 J/cm2 did not cause degradation, but rather a significant (p < .05) and apparent increase of total phenols. Total color difference (ΔE*) revealed only “slight differences,” compared to the untreated sample. In conclusion, higher reduction of microbial load and mycotoxins in red pepper powder could be achieved, when higher treatment intensity was applied. This suggests the PL as a potential technology for decontamination of red pepper powder and other spice powders.
The study aimed to investigate inactivation of naturally occurring microorganisms and quality of red pepper paste treated by high pressure processing (HPP). Central composite rotatable design was employed to determine the impacts of pressure (100–600 MPa) and holding time (30–600 s). HPP at 527 MPa for 517 s reduced aerobic mesophilic bacteria count by 4.5 log CFU/g. Yeasts and molds counts were reduced to 1 log CFU/g at 600 MPa for 315 s. Total phenols, carotenoids and antioxidants activity ranged from 0.28 to 0.33 g GAE/100 g, 96.0–98.4 mg βc/100 g and 8.70–8.95 μmol TE/g, respectively. Increase (2.5–6.7%) in these variables was observed with increasing pressure and holding time. Total color difference (ΔE∗) values (0.2–2.8) were within the ranges of ‘imperceptible’ to ‘noticeable’. Experimental results were fitted satisfactorily into quadratic model with higher R2 values (0.8619–0.9863). Optimization process suggested treatment of red pepper paste at 536 MPa for 125 s for maximum desirability (0.622). Validation experiments confirmed comparable percentage of relative errors. Overall, this technique could be considered as an efficient treatment for the inactivation of microorganisms that naturally occur in red pepper paste with minimal changes in its characteristics.
Red pepper (Capsicum annuum L.) is one of the major spices consumed globally, recognized for its aroma and nutrient properties, and it has a major economic value for high producing countries. However, characterization of its techno-functional properties and in-depth understanding of oxidative stability is needed to produce food of high quality and stability. Thus, this work focused on the chemical, functional, thermal, oxidative stability and rheological properties of red pepper powder and paste. Experiment was designed in a Completely Randomized Design (CRD) fashion. The red pepper powder contained 14.50 g/100 g, 44.00 g/100 g and 7.57 g/100 g of crude fat, crude fiber and ash, respectively. The concentration of total phenols, carotenoids and antioxidants activity of the powder were 1.04 g GAE/100 g, 374 mg βc/100 g and 38.61 μmol TE/g, respectively. Functional properties showed lower bulk density (395.1 kg/m3) and higher tapped density (583.4 kg/m3) indicating the higher compressibility of the powder. In contrast, Hausner ratio (1.48), Carr’s index (32%) and angle of repose (45°) indicated poor flowability of the powder. Particle size distribution also indicated that the volume weighted mean values D[4,3] of the powder and paste were 262.20 and 201.46, respectively. Emulsifying capacity of the powder was 47.5%. Oil and water absorption capacities varied from 1.41 to 1.73 and 0.86 to 2.29 g/g of initial weight, respectively. Higher glass transition temperature was observed for the powder (62.54°C) than the paste (45.64°C). The induction period indicated that red pepper was more stable against oxidation in powder (5.2 h) than in the paste form (3.2 h). Rheological analysis revealed that the paste exhibited shear-thinning behavior. Overall, understanding of the properties of red pepper could contribute to enhance quality.
The study aimed for the analysis of the impact of pulsed electric field (PEF) pre-treatment on convection (CD) and microwave (MW-CD) assisted air drying. Drying kinetics acceleration and retention of bioactive compounds of PEF pre-treated carrots and apples has been demonstrated. Moreover, the direct and indirect environmental energy impacts of CD and MW-CD technologies with consideration of bioactive compounds preservation has been evaluated. PEF assisted CD and MW-CD demonstrated lower energy use, especially for indirect energy consumption, in the case of carotenoids preservation in dried carrots.
Health-related quality of life (HRQOL) in parents of children suffering from renal disease is often diminished by the illness burden experienced in daily life and by unfavorable ways of coping. Our aim was to examine the relationship between psychosocial strains perceived by parents, their ways of coping, and HRQOL. In an anonymous cross-sectional study, parents completed a questionnaire concerning psychosocial strains, coping strategies, and HRQOL, as well as sociodemographic and illness parameters. Study participants were recruited in two outpatient dialysis centers. Participating in the study were 195 parents (105 mothers, 90 fathers; age 43 ± 8 years; representing 108 families) of children suffering from renal disease (age 12 ± 5 years). Parents of children with chronic renal failure reported moderate HRQOL with parents of children undergoing dialysis experiencing more limitations in quality of life than parents of children living with a kidney graft and parents of children undergoing conservative treatment. Mothers experienced lower HRQOL and higher psychosocial strains than fathers. HRQOL was predicted by the coping strategies “focusing on child” (β = –0.25), “improving marital relationship” (β = 0.24), “seeking social support” (β = –0.22) and “self-acceptation and growth” (β =0 .19) as well as parents′ perceived limitation by illness in daily life (β = –0.15; explained variance 57%). In the comprehensive care for families with a child suffering from a renal disease, screening for psychosocial strains and ways of coping, along with applying interventions to strengthen adaptive coping strategies, may be a preventative means of improving parents′ quality of life.
Purpose
In a knowledge-based society, one of the main driving forces of prosperity at city and regional levels is the ability to attract students and recently qualified graduates. The purpose of this paper is to identify the urban factors that contribute to “attractiveness” from a student perspective and to subsequently analyse the influence of these factors on students’ location choices. In the corresponding case study, the criteria identified were used to evaluate student satisfaction with the German city of Osnabrück.
Design/methodology/approach
A multi-level empirical research concept was used to analyse the relevant criteria. Focus group discussions served as the basis to design and conduct a survey of almost 2,300 students in the university city of Osnabrück at the end of 2016. A follow-up study in November 2017 at two universities in Berlin provided data for comparison.
Findings
Accessible – safe – clean and with affordable housing: these terms sum up the core expectations that students have of an attractive city. In contrast to the findings of Richard Florida regarding the creative class, cultural opportunities, exciting nightlife and a multicultural population were not important factors among the survey participants.
Research limitations/implications
As the study was conducted in two cities only, the findings cannot be generalised. It should be replicated in at least five other cities to validate and compare the research findings presented here. Furthermore, this study focused on the factors influencing the attractiveness of the city as a whole. However, the attractiveness of a specific neighbourhood may be of greater relevance to the decision-making process. As housing plays a major role, students’ particular needs with regard to accommodation should also be examined at greater depth.
Practical implications
Insights generated by the empirical study provide relevant information that may assist city stakeholders in taking effective measures regarding place management to attract and retain students.
Originality/value
To date, little research has been undertaken to empirically examine the specific factors that German students look for when deciding where to study and live. The goal of this paper is to present new empirical insights concerning the quality-of-life factors that influence students’ decision-making processes.
The break-up of the Soviet Union in 1991 triggered cropland abandonment on a continental scale, which in turn ledto carbon accumulation on abandoned land across Eurasia. Previous studies have estimated carbon accumulationrates across Russia based on large-scale modelling. Studies that assess carbon sequestration on abandoned land basedon robust field sampling are rare. We investigated soil organic carbon (SOC) stocks using a randomized samplingdesign along a climatic gradient from forest steppe to Sub-Taiga in Western Siberia (Tyumen Province). In total, SOCcontents were sampled on 470 plots across different soil and land-use types. The effect of land use on changes in SOCstock was evaluated, and carbon sequestration rates were calculated for different age stages of abandoned cropland.While land-use type had an effect on carbon accumulation in the topsoil (0–5 cm), no independent land-use effectswere found for deeper SOC stocks. Topsoil carbon stocks of grasslands and forests were significantly higher thanthose of soils managed for crops and under abandoned cropland. SOC increased significantly with time sinceabandonment. The average carbon sequestration rate for soils of abandoned cropland was 0.66 Mg C ha1yr1(1–20 years old, 0–5 cm soil depth), which is at the lower end of published estimates for Russia and Siberia. Therewas a tendency towards SOC saturation on abandoned land as sequestration rates were much higher for recentlyabandoned (1–10 years old, 1.04 Mg C ha1yr1) compared to earlier abandoned crop fields (11–20 years old,0.26 Mg C ha1yr1). Our study confirms the global significance of abandoned cropland in Russia for carbonsequestration. Our findings also suggest that robust regional surveys based on a large number of samples advancemodel-based continent-wide SOC prediction.
While recent studies have demonstrated that events are fundamentally climate sensitive, this seems to not be fully considered in event research or corporate event practice. Thus, this study aims to identify the influencing factors that affect the acceptance of climate adaptation measures among decision-makers in the event industry. The analysis was divided into three main parts. First, the existing literature related to climate change in an events context was reviewed. Using 15 semi structured interviews, the findings from this review were then critically discussed with stakeholders in Germany involved in event planning. Finally, explicit climate adaptation measures were proposed and discussed. Based on all findings, there appears to be a low level of awareness of and interest in climate adaptation amongst German event industry players. There is an imminent need for further research on climate adaptation and for decision-makers to better prepare for climate change in order to counteract resulting negative impacts.
Iron deficiency is still widespread as a major health problem even in countries with adequate food supply. It mainly affects women but also vegans, vegetarians, and athletes and can lead to various clinical pictures. Biofortification of vitamin C-rich vegetables with iron may be one new approach to face this nutritional challenge. However, so far, little is known about the consumer acceptance of iron-biofortified vegetables, particularly in developed countries. To address this issue, a quantitative survey of 1000 consumers in Germany was conducted. The results showed that depending on the type of vegetable, between 54% and 79% of the respondents were interested in iron-biofortified vegetables. Regression analysis showed a relationship between product acceptance, gender, and area of residence. In addition, relationships were found between consumer preferences for enjoyment, sustainability, and naturalness. Compared to functional food and dietary supplements, 77% of respondents would prefer fresh iron-rich vegetables to improve their iron intake. For a market launch, those iron-rich vegetables appear especially promising, which can additionally be advertised with claims for being rich in vitamin C and cultivated in an environmentally friendly way. Consumers were willing to pay EUR 0.10 to EUR 0.20 more for the iron-biofortified vegetables.
Consumer Acceptance and Market Potential of Iodine-Biofortified Fruit and Vegetables in Germany
(2021)
Biofortification of food crops with iodine is a novel approach to preventing iodine deficiency in humans. The present study analyses the consumer target groups and the market potential of iodine-biofortified fruit and vegetables in Germany. For this purpose, an online survey of 1016 German fruit and vegetable consumers was conducted to investigate the acceptance of different product categories as well as relevant criteria for the market launch. The results show that io-dine-biofortified fruit and vegetables are particularly attractive to consumers who purchase at farmers’ markets, organic food shops, and farm stores. Out of this group, 39% of consumers rate such iodine-rich foods as very appealing. They attach importance to food that naturally contains iodine and prefer produce from integrated domestic cultivation. With their focus on sustainability and naturalness, this group of consumers clearly differs from typical users of dietary supplements, who are primarily concerned with health benefits. However, overall about 85% of respondents would prefer biofortified fruits and vegetables to supplements to improve their iodine supply. The greatest market potential for iodine-biofortified fruit and vegetables is to be expected in super-markets, as this is the preferred food shopping location for most consumers. A total of 28% of those who buy here rate the biofortified foods presented as very appealing. Nevertheless, a successful market launch requires that the benefits of the new products are communicated according to the potential consumer group needs.
Background:
Contact tracing apps are potentially useful tools for supporting national COVID-19 containment strategies. Various national apps with different technical design features have been commissioned and issued by governments worldwide.
Objective:
Our goal was to develop and propose an item set that was suitable for describing and monitoring nationally issued COVID-19 contact tracing apps. This item set could provide a framework for describing the key technical features of such apps and monitoring their use based on widely available information.
Methods:
We used an open-source intelligence approach (OSINT) to access a multitude of publicly available sources and collect data and information regarding the development and use of contact tracing apps in different countries over several months (from June 2020 to January 2021). The collected documents were then iteratively analyzed via content analysis methods. During this process, an initial set of subject areas were refined into categories for evaluation (ie, coherent topics), which were then examined for individual features. These features were paraphrased as items in the form of questions and applied to information materials from a sample of countries (ie, Brazil, China, Finland, France, Germany, Italy, Singapore, South Korea, Spain, and the United Kingdom [England and Wales]). This sample was purposefully selected; our intention was to include the apps of different countries from around the world and to propose a valid item set that can be relatively easily applied by using an OSINT approach.
Results:
Our OSINT approach and subsequent analysis of the collected documents resulted in the definition of the following five main categories and associated subcategories: (1) background information (open-source code, public information, and collaborators); (2) purpose and workflow (secondary data use and warning process design); (3) technical information (protocol, tracing technology, exposure notification system, and interoperability); (4) privacy protection (the entity of trust and anonymity); and (5) availability and use (release date and the number of downloads). Based on this structure, a set of items that constituted the evaluation framework were specified. The application of these items to the 10 selected countries revealed differences, especially with regard to the centralization of the entity of trust and the overall transparency of the apps’ technical makeup.
Conclusions:
We provide a set of criteria for monitoring and evaluating COVID-19 tracing apps that can be easily applied to publicly issued information. The application of these criteria might help governments to identify design features that promote the successful, widespread adoption of COVID-19 tracing apps among target populations and across national boundaries.
Background:
Etomidate is typically used as an induction agent in cardiac surgery because it has little impact on hemodynamics. It is a known suppressor of adrenocortical function and may increase the risk for post-operative infections, sepsis, and mortality. The aim of this study was to evaluate whether etomidate increases the risk of postoperative sepsis (primary outcome) and infections (secondary outcome) compared to propofol.
Methods:
This was a retrospective before–after trial (IRB EA1/143/20) performed at a tertiary medical center in Berlin, Germany, between 10/2012 and 01/2015. Patients undergoing cardiac surgery were investigated within two observation intervals, during which etomidate and propofol were the sole induction agents.
Results:
One-thousand, four-hundred, and sixty-two patients, and 622 matched pairs, after caliper propensity-score matching, were included in the final analysis. Sepsis rates did not differ in the matched cohort (etomidate: 11.5% vs. propofol: 8.2%, p = 0.052). Patients in the etomidate interval were more likely to develop hospital-acquired pneumonia (etomidate: 18.6% vs. propofol: 14.0%, p = 0.031).
Conclusion:
Our study showed that a single-dose of etomidate is not statistically associated with higher postoperative sepsis rates after cardiac surgery, but is associated with a higher incidence of hospital-acquired pneumonia. However, there is a notable trend towards a higher sepsis rate.
The benzodiazepine, midazolam, is one of the most frequently used sedatives in intensive care medicine, but it has an unfavorable pharmacokinetic profile when continuously applied. As a consequence, patients are frequently prolonged and more deeply sedated than intended. Due to its distinct pharmacological features, including a cytochrome P450-independent metabolization, intravenous lormetazepam might be clinically advantageous compared to midazolam. In this retrospective cohort study, we compared patients who received either intravenous lormetazepam or midazolam with respect to their survival and sedation characteristics. The cohort included 3314 mechanically ventilated, critically ill patients that received one of the two drugs in a tertiary medical center in Germany between 2006 and 2018. A Cox proportional hazards model with mortality as outcome and APACHE II, age, gender, and admission mode as covariates revealed a hazard ratio of 1.75 [95% CI 1.46–2.09; p < 0.001] for in-hospital mortality associated with the use of midazolam. After additionally adjusting for sedation intensity, the HR became 1.04 [95% CI 0.83–1.31; p = 0.97]. Thus, we concluded that excessive sedation occurs more frequently in critically ill patients treated with midazolam than in patients treated with lormetazepam. These findings require further investigation in prospective trials to assess if lormetazepam, due to its ability to maintain light sedation, might be favorable over other benzodiazepines for sedation in the ICU.
Acute post-operative delirium (POD) and long-term post-operative cognitive dysfunction (POCD) are frequent and associated with increased mortality, dependency on care giving and institutionalization rates. The POCD-related cost burden on the German long-term care insurance provides an indication for the savings potential from risk-adapted treatment schemes. Comprehensive estimates have not been assessed or published so far.
A model-based cost-analysis was designed to estimate POCD-related costs in the long-term care insurance. Comprehensive analysis of inpatient operations and procedures (OPS-codes) served as the base for case number calculations, which were then used as input to the actual cost model. POCD-incidence rates were obtained from the BioCog study. Various sensitivity analyses were performed to assess uncertainty of the model results.
Total POCD related annual costs in the German long-term care insurance account for approximately 1.6 billion EUR according to the base case of our analysis. Total annual costs for all POCD cases depend on surgery numbers, incidence rates, other assumptions, and uncertain input parameters.
The financial burden to the long-term care insurance is substantial, even in a conservative scenario of the cost model. Variability of results stems from uncertain assumptions, POCD-incidence rates and from uncertain patient numbers who are undergoing surgery and are therefore at risk to develop POCD.
After foundation of the Wadden Sea National Park, grazing and artificial drainage was ceased or reduced on large areas of the salt marshes at the Schleswig-Holstein mainland coast (Northern Germany). The effect of grazing cessation versus intensive and moderate grazing on vegetation diversity was studied on small (plant species richness on plots between 0.01 and 100 m2) and large scale (vegetation type richness per hectare) over 18 to 20 years by analysing data from long-term monitoring programs. Plant species richness and vegetation type richness increased strongly over time in all management regimes, because grazing-sensitive species increased first in ungrazed marshes and later dispersed to and established in intensively grazed marshes. Dominance of the tall, late-successional grass Elymus athericus on 7% to 52% of all moderately and ungrazed (primarily high marsh) plots led to a decrease in species richness. After 18 to 20 years, species richness was highest in moderately and intensively grazed high marshes. Differences were significant only on small plots of up to 4 m2. On the large scale, vegetation type richness in the low marsh was higher without grazing, while no differences were found in the high marsh. Our results indicate that grazing effects differ between spatial scales and that different spatial scales have to be considered for monitoring and evaluation of vegetation diversity in salt marshes. To conserve vegetation diversity on all scales, a large-scale mosaic of different management regimes should be maintained.
BACKGROUND:
There is little knowledge regarding the association between psychological factors and complex regional pain syndrome (CRPS) in children. Specifically, it is not known which factors precipitate CRPS and which result from the ongoing painful disease.
OBJECTIVES:
To examine symptoms of depression and anxiety as well as the experience of stressful life events in children with CRPS compared with children with chronic primary headaches and functional abdominal pain.METHODS: A retrospective chart study examined children with CRPS (n=37) who received intensive inpatient pain treatment between 2004 and 2010. They were compared with two control groups (chronic primary headaches and functional abdominal pain; each n=37), who also received intensive inpatient pain treatment. Control groups were matched with the CRPS group with regard to admission date, age and sex. Groups were compared on symptoms of depression and anxiety as well as stressful life events.
RESULTS:
Children with CRPS reported lower anxiety and depression scores compared with children with abdominal pain. A higher number of stressful life events before and after the onset of the pain condition was observed for children with CRPS.
CONCLUSIONS:
Children with CRPS are not particularly prone to symptoms of anxiety or depression. Importantly, children with CRPS experienced more stressful life events than children with chronic headaches or abdominal pain. Prospective long-term studies are needed to further explore the potential role of stressful life events in the etiology of CRPS.
Chitin is an abundant waste product from shrimp and mushroom industries and as such, an appropriate secondary feedstock for biotechnological processes. However, chitin is a crystalline substrate embedded in complex biological matrices, and, therefore, difficult to utilize, requiring an equally complex chitinolytic machinery. Following a bottom-up approach, we here describe the step-wise development of a mutualistic, non-competitive consortium in which a lysine-auxotrophic Escherichia coli substrate converter cleaves the chitin monomer N-acetylglucosamine (GlcNAc) into glucosamine (GlcN) and acetate, but uses only acetate while leaving GlcN for growth of the lysine-secreting Corynebacterium glutamicum producer strain. We first engineered the substrate converter strain for growth on acetate but not GlcN, and the producer strain for growth on GlcN but not acetate. Growth of the two strains in co-culture in the presence of a mixture of GlcN and acetate was stabilized through lysine cross-feeding. Addition of recombinant chitinase to cleave chitin into GlcNAc2, chitin deacetylase to convert GlcNAc2 into GlcN2 and acetate, and glucosaminidase to cleave GlcN2 into GlcN supported growth of the two strains in co-culture in the presence of colloidal chitin as sole carbon source. Substrate converter strains secreting a chitinase or a β-1,4-glucosaminidase degraded chitin to GlcNAc2 or GlcN2 to GlcN, respectively, but required glucose for growth. In contrast, by cleaving GlcNAc into GlcN and acetate, a chitin deacetylase-expressing substrate converter enabled growth of the producer strain in co-culture with GlcNAc as sole carbon source, providing proof-of-principle for a fully integrated co-culture for the biotechnological utilization of chitin.
Alexithymia, or a lack of emotional awareness, is prevalent in some chronic pain conditions and has been linked to poor recognition of others' emotions. Recognising others' emotions from their facial expression involves both emotional and motor processing, but the possible contribution of motor disruption has not been considered. It is possible that poor performance on emotional recognition tasks could reflect problems with emotional processing, motor processing or both. We hypothesised that people with chronic facial pain would be less accurate in recognising others' emotions from facial expressions, would be less accurate in a motor imagery task involving the face, and that performance on both tasks would be positively related. A convenience sample of 19 people (15 females) with chronic facial pain and 19 gender-matched controls participated. They undertook two tasks; in the first task, they identified the facial emotion presented in a photograph. In the second, they identified whether the person in the image had a facial feature pointed towards their left or right side, a well-recognised paradigm to induce implicit motor imagery. People with chronic facial pain performed worse than controls at both tasks (Facially Expressed Emotion Labelling (FEEL) task P < 0·001; left/right judgment task P < 0·001). Participants who were more accurate at one task were also more accurate at the other, regardless of group (P < 0·001, r2 = 0·523). Participants with chronic facial pain were worse than controls at both the FEEL emotion recognition task and the left/right facial expression task and performance covaried within participants. We propose that disrupted motor processing may underpin or at least contribute to the difficulty that facial pain patients have in emotion recognition and that further research that tests this proposal is warranted.
The primary objective of this study was to determine the structural and known-group validity as well as the inter-rater reliability of a test battery to evaluate the motor control of the craniofacial region. Seventy volunteers without TMD and 25 subjects with TMD (Axes I) per the DC/TMD were asked to execute a test battery consisting of eight tests. The tests were video-taped in the same sequence in a standardised manner. Two experienced physical therapists participated in this study as blinded assessors. We used exploratory factor analysis to identify the underlying component structure of the eight tests. Internal consistency (Cronbach's α), inter-rater reliability (intra-class correlation coefficient) and construct validity (ie, hypothesis testing-known-group validity) (receiver operating curves) were also explored for the test battery. The structural validity showed the presence of one factor underlying the construct of the test battery. The internal consistency was excellent (0.90) as well as the inter-rater reliability. All values of reliability were close to 0.9 or above indicating very high inter-rater reliability. The area under the curve (AUC) was 0.93 for rater 1 and 0.94 for rater two, respectively, indicating excellent discrimination between subjects with TMD and healthy controls. The results of the present study support the psychometric properties of test battery to measure motor control of the craniofacial region when evaluated through videotaping. This test battery could be used to differentiate between healthy subjects and subjects with musculoskeletal impairments in the cervical and oro-facial regions. In addition, this test battery could be used to assess the effectiveness of management strategies in the craniofacial region.
Objective
To identify assessment tools used to evaluate patients with temporomandibular disorders (TMD) considered to be clinically most useful by a panel of international experts in TMD physical therapy (PT).
Methods
A Delphi survey method administered to a panel of international experts in TMD PT was conducted over three rounds from October 2017 to June 2018. The initial contact was made by email. Participation was voluntary. An e-survey, according to the Checklist for Reporting Results of Internet E-Surveys (CHERRIES), was posted using SurveyMonkey for each round. Percentages of responses were analysed for each question from each round of the Delphi survey administrations.
Results
Twenty-three experts (completion rate: 23/25) completed all three rounds of the survey for three clinical test categories: 1) questionnaires, 2) pain screening tools and 3) physical examination tests. The following was the consensus-based decision regarding the identification of the clinically most useful assessments. (1) Four of 9 questionnaires were identified: Jaw Functional Limitation (JFL-8), Mandibular Function Impairment Questionnaire (MFIQ), Tampa Scale for Kinesiophobia for Temporomandibular disorders (TSK/TMD) and the neck disability index (NDI). (2) Three of 8 identified pain screening tests: visual analog scale (VAS), numeric pain rating scale (NRS) and pain during mandibular movements. (3) Eight of 18 identified physical examination tests: physiological temporomandibular joint (TMJ) movements, trigger point (TrP) palpation of the masticatory muscles, TrP palpation away from the masticatory system, accessory movements, articular palpation, noise detection during movement, manual screening of the cervical spine and the Neck Flexor Muscle Endurance Test.
Conclusion
After three rounds in this Delphi survey, the results of the most used assessment tools by TMD PT experts were established. They proved to be founded on test construct, test psychometric properties (reliability/validity) and expert preference for test clusters. A concordance with the screening tools of the diagnostic criteria of TMD consortium was noted. Findings may be used to guide policymaking purposes and future diagnostic research.
The management of patients experiencing chronic orofacial pain is a great challenge, due to the complexity of chronic pain itself, combined with an increased peripheral sensitization in the craniofacial itself. Therefore, patients with orofacial pain may present a clear distortion of the somatorepresentation after some time. In this review, the authors develop a neurophysiological explanation of orofacial distortion, as well as propose assessment and treatment options, based on scarcely available scientific evidence and their own clinical experience. The assessments of facial somatosensory, cognitive-affective and motor dysfunctions are crucial to establish the most accurate treatment; the assessment tools are described in the article. Two-point discrimination, laterality recognition and emotion recognition are altered in patients with orofacial pain. Other sensorimotor assessment tools, such as motor acuity and auditory acuity, are also explained. Finally, the authors review their treatment proposals, based on the integration of brain training techniques and biobehavioral interventions. Somatosensory reintegration (tactile acuity training), facial emotion recognition, movement representation techniques, orofacial motor training and therapeutic patient education are explained in detail, and this may challenge new directions in rehabilitation and research.
Recognition of Emotional Facial Expressions and Alexithymia in Patients with Chronic Facial Pain
(2018)
Objectives
Alexithymia, conceived as difficulties to identify emotions, is said to be related with several pain syndromes. This study examined the recognition of facially expressed emotions and its relation to alexithymia in subjects with chronic facial pain.
Methods
A total of 62 subjects were recruited, with n=20 patients with chronic facial pain and n=42 healthy controls. All subjects were tested for the recognition of facially expressed emotions (Facially Expressed Emotion Labelling Test (FEEL test). The Toronto Alexithymia Scale (TAS-26) was used for the diagnosis of alexithymia.
Results
Patients with chronic facial pain performed worse than controls at the FEEL task (p<.001) and showed higher total TAS scores (p<.001). This indicates the presence of alexithymia and facial emotion recognition deficits in the facial pain group.
Discussion
It was concluded from the results that both the recognition of facially expressed emotions, and the ability to identify and describe one’s own feelings (TAS), are restricted in chronic orofacial pain patients. This relationship is particularly important in the treatment of chronic facial pain, indicating that it should become part of the treatment in addition to the therapeutic key issues, to influence the quality of life of the affected patients positively.
Chronic facial pain has many of the clinical characteristics found in other persistent musculoskeletal conditions, such as low back and cervical pain syndromes. Unique to this condition, however, is that painful facial movements may result in rigidity or altered ability to demonstrate mimicry, defined as the natural tendency to adopt the behavioral expressions of other persons involved in the interaction. Loss of ability to communicate through emotional expression can lead to impaired processing of emotions and ultimately social isolation. Diminished quality and quantity of facial expression is associated with chronic face pain, tempromandibular dysfunction, facial asymmetries, and neurological disorders. This report provides a framework for assessment of impaired emotional processing and associated somatosensory alterations. Principles for management for chronic facial pain should include graded motor imagery, in addition to standard treatments of manual therapy, exercise, and patient education. A case study is provided which illustrates these principles.
Introduction
Tests to evaluate the integrity of the alar ligaments are important clinical tools for manual therapists, but there is limited research regarding their validity.
Method
A single blinded examiner assessed alar ligament integrity using the lateral shear test (LST), rotation stress test (RST) and side-bending stress test (SBST) on a sample of convenience comprising 7 subjects with MRI confirmed alar ligament lesions and 11 healthy people. Alar ligament lesions were identified using both supine and high-field strength upright MRI.
Results
The RST had a sensitivity of 80% and a specificity of 69.2%. The SBST and the LST both showed a sensitivity of 80% and a specificity of 76.9%. In cases where all three tests were positive, the specificity increased to 84.6%.
Discussion
Tests of manual examination of alar ligament integrity have some diagnostic utility; however, these findings require further corroboration in a larger sample.
BACKGROUND: The Craniofacial Pain and Disability Inventory (CF-PDI) is a cross-culturally adapted instrument designed from a biopsychosocial perspective to measure pain, disability, and function in orofacial head and neck pain with shown psychometric properties; however, the German cross-cultural adaption is lacking.
OBJECTIVES: To carry out a transcultural translation of CF-PDI into German and assess its psychometric properties in patients with painful temporomandibular disorders (TMD) with respect to construct and clinical validity, internal consistency and reproducibility.
STUDY DESIGN: Multicenter, prospective, cross-sectional design.
SETTING: Patients (n = 398) were recruited from dental and physical therapy clinics in middle and south Germany.
METHODS: Structural validity was assessed using exploratory factor analysis (EFA) and confirmatory factor analysis (CFA). We investigated know-group validity by means of the scale’s potential to discriminate between affected and unaffected subjects. Multiple linear regression analysis was used to estimate convergent validity. We tested test-retest reliability by the intraclass correlation coefficient and the Internal consistency by Cronbach’s alpha, or each dimension separately, and the total score. Multiple linear regression analysis was used to estimate convergent validity.
RESULTS: Two hundred forty-six heterogeneous chronic craniofacial pain patients and 152 patients without complaints were recruited from the middle and south of Germany. The German version CF-PDI-G presents 21 items, 4 factors, and adequate psychometric properties. The test-retest reliability and internal consistency of the CF-PDI-G were both excellent for the entire instrument and also for all sub-scales (intraclass correlation coefficient [ICC] > 0.90) except for the comorbidities and interference with work which was acceptable (ICC = 0.69). Standard error of the measurement (SEM) and minimal detectable change values are sufficiently low. Assessment of clinical validity shows good potential of discrimination and classification into categories “no,” “mild,” “moderate,” and “severe.” The multiple linear regression model showed a strong association between neck disability index, Visual Analog Scale, and anamnestic questionnaire (supporting the scale’s convergent validity).
LIMITATIONS: Our sample has a higher prevalence of women and the sample was not recruited consecutively, which may lead to a biased estimation of psychometric properties.
CONCLUSIONS: The CF-PDI-G represents valid and reliable instrument to assess pain and disability in patients with orofacial pain and headache suitable for research and clinical practice.
There is evidence that temporomandibular disorder (TMD) may be a contributing factor to cervicogenic headache (CGH), in part because of the influence of dysfunction of the temporomandibular joint on the cervical spine. The purpose of this randomized controlled trial was to determine whether orofacial treatment in addition to cervical manual therapy, was more effective than cervical manual therapy alone on measures of cervical movement impairment in patients with features of CGH and signs of TMD. In this study, 43 patients (27 women) with headache for more than 3-months and with some features of CGH and signs of TMD were randomly assigned to receive either cervical manual therapy (usual care) or orofacial manual therapy to address TMD in addition to usual care. Subjects were assessed at baseline, after 6 treatment sessions (3-months), and at 6-months follow-up. 38 subjects (25 female) completed all analysis at 6-months follow-up. The outcome criteria were: cervical range of movement (including the C1-2 flexion-rotation test) and manual examination of the upper 3 cervical vertebra. The group that received orofacial treatment in addition to usual care showed significant reduction in all aspects of cervical impairment after the treatment period. These improvements persisted to the 6-month follow-up, but were not observed in the usual care group at any point. These observations together with previous reports indicate that manual therapists should look for features of TMD when examining patients with headache, particularly if treatment fails when directed to the cervical spine.
Objective:
The German version of the Social Phobia and Anxiety Inventory (SPAI-G) is avalidated measure for the detection of social anxiety disorder (SAD). The aim of the presentstudy was to develop optimal cut points (OC) for remission and response to treatment for theSPAI-G.
Methods:
We used Receiver Operating Characteristic methods and bootstrapping to analysethe data of 359 patients after psychotherapeutic treatment. OCs where defined as the cut pointswith the highest sensitivity and specificity after bootstrapping.
Results:
For remission, an OC of 2.79 was found, and for response, a change in score frompre- to posttreatment by 11% yielded best results.
Conclusions:
The OC we identified for remission may be used to improve the diagnostic utilityof the SPAI-G. However, the cut point for response achieved only borderline-acceptable levelsof sensitivity and specificity, calling into doubt their utility in clinical and research setting.
The Liebowitz Social Anxiety Scale (LSAS) is the most frequently used instrument to assess social anxiety disorder (SAD) in clinical research and practice. Both a self‐reported (LSAS‐SR) and a clinician‐administered (LSAS‐CA) version are available. The aim of the present study was to define optimal cut‐off (OC) scores for remission and response to treatment for the LSAS in a German sample.
Data of N = 311 patients with SAD were used who had completed psychotherapeutic treatment within a multicentre randomized controlled trial. Diagnosis of SAD and reduction in symptom severity according to the Structured Clinical Interview for Diagnostic and Statistical Manual of Mental Disorders, 4th edition, served as gold standard. OCs yielding the best balance between sensitivity and specificity were determined using receiver operating characteristics. The variability of the resulting OCs was estimated by nonparametric bootstrapping.
Using diagnosis of SAD (present vs. absent) as a criterion, results for remission indicated cut‐off values of 35 for the LSAS‐SR and 30 for the LSAS‐CA, with acceptable sensitivity (LSAS‐SR: .83, LSAS‐CA: .88) and specificity (LSAS‐SR: .82, LSAS‐CA: .87). For detection of response to treatment, assessed by a 1‐point reduction in the Structured Clinical Interview for Diagnostic and Statistical Manual of Mental Disorders, 4th edition, rating, a reduction of 28% for the LSAS‐SR and 29% for the LSAS‐CA yielded the best balance between sensitivity (LSAS‐SR: .75, LSAS‐CA: .83) and specificity (LSAS‐SR: .76, LSAS‐CA: .80).
To our knowledge, we are the first to define cut points for the LSAS in a German sample. Overall, the cut points for remission and response corroborate previously reported cut points, now building on a broader data basis.
Suicide is a major cause of death in adulthood and specifically in patients suffering from mental illnesses. The Depressive Symptom Inventory Suicidality Subscale (DSI-SS) is widely used to detect and prevent suicidal ideation. The aim of the present study was to determine optimal cut points for the DSI-SS in different populations.
We analysed the data of one population-based sample (n = 532), one outpatient sample (n = 180) and one inpatient sample (n = 244). Internal consistency, convergent validity and optimal cut points according to receiver operating characteristics were calculated.
In all samples, we found excellent item-total correlations and internal consistencies for the DSI-SS. Zero-order correlations between the DSI-SS and theoretically related constructs showed positive correlation coefficients, ranging from 0.50 to 0.67. The DSI-SS differentiated well between patients with and without suicide attempts in the population-based sample, but less so in the inpatient sample and only marginally in the outpatient sample. A bootstrapping analysis showed some variability in the cut points that emerged as optimal, but there was no overlap between the different samples.
The specific cut points that we identified may be used to improve the diagnostic utility of the DSI-SS and the chance to detect suicidal ideation.
Background:
One of the main problems of Internet-delivered interventions for a range of disorders is the high dropout rate, yet little is known about the factors associated with this. We recently developed and tested a Web-based 6-session program to enhance motivation to change for women with anorexia nervosa, bulimia nervosa, or related subthreshold eating pathology.
Objective:
The aim of the present study was to identify predictors of dropout from this Web program.
Methods:
A total of 179 women took part in the study. We used survival analyses (Cox regression) to investigate the predictive effect of eating disorder pathology (assessed by the Eating Disorders Examination-Questionnaire; EDE-Q), depressive mood (Hopkins Symptom Checklist), motivation to change (University of Rhode Island Change Assessment Scale; URICA), and participants’ age at dropout. To identify predictors, we used the least absolute shrinkage and selection operator (LASSO) method.
Results:
The dropout rate was 50.8% (91/179) and was equally distributed across the 6 treatment sessions. The LASSO analysis revealed that higher scores on the Shape Concerns subscale of the EDE-Q, a higher frequency of binge eating episodes and vomiting, as well as higher depression scores significantly increased the probability of dropout. However, we did not find any effect of the URICA or age on dropout.
Conclusions:
Women with more severe eating disorder pathology and depressive mood had a higher likelihood of dropping out from a Web-based motivational enhancement program. Interventions such as ours need to address the specific needs of women with more severe eating disorder pathology and depressive mood and offer them additional support to prevent them from prematurely discontinuing treatment.
The Brief Symptom Inventory (BSI)-18 is a widely-used tool to assess changes in general distress in patients despite an ongoing debate about its factorial structure and lack of evidence for longitudinal measurement invariance (LMI). We investigated BSI-18 scores from 1,081 patients from an outpatient clinic collected after the 2nd, 6th, 10th, 18th, and 26th therapy session. Confirmatory factor analysis (CFA) was used to compare models comprising one, three, and four latent dimensions that were proposed in the literature. LMI was investigated using a series of model comparisons, based on chi-square tests, effect sizes, and changes in comparative fit index (CFI). Psychological distress diminished over the course of therapy. A four-factor structure (depression, somatic symptoms, generalized anxiety, and panic) showed the best fit to the data at all measurement occasions. The series of model comparisons showed that constraining parameters to be equal across time resulted in very small decreases in model fit that did not exceed the cutoff for the assumption of measurement in variance. Our results show that the BSI-18 is best conceptualized as a four-dimensional tool that exhibits strict longitudinal measurement invariance. Clinicians and applied researchers do not have to be concerned about the interpretation of mean differences over time.
Investigation of the use of ceramic materials in innovative light water reactor – fuel rod concepts
(2001)
Purpose:
To analyse the willingness for postmortem cornea donation in Germany.
Methods:
Employees in two cities (UKM, UKS), and university hospitals (STM, STE), members of the German Ophthalmological Society (DOG), and employees of an automobile company (BO) participated in a questionnaire about postmortem cornea donation attitudes. The questionnaire consisted of demographic items, motives concerning postmortem cornea donation, general attitudes toward donation, and questions concerning the perceived needs for information about donation. The statistical analyses included logistic regression with the target parameter of 'willingness to donate cornea postmortem'.
Results:
Of the participants, 67.7 % (UKM, UKS), 70.9 % (STM, STE), 70.8 % (BO), and 79.4 % (DOG) declared their intention to donate their corneas postmortem. Younger age (p < 0.001), poorer general health (p < 0.05), faith in an eternal life (p < 0.05), disagreement with brain death diagnostics (p < 0.001), fear of receiving worse medical treatment (p < 0.001), and fear of the commercialization of organs (p < 0.001) were found to be risk factors for a negative attitude toward postmortem cornea. The majority of participants (57.4 %) indicated that additional information about donation would be appreciated, and the internet (69.9 %) was considered the most appropriate means for conveying this information.
Conclusions:
Emotional items were revealed to be the most relevant factors influencing the willingness to donate cornea postmortem, which may be counteracted by means of public education. The relatively low willingness among the medical staff contrasts with previous observations in a professional ophthalmologic society.
Attitudes Concerning Postmortem Organ Donation : A Multicenter Survey in Various German Cohorts
(2015)
BACKGROUND
The aim of this study was to characterize postmortem organ donation attitudes in various German cohorts.
MATERIAL AND METHODS
Employees of 2 German cities and 2 German university hospitals, employees of a German automobile enterprise, and members of a German Medical Society were administered a questionnaire about postmortem organ and tissue donation attitudes. Demographic data and general attitudes were questioned and focused on: I) willingness to donate organs, II) holding a donor card, and III) having discussed the topic with the family.
RESULTS
Of 5291 participants, 65.2% reported favoring postmortem organ donation. Missing negative experiences, the idea that donation is helpful, a non-medical professional environment, excellent general health, gender, agreement with the brain-death paradigm, and age significantly influenced the participants’ attitudes. Participants were more likely to possess donor cards and had discussed more often with family members if they agreed with the brain-death paradigm and considered donation to be helpful. Males and older participants were the most likely to neglect donor cards, and Catholics, Protestants, and participants with poor health were the least likely to donate organs. Interest in receiving more information was expressed by 38.1% and 50.6% of participants refusing donation of all or of specific organs, respectively, and suggested the internet (60.0%) and family doctors (35.0%) as preferred sources of information.
CONCLUSIONS
Public campaigns in Germany should focus on males and older people as regards donor cards, and females, younger, and religiously affiliated persons as regards the general willingness to donate organs postmortem.
Introduction: Patients undergoing revision total hip surgery (RTHS) have a high prevalence of mild and moderate preoperative anemia, associated with adverse outcomes. The aim of this study was to investigate the association of perioperative allogeneic blood transfusions (ABT) and postoperative complications in preoperatively mild compared to moderate anemic patients undergoing RTHS who did not receive a diagnostic anemia workup and treatment before surgery. Methods: We included 1,765 patients between 2007 and 2019 at a university hospital. Patients were categorized according to their severity of anemia using the WHO criteria of mild, moderate, and severe anemia in the first Hb level of the case. Patients were grouped as having received no ABT, 1–2 units of ABT, or more than 2 units of ABT. Need for intraoperative ABT was assessed in accordance with institutional standards. Primary endpoint was the compound incidence of postoperative complications. Secondary outcomes included major/minor complications and length of hospital and ICU stay. Results: Of the 1,765 patients, 31.0% were anemic of any cause before surgery. Transfusion rates were 81% in anemic patients and 41.2% in nonanemic patients. The adjusted risks for compound postoperative complication were significantly higher in patients with moderate anemia (OR 4.88, 95% CI: 1.54–13.15, p = 0.003) but not for patients with mild anemia (OR 1.93, 95% CI: 0.85–3.94, p < 0.090). Perioperative ABT was associated with significantly higher risks for complications in nonanemic patients and showed an increased risk for complications in all anemic patients. In RTHS, perioperative ABT as a treatment for moderate preoperative anemia of any cause was associated with a negative compound effect on postoperative complications, compared to anemia or ABT alone. Discussion: ABT is associated with adverse outcomes of patients with moderate preoperative anemia before RTHS. For this reason, medical treatment of moderate preoperative anemia may be considered.
Species dispersal, establishment, and assembly are crucial stages of the life history of plants, and clear understanding ofthe governing forces and rules that shape species composition in a particular community is vital for successful ecologicalrestoration. In this article, we focus on five aspects of seed dispersal and plant establishment, which should be consideredduring habitat restoration actions. In the first two sections, we discuss the success of spontaneous dispersal and establishmenton restoration based on either spatial dispersal or local seed banks. In the third section, we assess the possibilities ofspecies introduction and assisted dispersal. In the fourth section, we introduce some possibilities for the improvement ofestablishment success of spontaneously dispersed or introduced species. Finally, we highlight issues influencing long-termpersistence and sustainability of restored habitats, related to the alteration of management type and intensity, climate change,and spread of non-native species. With the present article, we introduce the special issue entitled “Seed dispersal and soil seedbanks – promising sources for ecological restoration” containing 15 papers by 62 authors from 10 countries arranged in theabovementioned five topics.
Clinically Significant Differences in Acute Pain Measured on Self-report Pain Scales in Children
(2015)
Objectives
The objective was to determine the minimum and ideal clinically significant differences (MCSD, ICSD) in pain intensity in children for the Faces Pain Scale–Revised (FPS-R) and the Color Analog Scale (CAS) and to identify any differences in these estimates based on patient characteristics.
Methods
This was a prospective study of children aged 4 to 17 years with acute pain presenting to two urban pediatric emergency departments. Participants self-reported their pain intensity using the FPS−R and CAS and qualitatively described their changes in pain. Changes in pain score reported using the FPS-R and CAS that were associated with “a little less” and “much less” pain (MCSD and ICSD, respectively) were identified using a receiver operating characteristic–based method and expressed as raw change score and percent reductions. Estimates of MCSD and ICSD were determined for each category of initial pain intensity (mild, moderate, and severe) and patient characteristics (age, sex, and ethnicity). Post hoc exploratory analyses evaluated categories of race, primary language, and etiology of pain.
Results
A total of 314 children with acute pain were enrolled; mean (±SD) age was 9.8 (±3.8) years. The FPS-R raw change score and percent reduction MCSD estimates were 2/10 and 25%, with ICSD estimates of 3/10 and 60%. For the CAS, raw change score and percent reduction MCSD estimates were 1/10 and 15%, with ICSD estimates of 2.75/10 and 52%. For both scales, raw change score and percent reduction estimates of the MCSD remained unchanged in children with either moderate or severe pain. For both scales, estimates of ICSD were not stable across categories of initial pain intensity. There was no difference in MCSD or ICSD based on age, sex, ethnicity, race, primary language, or etiology of pain.
Conclusions
The MCSD estimates can be expressed as raw change score and percent reductions for the FPS-R and CAS. These estimates appear stable for children with moderate to severe pain, irrespective of age, sex, and ethnicity. Estimates of ICSD were not stable across different categories of initial pain intensity, therefore limiting their potential generalizability.
Objectives: The aims of this study were to define the Faces Pain Scale-Revised (FPS-R) and Color Analog Scale (CAS) scores associated with no pain, mild pain, moderate pain, and severe pain in children with acute pain, and to identify differences based on age, sex, and ethnicity.
Methods: We conducted a prospective observational study in 2 pediatric emergency departments of children aged 4 to 17 years with painful and nonpainful conditions. We assessed their pain intensity using the FPS-R, CAS, and qualitative measures. Pain score cut points that best differentiated adjacent categories of pain were identified using a receiver operating characteristic-based method. Cut points were compared within subgroups based on age, sex, and ethnicity.
Results: We enrolled 620 patients, of whom 314 had painful conditions. The mean age was 9.2 years; 315 (50.8%) were in the younger age group (aged 4-7 years); 291 (46.8%) were female; and 341 (55%) were Hispanic. The scores best representing categories of pain for the FPS-R were as follows: no pain, 0 and 2; mild pain, 4; moderate pain, 6; and severe pain, 8 and 10. For the CAS, these were 0 to 1, 1.25 to 2.75, 3 to 5.75, and 6 to 10, respectively. Children with no pain frequently reported nonzero pain scores. There was considerable overlap of scores associated with mild and moderate pain. There were no clinically meaningful differences of scores representing each category of pain based on age, ethnicity, and race.
Conclusions: We defined pain scores for the FPS-R and CAS associated with categories of pain intensity in children with acute pain that are generalizable across subgroups based on patient characteristics. There were minor but potentially important differences in pain scores used to delineate categories of pain intensity compared to prior convention.
Background
Beta-blocker (BB) therapy plays a central role in the treatment of cardiovascular diseases. An increasing number of patients with cardiovascular diseases undergoe noncardiac surgery, where opioids are an integral part of the anesthesiological management. There is evidence to suggest that short-term intravenous BB therapy may influence perioperative opioid requirements due to an assumed cross-talk between G-protein coupled beta-adrenergic and opioid receptors. Whether chronic BB therapy could also have an influence on perioperative opioid requirements is unclear.
Methods
A post hoc analysis of prospectively collected data from a multicenter observational (BioCog) study was performed. Inclusion criteria consisted of elderly patients (≥ 65 years) undergoing elective noncardiac surgery as well as total intravenous general anesthesia without the use of regional anesthesia and duration of anesthesia ≥ 60 min. Two groups were defined: patients with and without BB in their regular preopreative medication. The administered opioids were converted to their respective morphine equivalent doses. Multiple regression analysis was performed using the morphine-index to identify independent predictors.
Results
A total of 747 patients were included in the BioCog study in the study center Berlin. 106 patients fulfilled the inclusion criteria. Of these, 37 were on chronic BB. The latter were preoperatively significantly more likely to have arterial hypertension (94.6%), chronic renal failure (27%) and hyperlipoproteinemia (51.4%) compared to patients without BB. Both groups did not differ in terms of cumulative perioperative morphine equivalent dose (230.9 (BB group) vs. 214.8 mg (Non-BB group)). Predictive factors for increased morphine-index were older age, male sex, longer duration of anesthesia and surgery of the trunk. In a model with logarithmised morphine index, only gender (female) and duration of anesthesia remained predictive factors.
Conclusions
Chronic BB therapy was not associated with a reduced perioperative opioid consumption.
Background:
Children with severe psychomotor impairment (SPMI) often experience sleep disturbances that severely distress both the child and his or her parents. Validated questionnaires for the assessment of parents’ distress related to their child’s sleep disturbances are lacking.
Methods:
We developed and validated a new questionnaire, the HOST (holistic assessment of sleep and daily troubles in parents of children with SPMI) to assess the effect of the sleep disturbances in children with SPMI on their parents. The questionnaire was developed based on published data and expert opinion, and it was refined via direct consultation with affected parents. Its psychometric characteristics were assessed in a sample of parents of 214 children with SPMI. It was retested using a random subsample of the participants.
Results:
Explorative factor analysis revealed that the HOST was composed of four scales. Fit indices, item analysis, and convergent validity (coherence with preexisting instruments of sleep disturbances and health status) were adequate. Retest analysis (n = 62) revealed high stability of the HOST questionnaire and adequate replication validity.
Conclusion:
Sleep-related difficulties significantly impact the sociomedical characteristics of the parents of children with complex neurologic diseases. Typically, parents are severely affected in various aspects of daily life (i.e., medical health, social life, professional life). The HOST proved to be a valid, reliable and economical assessment tool of sleep-related difficulties in parents and relatives of children with SPMI. The HOST is capable of identifying individuals and specific areas requiring intervention.
In this experimental work, the quasi static and fatigue properties of a 40 wt.% long carbon fiber reinforced partially aromatic polyamide (Grivory GCL-4H) were investigated. For this purpose, microstructural parameter variations in the form of different thicknesses and different removal directions from injectionmolded plates were evaluated. Mechanical properties decreased by increasing misalignment away from the melt flow direction. By changing the specimen thickness, no change in the general fiber distribution pattern transversal and normal to the axis of melt flow was observed. It has shown that with increasing specimen thickness the quasi static properties along the melt flow direction decreased and vice versa resulting in superior properties normal to the melt flow axis. At around 5 mm, an intersection suggests quasi-isotropic behavior. In addition, the fatigue strength of the material was significantly higher in the flow direction than normal to the flow direction. No change in fatigue life was observed while changing specimen thickness. The Basquin equation seems to describe the effect of stress amplitude on the fatigue strength of this composite. Scanning electron microscopy was used to investigate fracture surfaces of tested specimens. Results show that mechanical properties and morphological structures depend highly on fiber orientation.
Professionalization in low-threshold drug aid : between managerialism and practitioner knowledge
(2021)
Background
Against the background of a steadily increasing degree of digitalization in health care, a professional information management (IM) is required to successfully plan, implement, and evaluate information technology (IT). At its core, IM has to ensure a high quality of health data and health information systems to support patient care.
Objectives
The goal of the present study was to define what constitutes professional IM as a construct as well as to propose a reliable and valid measurement instrument.
Methods
To develop and validate the construct of professionalism of information management (PIM) and itsmeasurement, a stepwise approach followed an established procedure from information systems and behavioral research. The procedure included an analysis of the pertaining literature and expert rounds on the construct and the
instrument, two consecutive and comprehensive surveys at the national and international level, exploratory and confirmatory factor analyses as well as reliability and validity testing.
Results
Professionalism of information management was developed as a construct consisting of the three dimensions of strategic, tactical, and operational IMas well as of the regularity and cyclical phases of IM procedures as the two elements of professionalism.
The PIM instrument operationalized the construct providing items that incorporated IM procedures along the three dimensions and cyclical phases. These procedures had to be evaluated against their degree of regularity in the instrument. The instrument proved to be reliable and valid in two consecutive measurement phases
and across three countries.
Conclusion
It can be concluded that professionalism of information management is a meaningful construct that can be operationalized in a scientifically rigorous manner. Both science and practice can benefit from these developments in terms of improved self-assessment, benchmarking capabilities, and eventually, obtaining a better understanding of health IT maturity.
Usability is a core construct of website evaluation and inherently defined as interactive. Yet, when analysing first impressions of websites, expected usability, i.e., before use, is of interest. Here we investigate to what extend ratings of expected usability are related to (a) experienced usability, i.e., ratings after use, and (b) objective usability measures, i.e., task performance. Furthermore, we try to elucidate how ratings of expected usability are correlated to aesthetic judgments. In an experiment, 57 participants submitted expected usability ratings after the presentation of website screenshots in three viewing-time conditions (50, 500, and 10,000 ms) and after an interactive task (experienced usability). Additionally, objective usability measures (task completion and duration) and subjective aesthetics evaluations were recorded for each website. The results at both the group and individual level show that expected usability ratings are not significantly related either to experienced usability or objective usability measures. Instead, they are highly correlated with aesthetics ratings. Taken together, our results highlight the need for interaction in empirical website usability testing, even when exploring very early usability impressions. In our study, user ratings of expected usability were no valid proxy neither for objective usability nor for experienced website usability.
Purpose
This paper aims to investigate the strategic consequences of manufacturing location decisions, with a focus on understanding the link between collocating manufacturing with other value chain activities, via reshoring or retaining and organizational agility.
Design/methodology/approach
The paper uses qualitative data from 115 interviews with executives from UK high value manufacturing companies to explore the recent phenomenon of reshoring and the strategic effects of manufacturing location.
Findings
The location of manufacturing is operationally and strategically important for multinational companies. The spatial dispersion of manufacturing is determined by firm-specific and external factors, both of which are subject to constant change. The analysis shows that concentrating on manufacturing in their home countries enables firms to increase organizational agility and stimulate innovation. Better integration with and more extensive collaboration between related value chain activities, such as research and development, sales and marketing, leads to higher flexibility, speed and responsiveness to customer requirements. However, under certain conditions, firms also continue to benefit from the known advantages of offshoring.
Originality/value
This research sheds light on possible strategic downsides of global value chains, characterized by dispersed activities and intermitted processes. The results provide evidence that retaining manufacturing or bringing back manufacturing operations to a company’s home country can increase organizational flexibility, speed, adaptability, innovativeness and responsiveness to customer requirements. As these capabilities are critical for long-term survival, especially in dynamic environments, firms need to review their global factory configurations and determine whether the short-term advantages of foreign locations continue to justify offshoring practices.
Injection of slurry or digestate below maize seeds is a relatively new technique developed to improve nitrogen use efficiency. However, this practice has the major drawback of increasing nitrous oxide (N2O) emissions. The application of a nitrification inhibitor (NI) is an effective method to reduce these emissions. To evaluate the effect of the NI 3,4‐dimethypyrazole phosphate (DMPP) on N2O emissions and the stabilization of ammonium, a two‐factorial soil‐column experiment was conducted. PVC pipes (20 cm diameter and 30 cm length) were used as incubation vessels for the soil‐columns. The trial consisted of four treatments in a randomized block design with four replications: slurry injection, slurry injection + DMPP, digestate injection, and digestate injection + DMPP. During the 47‐day incubation period, N2O fluxes were measured twice a week and cumulated by linear interpolation of the gas‐fluxes of consecutive measurement dates. After completion of the gas flux measurement, concentration of ammonium and nitrate within the soil‐columns was determined. DMPP delayed the conversion of ammonium within the manure injection zone significantly. This effect was considerably more pronounced in treatment digestate + NI than in treatment slurry + NI. Regarding the cumulated N2O emissions, no difference between slurry and digestate treatments was determined. DMPP reduced the release of N2O significantly. Transferring the results into practice, the use of DMPP is a promising way to reduce greenhouse gas emissions and nitrate leaching, following the injection of slurry or digestate.
Easy and inexpensive methods for measuring ammonia emissions in multi-plot field trials allow the comparison of several treatments with liquid manure application. One approach that might be suitable under these conditions is the dynamic tube method (DTM). Applying the DTM, a mobile chamber system is placed on the soil surface, and the air volume within is exchanged at a constant rate for approx. 90 s. with an automated pump. This procedure is assumed to achieve an equilibrium ammonia concentration within the system. Subsequently, a measurement is performed using an ammonia-sensitive detector tube. Ammonia fluxes are calculated based on an empirical model that also takes into account the background ammonia concentration measured on unfertilized control plots. Between measurements on different plots, the chamber system is flushed with ambient air and cleaned with paper towels to minimize contamination with ammonia. The aim of this study was to determine important prerequisites and boundary conditions for the application of the DTM.
We conducted a laboratory experiment to test if the ammonia concentration remains stable while performing a measurement. Furthermore, we investigated the cleaning procedure and the effect of potential ammonia carryover on cumulated emissions under field conditions following liquid manure application. The laboratory experiment indicated that the premeasurement phase to ensure a constant ammonia concentration is not sufficient. The concentration only stabilized after performing more than 100 pump strokes, with 20 pump strokes (lasting approximately 90 s) being the recommendation.
However, the duration of performing a measurement can vary substantially, and linear conversion accounts for those differences, so a stable concentration is mandatory. Further experiments showed that the cleaning procedure is not sufficient under field conditions. Thirty minutes after performing measurements on high emitting plots, which resulted in an ammonia concentration of approx.
10 ppm in the chamber, we detected a residual concentration of 2 ppm. This contamination may affect measurements on plots with liquid manure application as well as on untreated control plots. In a field experiment with trailing hose application of liquid manure, we subsequently demonstrated that the calculation of cumulative ammonia emissions can vary by a factor of three, depending on the degree of chamber system contamination when measuring control plots. When the ammoni background values were determined by an uncontaminated chamber system that was used to measure only control plots, cumulative ammonia emissions were approximately 9 kg NH3-N ha1.
However, when ammonia background values were determined using the contaminated chamber system that was also used to measure on plots with liquid manure application, the calculation of cumulative ammonia losses indicated approximately 3 kg NH3-N ha1. Based on these results, it can be concluded that a new empirical DTM calibration is needed for multi-plot field experiments with high-emitting treatments.
15 δ N signals in plant and soil material integrate over a number of biogeochemical processes
related to nitrogen (N) and therefore provide information on net effects of multiple
processes on N dynamics. In general little is known in many grassland restoration projects
on soil–plant N dynamics in relation to the restoration treatments. In particular, 15 δ N signals
may be a useful tool to assess whether abiotic restoration treatments have produced the
desired result. In this study we used the range of abiotic and biotic conditions provided
by a restoration experiment to assess to whether the restoration treatments and/or plant
functional identity and legume neighborhood affected plant 15 δ N signals. The restoration
treatments consisted of hay transfer and topsoil removal, thus representing increasing
restoration effort, from no restoration measures, through biotic manipulation to major
abiotic manipulation. We measured 15 δ N and %N in six different plant species (two nonlegumes and four legumes) across the restoration treatments. We found that restoration
treatments were clearly reflected in 15 δ N of the non-legume species, with very depleted
15 δ N associated with low soil N, and our results suggest this may be linked to uptake of
ammonium (rather than nitrate). The two non-legume species differed considerably in their
15 δ N signals, which may be related to the two species forming different kinds of mycorrhizal
symbioses. Plant 15 δ N signals could clearly separate legumes from non-legumes, but our
results did not allow for an assessment of legume neighborhood effects on non-legume
15 δ N signals. We discuss our results in the light of what the 15 δ N signals may be telling
us about plant–soil N dynamics and their potential value as an indicator for N dynamics in
restoration.
Background/Aim
This study aimed to establish the somatosensory profile of patients with lumbar radiculopathy at pre-and post-microdiscectomy and to explore any association between pre-surgical quantitative sensory test (QST) parameters and post-surgical clinical outcomes.
Methods
A standardized QST protocol was performed in 53 patients (mean age 38 ± 11 years, 26 females) with unilateral L5/S1 radiculopathy in the main pain area (MPA), affected dermatome and contralateral mirror sites and in age- and gender-,and body site-matched healthy controls. Repeat measures at 3 months included QST, the Oswestry Disability Index (ODI) and numerous other clinical measures; at 12 months, only clinical measures were repeated. A change <30% on the ODI was defined as ‘no clinically meaningful improvement’.
Results
Patients showed a significant loss of function in their symptomatic leg both in the dermatome (thermal, mechanical, vibration detection p < .002), and MPA (thermal, mechanical, vibration detection, mechanical pain threshold, mechanical pain sensitivity p < .041) and increased cold sensitivity in the MPA (p < .001). Pre-surgical altered QST parameters improved significantly post-surgery in the dermatome (p < .018) in the symptomatic leg and in the MPA (p < .010), except for thermal detection thresholds and cold sensitivity. Clinical outcomes improved at 3 and 12 months (p < .001). Seven patients demonstrated <30% change on the ODI at 12 months. Baseline loss of function in mechanical detection in the MPA was associated with <30% change on the ODI at 12 months (OR 2.63, 95% CI 1.09–6.37, p = .032).
Conclusion
Microdiscectomy resulted in improvements in affected somatosensory parameters and clinical outcomes. Pre-surgical mechanical detection thresholds may be predictive of clinical outcome.
Significance
This study documented quantitative sensory testing (QST) profiles in patients with lumbar radiculopathy in their main pain area (MPA) and dermatome pre- and post-microdiscectomy and explored associations between QST parameters and clinical outcome. Lumbar radiculopathy was associated with loss of function in modalities mediated by large and small sensory fibres. Microdiscectomy resulted in significant improvements in loss of function and clinical outcomes in 85% of our cohort. Pre-surgical mechanical detection thresholds in the MPA may be predictive of clinical outcome.
Background and aims
In 2008, the International Association for the Study of Pain Special Interest Group on Neuropathic Pain (NeuPSIG) proposed a clinical grading system to help identify patients with neuropathic pain (NeP). We previously applied this classification system, along with two NeP screening tools, the painDETECT (PD-Q) and Leeds Assessment of Neuropathic Symptoms and Signs pain scale (LANSS), to identify NeP in patients with neck/upper limb pain. Both screening tools failed to identify a large proportion of patients with clinically classified NeP, however a limitation of our study was the use of a single clinician performing the NeP classification. In 2016, the NeuPSIG grading system was updated with the aim of improving its clinical utility. We were interested in field testing of the revised grading system, in particular in the application of the grading system and the agreement of interpretation of clinical findings. The primary aim of the current study was to explore the application of the NeuPSIG revised grading system based on patient records and to establish the inter-rater agreement of detecting NeP. A secondary aim was to investigate the level of agreement in detecting NeP between the revised NeuPSIG grading system and the LANSS and PD-Q.
Methods
In this retrospective study, two expert clinicians (Specialist Pain Medicine Physician and Advanced Scope Physiotherapist) independently reviewed 152 patient case notes and classified them according to the revised grading system. The consensus of the expert clinicians’ clinical classification was used as “gold standard” to determine the diagnostic accuracy of the two NeP screening tools.
Results
The two clinicians agreed in classifying 117 out of 152 patients (ICC 0.794, 95% CI 0.716–850; κ 0.62, 95% CI 0.50–0.73), yielding a 77% agreement. Compared to the clinicians’ consensus, both LANSS and PD-Q demonstrated limited diagnostic accuracy in detecting NeP (LANSS sensitivity 24%, specificity 97%; PD-Q sensitivity 53%, specificity 67%).
Conclusions
The application of the revised NeP grading system was feasible in our retrospective analysis of patients with neck/upper limb pain. High inter-rater percentage agreement was demonstrated. The hierarchical order of classification may lead to false negative classification. We propose that in the absence of sensory changes or diagnostic tests in patients with neck/upper limb pain, classification of NeP may be further improved using a cluster of clinical findings that confirm a relevant nerve lesion/disease, such as reflex and motor changes. The diagnostic accuracy of LANSS and PD-Q in identifying NeP in patients with neck/upper limb pain remains limited. Clinical judgment remains crucial to diagnosing NeP in the clinical practice.
Implications
Our observations suggest that in view of the heterogeneity in patients with neck/upper limb pain, a considerable amount of expertise is required to interpret the revised grading system. While the application was feasible in our clinical setting, it is unclear if this will be feasible to apply in primary health care settings where early recognition and timely intervention is often most needed. The use of LANSS and PD-Q in the identification of NeP in patients with neck/upper limb pain remains questionable.
Background: The painDETECT questionnaire (PD-Q) has been used as a tool to characterize sensory abnormalities in patients with persistent pain. This study investigated whether the self-reported sensory descriptors of patients with painful cervical radiculopathy (CxRAD) and patients with fibromyalgia (FM), as characterized by responses to verbal sensory descriptors from PD-Q (sensitivity to light touch, cold, heat, slight pressure, feeling of numbness in the main area of pain), were associated with the corresponding sensory parameters as demonstrated by quantitative sensory testing (QST).
Methods: Twenty-three patients with CxRAD (eight women, 46.3 ± 9.6 years) and 22 patients with FM (20 women, 46.1 ± 11.5 years) completed the PD-Q. Standardized QST of dynamic mechanical allodynia, cold and heat pain thresholds, pressure pain thresholds, mechanical and vibration detection thresholds, was recorded from the maximal pain area. Comparative QST data from 31 age-matched healthy controls (HCs; 15 women) were obtained.
Results: Patients with CxRAD demonstrated a match between their self-reported descriptors and QST parameters for all sensory parameters except for sensitivity to light touch, and these matches were statistically significant compared with HC data (p ≤ 0.006). The FM group demonstrated discrepancies between the PD-Q and QST sensory phenotypes for all sensory descriptors, indicating that the self-reported sensory descriptors did not consistently match the QST parameters (p = ≤0.017).
Conclusion: Clinicians and researchers should be cautious about relying on PD-Q as a stand-alone screening tool to determine sensory abnormalities in patients with FM.
Identification of differences in clinical presentation and underlying pain mechanisms may assist the classification of patients with neck–arm pain which is important for the provision of targeted best evidence based management. The aim of this study was to: (i) assess the inter-examiner agreement in using specific systems to classify patients with cervical radiculopathy and patients with non-specific neck–arm pain associated with heightened nerve mechanosensitivity (NSNAP); (ii) assess the agreement between two clinical examiners and two clinical experts in classifying these patients, and (iii) assess the diagnostic accuracy of the two clinical examiners. Forty patients with unilateral neck–arm pain were examined by two clinicians and classified into (i) cervical radiculopathy, (ii) NSNAP, (iii) other. The classifications were compared to those made independently by two experts, based on a review of patients' clinical assessment notes. The experts' opinion was used as the reference criterion to assess the diagnostic accuracy of the clinical examiners in classifying each patient group. There was an 80% agreement between clinical examiners, and between experts and 70%–80% between clinical examiners and experts in classifying patients with cervical radiculopathy (kappa between 0.41 and 0.61). Agreement was 72.5%–80% in classifying patients with NSNAP (kappa between 0.43 and 0.52). Clinical examiners' diagnostic accuracy was high (radiculopathy: sensitivity 79%–84%; specificity 76%–81%; NSNAP: sensitivity 78%–100%; specificity 71%–81%). Compared to expert opinion, clinicians were able to identify patients with cervical radiculopathy and patients with NSNAP in 80% of cases, our data supporting the reliability of these classification systems.
Background: Lumbar discectomy is considered a safe, efficacious and cost-effective treatment for selected cases of patients with leg pain associated with the presence of a disc protrusion. But despite technically successful surgery, 30 % of patients complain of persistent pain on long-term follow up. Identification of possible predictors for a negative outcome is important, in the search for appropriate pre- and/or post-operative care and prevention of persistent disability. There is some evidence in the literature that quantitative sensory testing (QST) measures may play a role in prediction of patients’ pain persistency, however, this has never been investigated in patients undergoing lumbar discectomy.
Objective: The aim of this study is to determine the predictive value of QST parameters, in combination with previously documented predictor variables such as medical/psychological/cognitive behavioural factors, in patients with lumbar radiculopathy and/or radicular pain, for predicting patients’ clinical outcome after lumbar discectomy.
Method: Participants with radiculopathy and/or radicular pain and confirmed imaging diagnosis of nerve root compression will be recruited from the elective surgery waitlist at one hospital. All participants will undergo lumbar discectomy performed by one neurosurgeon. A standardized QST protocol comprising all of the somatosensory sub-modalities that are mediated by different primary afferents (C-, Aδ-, Aβ-) will be performed prior to surgery. QST will be conducted in the patients’ main pain area and contralateral side, in the affected dermatome and at a remote control site. The presence of other predictor variables will be captured by questionnaires. Follow-up at 3 months will include QST and measurements of pain intensity, pain descriptors, functional status, health related quality of life, return to work and health care utilisation. A further 1-year follow-up will include the same measurements except QST.
Results/Conclusions: Identification of new predictor variables may assist in the development of pre-surgical screening methods and in targeted pre- and/or post-operative patient care, with the potential to improve patients’ functional status, quality of life, work capacity whilst also reducing health care costs associated with persistent disability
Objectives
To investigate differences in pressure pain thresholds (PPTs) and longitudinal mechanosensitivity of the greater occipital nerve (GON) between patients with side-dominant head and neck pain (SDHNP) and healthy controls. Evaluation of neural sensitivity is not a standard procedure in the physical examination of headache patients but may influence treatment decisions.
Methods
Two blinded investigators evaluated PPTs on two different locations bilaterally over the GON as well as the occipitalis longsitting-slump (OLSS) in subjects with SDHNP (n = 38)) and healthy controls (n = 38).
Results
Pressure pain sensitivity of the GON was lower at the occiput in patients compared to controls (p = 0.001). Differences in pressure sensitivity of the GON at the nucheal line, or between the dominant headache side and the non-dominant side were not found (p > 0.05). The OLSS showed significant higher pain intensity in SDHNP (p < 0.001). In comparison to the non-dominant side, the dominant side was significantly more sensitive (p = 0.004).
Discussion
Palpation of the GON at the occiput and the OLSS may be potentially relevant tests in SDHNP. One explanation for an increased bilateral sensitivity may be sensitization mechanisms. Future research should investigate the efficacy of neurodynamic techniques directed at the GON.
Freshwater ecosystems host disproportionately high numbers of species relative to their surface area yet are poorly protected globally. We used data on the distribution of 1631 species of aquatic plant, mollusc, odonate and fish in 18,816 river and lake catchments in Europe to establish spatial conservation priorities based on the occurrence of threatened, rangerestricted and endemic species using the Marxan systematic conservation planning tool. We found that priorities were highest for rivers and ancient lakes in S Europe, large rivers and lakes in E and N Europe, smaller lakes in NW Europe and karst/limestone areas in the Balkans, S France and central Europe. The a priori inclusion of well-protected catchments resulted in geographically more balanced priorities and better coverage of threatened (critically endangered, endangered and vulnerable) species. The a priori exclusion of well-protected catchments showed that priority areas that need further conservation interventions are in S and E Europe. We developed three ways to evaluate the correspondence between conservation priority and current protection by assessing whether a cathment has more (or less) priority given its protection level relative to all other catchments. Each method found that priority relative to protection was high in S and E Europe and generally low in NW Europe. The inclusion of hydrological connectivity had little influence on these patterns but decreased the coverage of threatened species, indicating a trade-off between connectivity and conservation of threatened species. Our results suggest that catchments in S and E Europe need urgent conservation attention (protected areas, restoration, management, species protection) in the face of imminent threats such as river regulation, dam construction, hydropower development and climate change. Our study presents continental-scale conservation priorities for freshwater ecosystems in ecologically meaningful planning units and will thus be important in freshwater biodiversity conservation policy and practice, and water management in Europe.
Quantitative Umweltbewertungen sind von entscheidender Bedeutung, wenn es darum geht, effektiv auf nachhaltige Produktions- und Verbrauchsmuster hinzuarbeiten. In den letzten Jahrzehnten haben sich Lebenszyklusanalysen (LCA) als praktikables Mittel zur Messung der Umweltauswirkungen von Produkten entlang der Lieferkette etabliert. In Bezug auf Nutzer- und Konsummuster wurden jedoch methodische Schwächen festgestellt, und es wurden mehrere Versuche unternommen, Ökobilanzen entsprechend zu verbessern, beispielsweise durch Einbeziehung von Effekten höherer Ordnung und verhaltenswissenschaftlicher Unterstützung. In einer Diskussion solcher Ansätze zeigen wir, dass den Konzepten des Verbrauchs keine explizite Aufmerksamkeit geschenkt wurde, was häufig zu produktzentrierten Bewertungen führt. Wir führen Theorien sozialer Praktiken ein, um Konsummuster für LCA zugänglich zu machen. Soziale Praktiken sind routinierte Handlungen, die aus miteinander verbundenen Elementen (Materialien, Kompetenzen und Bedeutungen) bestehen, die sie als eine Einheit denkbar machen (z. B. Kochen). Da die meisten sozialen Praktiken eine Art von Verbrauch (Material, Energie, Luft) beinhalten, konnten wir einen Rahmen entwickeln, der soziale Praktiken mit der Ökobilanz verbindet. Der vorgeschlagene Rahmen bietet eine neue Perspektive für quantitative Umweltbewertungen, indem er den Schwerpunkt von Produkten oder Nutzern auf soziale Praktiken verlagert. Dementsprechend sehen wir die Chance darin, die reduktionistische Sichtweise zu überwinden, dass Menschen nur Nutzer von Produkten sind, und sie stattdessen als Praktiker in sozialen Praktiken zu sehen. Dieser Wandel könnte neue Methoden der interdisziplinären Konsumforschung ermöglichen, die beabsichtigte Sozialwissenschaften und wirkungsorientierte Bewertungen einbeziehen. Allerdings bedarf der Rahmen einer weiteren Überarbeitung und vor allem einer empirischen Validierung.
Die Autoren präsentieren einen innovativen Ansatz für Circular Economy (CE)-Strategien mit hohem Umweltpotenzial. Dieser Ansatz betont Verweigerung, Umdenken und Reduzieren, indem die Perspektive vom Verbraucher auf soziale Praktiken verschoben wird. Dabei wird Suffizienz als Schlüsselkonzept eingeführt, das entscheidend ist, um Umweltbelastungen durch CE zu reduzieren und die Transformation des Produktions- und Konsumsystems zu fördern. Die Autoren betonen, dass nachhaltige Kreislaufwirtschaft ohne Suffizienz nicht möglich ist.
Die Studie berichtet über einen Fall, in dem die Über-Verfügbarkeit durch das Konzept der Genügsamkeit ersetzt wurde. Die Autoren beobachten, wie Suffizienz in die Lebensstile der Teilnehmer integriert wird, insbesondere durch einen Farmboxing-Ansatz. Obwohl die Verallgemeinerung des Falls begrenzt ist, zeigt die Studie das Potenzial niedrigschwelliger Einführungen von Suffizienz, wie beispielsweise öffentlicher Raum für gärtnerische Aktivitäten.
Die Autoren betonen jedoch die Notwendigkeit weiterer empirischer Arbeit, um die Bedeutung von Suffizienz in verschiedenen sozialen Praktiken zu verstehen. Sie schlagen vor, dass der Zeitpunkt für solche Forschungen günstig ist, angesichts aktueller globaler Entwicklungen und der steigenden Nachfrage nach Strategien zur Ressourcenverringerung.
Die Studie identifiziert Herausforderungen, darunter unklare Faktoren, die nicht berücksichtigt wurden, wie die Rolle des sozialen Umfelds und die Übertragbarkeit der Ergebnisse auf andere Lebensbereiche. Die Autoren planen bereits eine Folgestudie mit quantitativen und qualitativen Analysen sowie Lebenszyklusanalysen.
Abschließend schlagen die Autoren vor, klare Definitionen für Wiederverwendung, Umdenken und Reduzierung zu entwickeln, basierend auf empirischen Daten und Beispielen sozialer Praktiken. Sie fordern auch dazu auf, Theorien sozialer Praktiken auf andere CE-Strategien anzuwenden, da soziale Praktiken nicht auf privaten Konsum beschränkt sind.
One challenge of the EU energy transition is the integration of renewable electricity generation in the distribution system. EU energy law proposes a possible solution by introducing “citizen energy communities” (Directive 2019/944/EU) which may be open for “cross-border participation”. This article proposes an innovative way of implementing such cross-border communities by linking distribution systems via a “switchable element”, a generation, storage, or consumption asset with a connection to each country. An optimization model has been developed to calculate the system cost savings of such a connection. Linking regions with complementary characteristics regarding electricity generation and demand via a switchable element leads to more efficient system utilization. Findings are relevant for the transposition of “citizen energy communities” in national laws.
Objective: To understand the significance of healthy living for users, professionals and managers of the Family Health Strategy (FHS) team.
Methods: Research of a qualitative nature, based on grounded theory. For data collection, interviews were conducted with 25 participants, including users, professionals and managers of a FHS team, during the period between March and December, 2009. Results: The collection and analysis of data was conducted in a systematic and comparative manner, demonstrating that healthy living can be characterized as a selforganizing process, mediated by the action of the FHS team professionals, especially by the community health agent, through creation of bonds of trust and stimulation of interactions and community associations. Conclusion: We concluded that healthy living is a singular phenomenon, complex, interactive, associative, political and social, coupled with the active involvement and participation of the users and by the engagement of effective and socially responsible professionals, managers and established political authorities.
Objective: To understand the meaning of the Learning Incubator as a teaching and learning technology in the nursing area.
Method: Qualitative research, supported by grounded theory. Data was collected from March to November 2019, through interviews with guiding questions and hypotheses directed at two different groups. The analysis was done by comparative data analysis and included open, axial and integrated coding, as proposed by the method. The theoretical sample included 23 participants, which were nurses, technicians, and nursing students.
Results: The delimitation of the categories converged in the phenomenon (Re)signifying knowledge and practices in the Learning Incubator. Guided by the paradigmatic model, the categories were named according to the three following components: Condition: Recognizing that the being and the professional practice are inextricable; Action/interaction: Revisiting professional practices that are repetitive and mechanic; Consequence: Referring to the reflections and knowledge constructed in the Learning Incubator.
Conclusion: The Learning Incubator, as seen by the study participants, is not limited to the Incubator meetings or the themes addressed in it. Beyond a welcoming physical space, the Incubator expands itself and becomes a tool that promotes self-reflection and self-assessment of professional behaviors and attitudes.
Objective:
to carry out a theoretical reflection on the Nursing Now Campaign and the experience of the unexpected irruptions facing the pandemic period.
Method:
a theoretical-reflective study, supported by the theoretical framework of complexity thinking. It aims at understanding the dialogic between the notions of order, disorder and organization, which translate the transition from simplification to complexity of the pandemic phenomenon and its relation to the theme of Nursing Now and Nursing in the future.
Results:
the universe of phenomena is simultaneously composed of order, disorder and organization. Reasserting the central role of Nursing in the health team, facing the irruptions and uncertainties caused by the current pandemic, implies the ability to dialog with disorder and raise a new and more complex global (re)organization of the being and doing Nursing.
Conclusion:
in addition to answers, theoretical reflection raises new questions and irruptions. The inseparability between the notions of order and disorder in the evolutionary dynamics of the Nursing system is conceived and the promotion of even more complex levels of organization, management and Nursing assistance to achieve universal access to health is advocated.
Objective:
To understand the meaning of entrepreneurial nursing care as inducer of healthy practices in vulnerable communities.
Method:
Grounded theory, whose data collection took place between March and December 2019, from interviews with 19 participants from the central region of Rio Grande do Sul, Brazil and comparative data analysis.
Results:
The phenomenon was delimited: Experiencing small/big transformations in the invisibility of everyday life in promoting healthy practices in vulnerable communities. Conducted by the paradigmatic model, the categories were named based on the components: Condition: Making choices and negotiating non-negotiable exchanges; Action/interaction: Motivating oneself to maintain basic human needs; Consequence: Broadening perspectives and transcending personal and collective boundaries.
Conclusion:
Entrepreneurial nursing care as inducer of healthy practices in vulnerable communities is not reduced to a scientific theory or to the linear and decontextualized apprehension of healthy living, but extends to reach small/big transformations that occur in the invisibility of everyday life.
Objectives: To identify emancipatory strategies to strengthen the social protagonism of recyclable materials collectors in the light of entrepreneurial Nursing care.
Methods: Qualitative study carried out in two stages: field approach from healthcare interventions in a Recycling Materials Association, and individual interviews conducted between October and December 2018.
Results: The analysis resulted in three thematic categories: Social contribution of recyclable materials collectors; From the assistentialist perception to entrepreneurial Nursing care; Emancipatory strategies of recycling work.
Final Considerations: The emancipatory strategies to strengthen the social protagonism of recyclable materials collectors in the light of entrepreneurial Nursing care are related to the appreciation, recognition and enhancement of social work that has been already performed by these professionals, and to the creation of spaces for the socialization of experiences, expectations and perspectives.
Are natural floods accelerators for streambank vegetationdevelopment in floodplain restoration?
(2021)
Riverbanks are very dynamic habitats for riparian vegetation strongly influenced byfluvial and geomorphic processes. This habitat type was severely reduced in the pastby river straightening and bank stabilisation. Restoration and establishment of newfloodplain streams promote this habitat, but a directed succession to later stages wasobserved many times. Our study aimed to analyse whether the often observeddirected succession of the streambank vegetation after restoration implementationcould be reversed by a natural flood along a newly created floodplain stream. Weinvestigated the effects of a natural flood in 2013 and different prerestorationconditions on species development in the riparian zone. Vegetation was studiedalong 12 transects in four different sections from 2011 to 2014. Species composi-tion differed strongly between the sections. Species richness was lowest in a newlydug steep section with high morphological dynamics and highest on wider flatstreambanks. Changes during the years reflecting different hydrological eventsvaried between sections. The high natural flood in 2013 reduced the cover of theherb layer and increased bare ground, which led in most sections to a loss of non-target species. Total target species richness did not change due to the natural flood,while target species showed a high turnover rate. In the following year, however,the flood‐induced development of species composition, in general, was reversed.Natural floods changed abiotic and biotic conditions along the streambank, but theydid not accelerate ecological restoration towards predefined target ecosystems.However, they were necessary to preserve the needed dynamic vegetation changesand species turnover to hinder the succession to later stages dominated by a fewspecies. Our study shows that riparian vegetation near the streambank can bemonitored most effectively in cross‐profile transects, both in the long‐term andevent‐related.
Background
There is no consensus on the instruments for diagnosis of post-intensive care syndrome (PICS). We present a proposal for a set of outcome measurement instruments of PICS in outpatient care.
Methods
We conducted a three-round, semi-structured consensus-seeking process with medical experts, followed each by exploratory feasibility investigations with intensive care unit survivors (n1 = 5; n2 = 5; n3 = 7). Fourteen participants from nine stakeholder groups participated in the first and second consensus meeting. In the third consensus meeting, a core group of six clinical researchers refined the final outcome measurement instrument set proposal.
Results
We suggest an outcome measurement instrument set used in a two-step process. First step: Screening with brief tests covering PICS domains of (1) mental health (Patient Health Questionnaire-4 (PHQ-4)), (2) cognition (MiniCog, Animal Naming), (3) physical function (Timed Up-and-Go (TUG), handgrip strength), and (4) health-related quality of life (HRQoL) (EQ-5D-5L). Single items measure subjective health before and after the intensive care unit stay. If patients report new or worsened health problems after intensive care unit discharge and show relevant impairment in at least one of the screening tests, a second extended assessment follows: (1) Mental health (Patient Health Questionnaire-8 (PHQ-8), Generalized Anxiety Disorder Scale-7 (GAD-7), Impact of Event Scale – revised (IES-R)); (2) cognition (Repeatable Battery for the Assessment of Neuropsychological Status (RBANS), Trail Making Test (TMT) A and B); (3) physical function (2-Minute Walk Test (2-MWT), handgrip strength, Short Physical Performance Battery (SPPB)); and (4) HRQoL (EQ-5D-5L, 12-Item WHO Disability Assessment Schedule (WHODAS 2.0)).
Conclusions
We propose an outcome measurement instrument set used in a two-step measurement of PICS, combining performance-based and patient-reported outcome measures. First-step screening is brief, free-of-charge, and easily applicable by health care professionals across different sectors. If indicated, specialized healthcare providers can perform the extended, second-step assessment. Usage of the first-step screening of our suggested outcome measurement instrument set in outpatient clinics with subsequent transfer to specialists is recommended for all intensive care unit survivors. This may increase awareness and reduce the burden of PICS.
Der Konsumbereich „Ernährung“ verantwortet in Deutschland rund 15 % der Treibhausgasemissionen (THG). Aufgrund der hohen Nachfrage hat die Außer-Haus-Verpflegung eine starke Hebelwirkung, um eine klimaeffiziente Ernährung voranzutreiben – so auch Schulküchen. Aktuell sind der Verpflegung jedes Schulkindes, welches ganzjährig an der Mittagsverpflegung teilnimmt, pro Jahr etwa 264 kg THG-Emissionen zuzuschreiben.
Im Projekt KEEKS „Klima- und energieeffiziente Küche in Schulen” wurden deshalb in 22 Schulküchen von Ganztagsschulen mit täglich insgesamt 5 000 Mittagessen der Status Quo ermittelt. Dies geschah mittels Energiemessungen, Analyse von Ausstattung, Technik und Prozessen der Küchen sowie leitfadengestützten Interviews mit den Küchenleitungen. Treibhausgasemissionen von Menüs und Küchenprozessen wurden bilanziert, Einsparpotenziale identifiziert sowie Handlungsempfehlungen entwickelt und erprobt. Die effektivsten Maßnahmen – die Reduktion und Substitution von Fleisch und Fleischprodukten sowie der Aufbau eines effizienten Abfallmanagements – sparen rund 10 % des Treibhausgasausstoßes einer Schulküche ein. Die erarbeiteten Handlungsempfehlungen können Küchenpersonal dabei unterstützen, eine klimafreundliche, kindgerechte, gesunde und bezahlbare Schulküche zu gestalten.
Dieser Artikel befasst sich mit dem bundesweiten Niveau der Treibhausgas (THG)-Emissionen und des Ressourcenverbrauchs in der deutschen Gemeinschaftsverpflegung im Segment "Business" und deren Einsparpotenziale durch verschiedene Szenarien der ungestützten und gestützten Rezeptüberarbeitung. Unter "ungestützt" verstehen wir in dieser Arbeit die intuitive Optimierung von Rezepten durch die Mitarbeiter von Betrieben der Gemeinschaftsverpflegung. Die "unterstützten" Ansätze mussten dagegen bestimmte Zielvorgaben erfüllen, zum Beispiel der Deutschen Gesellschaft für Ernährung; engl. Deutschen Gesellschaft für Ernährung oder das nachhaltige Niveau. Konkret Konkret wird getestet, wie sich (A) eine ungestützte Rezeptüberarbeitung, (B) eine Rezeptüberarbeitung auf Basis von Ernährungsempfehlungen und (C) eine Rezeptüberarbeitung mit wissenschaftlicher Begleitung die Umweltauswirkungen eines Gerichtes beeinflussen. Als methodischer Rahmen diente eine Online-Befragung von Unternehmen der Gemeinschaftsverpflegung sowie eine Szenarioanalyse auf Menüebene und auf nationaler Ebene durchgeführt. Die Ergebnisse basieren zum einen auf empirischen Daten und zum anderen auf Hochrechnungen. Die Ergebnisse zeigen, dass die die bundesweite Umsetzung der Rezepturüberarbeitung nach wissenschaftlichen Vorgaben - wie konkrete Zielvorgaben für die THG Emissionen pro Portion - bis zu 44% des Ressourcenverbrauchs in der deutschen Betriebsgastronomie einsparen kann (das entspricht 3,4 Millionen Tonnen Ressourcen pro Jahr) und bis zu 40 % der THG-Emissionen (0,6 Millionen Tonnen THG-Emissionen pro Jahr) einsparen.
The nutritional footprint – An assessment tool for health and environmental effects of nutrition
(2014)
Aufgrund seiner starken Umweltauswirkungen gilt der Lebensmittelsektor durch Effekte in Produktion, Verarbeitung, Konsum und Entsorgung gemeinhin als ein wichtiges Handlungsfeld, soll eine gesellschaftliche Transformation in Richtung Nachhaltigkeit weiter vorangetrieben werden. Da Ernährungsgewohnheiten sowohl ökologische Auswirkungen induzieren als auch gesundheitliche Folgen für die Verbraucher haben, sind Konzepte gefragt, die ökologische mit gesundheitlichen Indikatoren kombinieren; diese sind jedoch bislang rar. Das vorgestellte Instrument des Nutritional Footprint greift diese Problemstellung auf und verbindet in einem innovativen Konzept jeweils vier Kernindikatoren beider Dimensionen. Mithilfe des Konzepts erhalten Verbraucher einen Überblick zu Umwelt- und Gesundheitswirkungen ihrer Ernährung. Unternehmen können wiederum interne Datensätze verwalten, Benchmarking betreiben und ihre externe Kommunikationsleistung erweitern.
Der Bereich Ernährung wird in den kommenden Jahrzehnten vor zahlreichen Herausforderungen stehen, die sich aus den veränderten Lebensgewohnheiten und globalen Konsummustern bei gleichzeitig hohem Ressourcenverbrauch ergeben. Vor diesem Hintergrund wird in diesem Papier ein neu entwickeltes Instrument zur Verringerung der Auswirkungen auf die Ernährung vorgestellt, der sogenannte Ernährungsfußabdruck. Das Instrument basiert auf der Umsetzung des Konzepts einer nachhaltigen Ernährung in Entscheidungsprozessen und der Unterstützung einer ressourcenschonenden Gesellschaft. Das Konzept integriert jeweils vier Indikatoren aus den beiden ernährungsrelevanten Bereichen Gesundheit und Umwelt und verdichtet sie zu einem leicht kommunizierbaren Ergebnis, das sich in seinen Ergebnissen auf eine Wirkungsebene beschränkt. Anhand von acht Mittagsmahlzeiten werden die Methodik und ihre Berechnungsverfahren detailliert vorgestellt. Die Ergebnisse unterstreichen die allgemeine wissenschaftliche Sichtweise auf Lebensmittel; Mahlzeiten auf der Basis von tierischem Eiweiß sind im Hinblick auf ihre Gesundheits- und Umweltauswirkungen relevanter. Das Konzept scheint für Verbraucher nützlich zu sein, um ihre eigene Wahl zu bewerten, und für Unternehmen, um ihre internen Daten, ihre Benchmarking-Prozesse oder ihre externe Kommunikationsleistung zu erweitern. Methodische Unzulänglichkeiten und die Interpretation der Ergebnisse werden erörtert, und die Schlussfolgerung zeigt das Potenzial der Instrumente für die Gestaltung von Übergangsprozessen und für die Verringerung des Verbrauchs natürlicher Ressourcen durch die Unterstützung von Entscheidungen und Wahlmöglichkeiten von Lebensmittelanbietern und Verbrauchern.
In einer konsumorientierten Kultur ein Leben der Suffizienz zu führen, ist vielleicht eines der ehrgeizigsten Experimente, die ein Mensch unternehmen kann. Um diese Herausforderung zu untersuchen, haben wir einen sozial-praktischen Ansatz gewählt. Dieser Artikel basiert auf 42 qualitativen Interviews, in denen die Befragten gefragt wurden, warum und wie sie sich in einer westlichen Infrastruktur und Kultur suffizient verhalten. Die Ergebnisse deuten darauf hin, dass Suffizienz-orientierte Menschen bei ihren ressourcenintensiven Handlungen auf bestimmte Bedeutungen in der Alltagspraxis zurückgreifen. Diese Bedeutungen umfassen eine Mischung aus umweltfreundlichen Einstellungen, positiven sozialen Absichten und/oder persönlichen Verpflichtungen zur Sparsamkeit. Darüber hinaus haben wir eine Reihe spezifischer Praktiken identifiziert, darunter Teilen, Recycling und Wiederverwendung, die für einen ressourcenschonenden Lebensstil nützlich sind. Für unsere Befragten sind viele dieser Suffizienz-Praktiken regelmäßig im täglichen Leben anzutreffen und wurden nur selten hinterfragt. Anhand einer zusätzlichen Umfrage zeigen wir, dass diese Routinen zu einem weniger ressourcenintensiven Lebensstil führen und wie es einer kleinen Gruppe von Menschen gelungen ist, sich Suffizienzpraktiken anzugewöhnen. Die Mehrheit sieht jedoch keine Notwendigkeit für eine häufigere Umsetzung solcher Routinen, da die täglichen Entscheidungsprozesse weitgehend auf den Konsum von Produkten ausgerichtet sind.
Jede Ernährung hat Auswirkungen auf den Gesundheitszustand des Einzelnen, auf die Umwelt und auf soziale Belange. Eine wachsende Zahl von Mahlzeiten wird in der Außer-Haus-Verpflegung eingenommen, in der eine systematische Nachhaltigkeitsbewertung nicht zur gängigen Praxis gehört. Um diese Lücke zu schließen, wurde im Rahmen des NAHGAST-Projekts ein Instrument entwickelt. Nach mehr als einem Jahr des Einsatzes des NAHGAST-Online-Tools soll nun überprüft werden, welche positiven Umwelteinflüsse durch den Einsatz des Instruments realisiert werden können. Aus diesem Grund befasst sich dieser Artikel mit der Frage, ob ein Online-Tool Akteure aus dem Bereich des Außer-Haus-Verzehrs in die Lage versetzen kann, ihre Mahlzeiten im Hinblick auf Aspekte einer nachhaltigen Ernährung zu überarbeiten. Darüber hinaus wird die Frage beantwortet, wie sich konkrete Rezeptüberarbeitungen der beliebtesten Mittagsgerichte auf den materiellen Fußabdruck sowie auf den Carbon Footprint auswirken.
Zusammenfassend lässt sich sagen, dass ein Online-Tool individuelle Nachhaltigkeitspfade für Akteure im und ermöglicht eine eigenständige Rezeptüberarbeitung für bereits bestehende Gerichte. Mahlzeiten. Die Ergebnisse zeigen, dass bereits geringfügige Änderungen in den Rezepten zu Einsparungen von bis zu einem Drittel führen können beim CO2-Fußabdruck sowie beim Materialfußabdruck führen können. Bezogen auf den Außer-Haus-Verzehr Außer-Haus-Verzehr ergibt sich daraus das Potenzial für erhebliche Multiplikationseffekte, die den Weg für die Verbreitung von nachhaltiger Ernährung ebnen.
Quantitative Umweltbewertungen sind von entscheidender Bedeutung, wenn es darum geht, effektiv auf nachhaltige Produktions- und Verbrauchsmuster hinzuarbeiten. In den letzten Jahrzehnten haben sich Lebenszyklusanalysen (LCA) als praktikables Mittel zur Messung der Umweltauswirkungen von Produkten entlang der Lieferkette etabliert. In Bezug auf Nutzer- und Konsummuster wurden jedoch methodische Schwächen festgestellt, und es wurden mehrere Versuche unternommen, Ökobilanzen entsprechend zu verbessern, beispielsweise durch Einbeziehung von Effekten höherer Ordnung und verhaltenswissenschaftlicher Unterstützung. In einer Diskussion solcher Ansätze zeigen wir, dass den Konzepten des Verbrauchs keine explizite Aufmerksamkeit geschenkt wurde, was häufig zu produktzentrierten Bewertungen führt. Wir führen Theorien sozialer Praktiken ein, um Konsummuster für LCA zugänglich zu machen. Soziale Praktiken sind routinierte Handlungen, die aus miteinander verbundenen Elementen (Materialien, Kompetenzen und Bedeutungen) bestehen, die sie als eine Einheit denkbar machen (z. B. Kochen). Da die meisten sozialen Praktiken eine Art von Verbrauch (Material, Energie, Luft) beinhalten, konnten wir einen Rahmen entwickeln, der soziale Praktiken mit der Ökobilanz verbindet. Der vorgeschlagene Rahmen bietet eine neue Perspektive für quantitative Umweltbewertungen, indem er den Schwerpunkt von Produkten oder Nutzern auf soziale Praktiken verlagert. Dementsprechend sehen wir die Chance darin, die reduktionistische Sichtweise zu überwinden, dass Menschen nur Nutzer von Produkten sind, und sie stattdessen als Praktiker in sozialen Praktiken zu sehen. Dieser Wandel könnte neue Methoden der interdisziplinären Konsumforschung ermöglichen, die beabsichtigte Sozialwissenschaften und wirkungsorientierte Bewertungen einbeziehen. Allerdings bedarf der Rahmen einer weiteren Überarbeitung und vor allem einer empirischen Validierung.
Chronic illness can have a profound impact on couples’ relationships. In dealing with relational changes, new constructions and forms of relationships may arise. In the context of a larger grounded theory study on relational processes and practices in couples faced with chronic illness, this article focuses on concurrent relationships as an alternative form of relationship construction which embodies an additional relationship existing parallel to that of the couple confronted with chronic illness. Based on qualitative interviews with a subsample of five persons within the larger study, conditions for the development and characteristics of concurrent relationships are presented. From an individual and shared life questioned by chronic illness, concurrent relationships are formed in the attempt to be able to live on together in a new partner relationship and a continued care relationship. This can lead to new constructions of relationships, family life, and social relations in everyday life.
Intensive care units (ICU) are often overflooded with alarms from monitoring devices which constitutes a hazard to both staff and patients. To date, the suggested solutions to excessive monitoring alarms have remained on a research level. We aimed to identify patient characteristics that affect the ICU alarm rate with the goal of proposing a straightforward solution that can easily be implemented in ICUs. Alarm logs from eight adult ICUs of a tertiary care university-hospital in Berlin, Germany were retrospectively collected between September 2019 and March 2021. Adult patients admitted to the ICU with at least 24 h of continuous alarm logs were included in the study. The sum of alarms per patient per day was calculated. The median was 119. A total of 26,890 observations from 3205 patients were included. 23 variables were extracted from patients' electronic health records (EHR) and a multivariable logistic regression was performed to evaluate the association of patient characteristics and alarm rates. Invasive blood pressure monitoring (adjusted odds ratio (aOR) 4.68, 95%CI 4.15–5.29, p < 0.001), invasive mechanical ventilation (aOR 1.24, 95%CI 1.16–1.32, p < 0.001), heart failure (aOR 1.26, 95%CI 1.19–1.35, p < 0.001), chronic renal failure (aOR 1.18, 95%CI 1.10–1.27, p < 0.001), hypertension (aOR 1.19, 95%CI 1.13–1.26, p < 0.001), high RASS (aOR 1.22, 95%CI 1.18–1.25, p < 0.001) and scheduled surgical admission (aOR 1.22, 95%CI 1.13–1.32, p < 0.001) were significantly associated with a high alarm rate. Our study suggests that patient-specific alarm management should be integrated in the clinical routine of ICUs. To reduce the overall alarm load, particular attention regarding alarm management should be paid to patients with invasive blood pressure monitoring, invasive mechanical ventilation, heart failure, chronic renal failure, hypertension, high RASS or scheduled surgical admission since they are more likely to have a high contribution to noise pollution, alarm fatigue and hence compromised patient safety in ICUs.
Background
To detect changes in biological processes, samples are often studied at several time points. We examined expression data measured at different developmental stages, or more broadly, historical data. Hence, the main assumption of our proposed methodology was the independence between the examined samples over time. In addition, however, the examinations were clustered at each time point by measuring littermates from relatively few mother mice at each developmental stage. As each examination was lethal, we had an independent data structure over the entire history, but a dependent data structure at a particular time point. Over the course of these historical data, we wanted to identify abrupt changes in the parameter of interest - change points.
Results
In this study, we demonstrated the application of generalized hypothesis testing using a linear mixed effects model as a possible method to detect change points. The coefficients from the linear mixed model were used in multiple contrast tests and the effect estimates were visualized with their respective simultaneous confidence intervals. The latter were used to determine the change point(s). In small simulation studies, we modelled different courses with abrupt changes and compared the influence of different contrast matrices. We found two contrasts, both capable of answering different research questions in change point detection: The Sequen contrast to detect individual change points and the McDermott contrast to find change points due to overall progression. We provide the R code for direct use with provided examples. The applicability of those tests for real experimental data was shown with in-vivo data from a preclinical study.
Conclusion
Simultaneous confidence intervals estimated by multiple contrast tests using the model fit from a linear mixed model were capable to determine change points in clustered expression data. The confidence intervals directly delivered interpretable effect estimates representing the strength of the potential change point. Hence, scientists can define biologically relevant threshold of effect strength depending on their research question. We found two rarely used contrasts best fitted for detection of a possible change point: the Sequen and McDermott contrasts.
The kiwifruit processing industry is focused on product yield maximization and keeping energy costs and waste effluents to a minimum while maintaining high product quality. In our study, pulsed electric field (PEF) pretreatment enhanced kiwifruit processing to facilitate peelability and specific peeling process and enhanced valorization of kiwifruit waste. PEF optimization was applied to obtain the best treatment parameters. A 32 factorial design of response surface methodology was applied to find the effect of time elapsed after PEF treatment and the PEF-specific energy input on specific peeling force and kiwifruit firmness as response criteria. Under the optimized condition, the specific peeling force decreased by 100, and peelability increased by 2 times. The phenolic content and antioxidant capacity of PEF-treated kiwifruit bagasse were 5.1% and 260% richer than the control sample. Overall, the optimized PEF pretreatments incorporated into kiwifruit processing led to decreased energy demand and increased productivity.
Currently, the modelling of drying processes of plant tissues pre-treated by pulsed electric field (PEF) is following experimentally identified curves or separate heat and mass transfer and diffusion models with different levels of accuracy. This research had two major objectives: mathematical modeling and control of drying process of different vegetables pretreated by PEF during convective drying. The mathematical modeling was based on Luikov's heat and mass transfer model along with properties of different vegetables. Computer modelling was done using the difference method for predicting moisture and the temperature potentials of untreated and PEF-treated vegetables. The formulation and the solution procedures were applied to simulate the simultaneous heat and mass transfer in selected vegetables subjected to the convective drying. Suggested model had a good correlation with experimental results. Moreover, cell disintegration index can be used as a controllable parameter in heat and mass transfer models to predict drying behavior of potato, onion, and carrot tissues. Obtained drying models can be used as a mathematical tool to predict drying behavior for various types of agricultural products pre-treated by pulsed electric field.
The deployment of containers as building modules has grown in popularity over the past years due to their inherent strength, modular construction, and relatively low cost. The upcycled container architecture is being accepted since it is more eco-friendly than using the traditional building materials with intensive carbon footprint. Moreover, owing to the unquestionable urgency of climate change, existing climate-adaptive design strategies may no longer respond effectively as they are supposed to work in the previous passive design. Therefore, this paper explores the conceptual design for an upcycled shipping container building, which is designed as a carbon-smart modular living solution to a single family house under three design scenarios, related to cold, temperate, and hot–humid climatic zones, respectively. The extra feature of future climate adaption has been added by assessing the projected future climate data with the ASHRAE Standard 55 and Current Handbook of Fundamentals Comfort Model. Compared with the conventional design, Rome would gradually face more failures in conventional climate-adaptive design measures in the coming 60 years, as the growing trends in both cooling and dehumidification demand. Consequently, the appropriate utilization of internal heat gains are proposed to be the most promising measure, followed by the measure of windows sun shading and passive solar direct gain by using low mass, in the upcoming future in Rome. Future climate projection further shows different results in Berlin and Stockholm, where the special attention is around the occasional overheating risk towards the design goal of future thermal comfort.
In modern times, closed-loop control systems (CLCSs) play a prominent role in a wide application range, from production machinery via automated vehicles to robots. CLCSs actively manipulate the actual values of a process to match predetermined setpoints, typically in real time and with remarkable precision. However, the development, modeling, tuning, and optimization of CLCSs barely exploit the potential of artificial intelligence (AI). This paper explores novel opportunities and research directions in CLCS engineering, presenting potential designs and methodologies incorporating AI. Combining these opportunities and directions makes it evident that employing AI in developing and implementing CLCSs is indeed feasible. Integrating AI into CLCS development or AI directly within CLCSs can lead to a significant improvement in stakeholder confidence. Integrating AI in CLCSs raises the question: How can AI in CLCSs be trusted so that its promising capabilities can be used safely? One does not trust AI in CLCSs due to its unknowable nature caused by its extensive set of parameters that defy complete testing. Consequently, developers working on AI-based CLCSs must be able to rate the impact of the trainable parameters on the system accurately. By following this path, this paper highlights two key aspects as essential research directions towards safe AI-based CLCSs: (I) the identification and elimination of unproductive layers in artificial neural networks (ANNs) for reducing the number of trainable parameters without influencing the overall outcome, and (II) the utilization of the solution space of an ANN to define the safety-critical scenarios of an AI-based CLCS.
While developing traffic-based cognitive enhancement technology (CET), such as bike accident prevention systems, it can be challenging to test and evaluate them properly. After all, the real-world scenario could endanger the subjects’ health and safety. Therefore, a simulator is needed, preferably one that is realistic yet low cost. This paper introduces a way to use the video game Grand Theft Auto V (GTA V) and its sophisticated traffic system as a base to create such a simulator, allowing for the safe and realistic testing of dangerous traffic situations involving cyclists, cars, and trucks. The open world of GTA V, which can be explored on foot and via various vehicles, serves as an immersive stand-in for the real world. Custom modification scripts of the game give the researchers control over the experiment scenario and the output data to be evaluated. An off-the-shelf bicycle equipped with three sensors serves as a realistic input device for the subject’s movement direction and speed. The simulator was used to test two early-stage CET concepts enabling cyclists to sense dangerous traffic situations, such as trucks approaching from behind the cyclist. Thus, this paper also presents the user evaluation of the cycling simulator and the CET used by the subjects to sense dangerous traffic situations. With the knowledge of the first iteration of the user-centered design (UCD) process, this paper concludes by naming improvements for the cycling simulator and discussing further research directions for CET that enable users to sense dangerous situations better.
Stainless steel made to rust: a robust water-splitting catalyst with benchmark characteristics
(2015)
The oxygen evolution reaction (OER) is known as the efficiency-limiting step for the electrochemical cleavage of water mainly due to the large overpotentials commonly used materials on the anode side cause. Since Ni–Fe oxides reduce overpotentials occurring in the OER dramatically they are regarded as anode materials of choice for the electrocatalytically driven water-splitting reaction. We herewith show that a straightforward surface modification carried out with AISI 304, a general purpose austenitic stainless steel, very likely, based upon a dissolution mechanism, to result in the formation of an ultra-thin layer consisting of Ni, Fe oxide with a purity >99%. The Ni enriched thin layer firmly attached to the steel substrate is responsible for the unusual highly efficient anodic conversion of water into oxygen as demonstrated by the low overpotential of 212 mV at 12 mA cm−2 current density in 1 M KOH, 269.2 mV at 10 mA cm−2 current density in 0.1 M KOH respectively. The Ni, Fe-oxide layer formed on the steel creates a stable outer sphere, and the surface oxidized steel samples proved to be inert against longer operating times (>150 ks) in alkaline medium. In addition Faradaic efficiency measurements performed through chronopotentiometry revealed a charge to oxygen conversion close to 100%, thus underpinning the conclusion that no “inner oxidation” based on further oxidation of the metal matrix below the oxide layer occurs. These key figures achieved with an almost unrivalled-inexpensive and unrivalled-accessible material, are among the best ever presented activity characteristics for the anodic water-splitting reaction at pH 13.
Purpose
This study operationalizes risks in stakeholder dialog (SD). It conceptualizes SD as co-produced organizational discourse and examines the capacities of organizers' and stakeholders' practices to create a shared understanding of an organization’s risks to their mutual benefit. The meetings and online forum of a German public service media (PSM) organization were used as a case study.
Design/methodology/approach
The authors applied corpus-driven linguistic discourse analysis (topic modeling) to analyze citizens' (n = 2,452) forum posts (n = 14,744). Conversation analysis was used to examine video-recorded online meetings.
Findings
Organizers suspended actors' reciprocity in meetings. In the forums, topics emerged autonomously. Citizens' articulation of their identities was more diverse than the categories the organizer provided, and organizers did not respond to the autonomous emergence of contextualizations of citizens' perceptions of PSM performance in relation to their identities. The results suggest that risks arise from interactionally achieved occasions that prevent reasoned agreement and from actors' practices, which constituted autonomous discursive formations of topics and identities in the forums.
Originality/value
This study disentangles actors' practices, mutuality orientation and risk enactment during SD. It advances the methodological knowledge of strategic communication research on SD, utilizing social constructivist research methods to examine the contingencies of organization-stakeholder interaction in SD.
Reconnection of floodplains to rivers to enhance fluvial dynamics is a favored method of floodplain restoration in Europe. It is believed that the restoration of hydrological conditions of the floodplain facilitates natural dispersal of target species, and hence the reestablishment, of typical plant communities. The aim of our study was to investigate whether floodplain target species could reach restoration sites via hydrochorous dispersal. We analyzed seed inflow from the river and seed dispersal in different sectors of a new watercourse in the Danube floodplain. Seeds were captured using 27 seed traps during three sampling periods of 3 weeks each from summer 2011 to spring 2012. After germination seedlings were identified, we detected a total of almost 39,000 seeds of 176 species, including 80 target species of riparian habitats. We found significant differences between seasons (most seeds in autumn/winter) and between stream sectors. Fewer seeds came in from the Danube (2,800 seeds) than were transported within the floodplain. Several new floodplain target species were detected, which had not been found in the aboveground vegetation or soil seed bank before the start of the restoration. Seeds of nonnative species did not disperse further than approximately 1 km. Our results indicated that hydrochorous seed dispersal from upstream habitats along the new watercourse was important for the establishment of target species and hence for the success of floodplain restoration. Technical water diversion weirs must be traversable for seeds, and small donor sectors upstream might enhance the reestablishment of target vegetation along new sectors downstream.
Soil seed banks have a high potential for vegetation re-establishment in restoration projects. We studied the soil seed bank in an oxbow system of a disconnected floodplain of the Danube River in Southern Germany. The aim of the study was to analyze if floodplain target species were still present in the seed bank after more than 150 years of embankment and disconnection from fluvial dynamics. In this context we investigated seed density, seed bank species richness and species composition in four broad habitat types with and without water-level fluctuations during the time of embankment (permanent water, fluctuating water, reed bed, hardwood floodplain forest). In addition, the similarity between seed bank and above-ground vegetation in these habitat types was studied in order to predict the success of future restoration measures. In total, 124 vascular plant species were determined in the seed bank samples. More than 50 % (66 species) were target species typical for floodplain habitats and 26 of these target species were lost or very rare in the above-ground vegetation. The four habitat types differed significantly in mean seed density and mean species richness. Mean species richness and the number of target species in the seed bank as well as the mean seed density were greatest in the habitats with fluctuating water level whereas mean seed density was much lower in the parts with more or less stable conditions like permanently standing water and hardwood floodplain forest. Sørensen similarity between seed bank and above-ground vegetation was very low in habitats with more or less stable water levels and desirable floodplain target species were very rare or completely absent. Our results indicate that the soil seed bank can be an important seed reservoir for the ecological restoration of floodplain plant communities especially for habitats with unstable environmental conditions during the period of disconnection. Restoration of water level dynamics is important to maintain the seed bank of populations of floodplain target species.
Ecological restoration of an urban demolition site through introduction of native forb species
(2020)
Urban brownfields can provide habitats for endangered native plant species but may also support invasive non native species. The aim of our study was to develop and test different measures for the ecological restoration of an urban brownfield. We aimed to enhance native plant species richness by seeding two native forb seed mix tures containing 25 (HD-mixture) and 13 species (LD-mixture), respectively, without affecting spontaneously colonizing plant species of nature conservation value. Additionally, we tested the effects of species introduction and mowing on the establishment of invasive non-native plant species, woody species development, and per ennial grass dominance.
With establishment rates of 84% (HD) and 92% (LD) in the first study year and about 60% for both seed mixtures in the fourth year, species introduction was successful and led to rapid re-vegetation of the formerly bare brownfield. Although seeding did not result in increased species richness using either seed mixture, the dominance of perennial grasses was significantly reduced by the establishment of forb species from the HD mixture. Overall, we observed the spontaneous establishment of ten plant species of conservation value; seeding affected these species only temporarily, in the second year. Seeding (with either seed mixture) and mowing both resulted in reduced cover of woody species and a lower frequency of the invasive non-native Robinia pseudoa cacia.
Our study demonstrated the high nature conservation value of dry and nutrient-poor urban brownfields for spontaneously occurring threatened plant species. Further studies have to evaluate if the reduced cover of grasses and woody species through introduction of site-specific native forb species might increase these sites’ aesthetic value and broader acceptance of brownfields by urban residents.
Extensive green roofs (EGRs) offer several beneficial ecosystem services for sustainable urban development. However, most standard green roofs have been designed with species-poor plant mixtures containing non-native species. Aiming to increase the nature conservation values of EGRs, we developed and tested a vascular plant seed mixture including regionally occurring native sandy dry grassland species in experimental miniature roofs in Northwestern Germany (temperate oceanic climate) over 4 years. We tested the mixture at two seed densities (1 and 2 g/m2). Additionally, we tested seeding at 1 g/m2 and introducing raked plant material collected from an ancient dry grassland. The total establishment rates of sown species reached 92–96% in the first year, but dropped to 40–60% in the last 2 years, with the highest values for the plots with raked material. Twenty-four additional species (11 vascular, 7 lichen, and 6 moss species, including 7 red-list species) typical of sandy dry grasslands were introduced through the raked material. Vascular plants reached 60–70% cover in the second year. Severe drought periods in the third and the fourth year led to a strong decline of vascular plant cover then. As this cover was higher in the plots with raked material, we assume facilitative effects through the well-developed cryptogam layer containing a mix of pleurocarpous and acrocarpous mosses and lichens. Spontaneously establishing acrocarpous mosses in sown plots did not seem to provide this same function. We conclude that EGRs designed with regionally occurring sandy dry grassland plant species and especially the application of raked plant material from ancient grassland is a fruitful approach to increase the value of green roofs for native phytodiversity.
Green roofs are known to mitigate the negative effects of urban consolidation by offering diverse ecosystem functions compared to non-vegetated roofs. However, the support for native biodiversity might be improved by using native plant species. In a mesocosm experiment, we studied the suitability of three commercial green-roof growth substrates for the establishment of 27 native plant species from dry sandy grasslands of northwestern Germany over the course of four years. The substrates were mineral-based, but differed in the layering of organic matter. Total establishment rates reached 44–59% in Year 4, indicating the general suitability of the substrates. During the first weeks after seeding, with light irrigation, the vascular plant cover was greater in the similar substrates Zincolit® Plus (Z) and Zincolit® Plus-Leicht (ZL) with their compost-based organic mulch layers than in the substrate Sedumteppich (ST) with its organic matter evenly admixed with the mineral aggregates. In Years 2 and 3, however, the vascular plant cover was greater in the ST substrate, likely due to the better availability of water and nutrients from the organic matter compared to the dry surface-mulch layer variants Z and ZL. After severe drought events, the decline in plant cover was more pronounced in the ST substrate, likely representing a trade-off between lush growth and a susceptibility to drought. An indicator-species analysis revealed differences in species composition between the ST and Z/ZL substrates. Annual plant species were indicators of the ST substrate. Perennials, such as Thymus pulegioides and Achillea millefolium, were typical of the Z and ZL substrates. In addition to the general suitability of the tested standard substrates for target species establishment, the study indicated that a combination of different layers of substrate components resulted in different vegetation patterns that may have a positive effect on green-roof biodiversity.
In urban areas, open space including brownfields often became rare due to increasing urbanisation. Urban brownfields can be important for biodiversity, but especially brownfields in early successional stages seem to be refused by urban residents due to their sparse vegetation and less aesthetic appearance. The aim of this study was to revegetate a young demolition site in the city core of Osnabrück, Germany and thereby to support native plant diversity and aesthetic values. We developed two seed mixtures of native plant species and tested them in a large-scale field experiment over two growing seasons.
Both seed mixtures developed towards structurally diverse and flower-rich vegetation. Establishment rates of sown species were consistently larger than 75%. Revegetation of the predominantly bare anthropogenically transformed soil by introduced species occurred fast. Vascular plant cover and vegetation height were higher on sown plots than in controls, but did not differ between the seed mixtures. Seeding did not increase plant species richness and did not reduce the establishment of a potentially invasive non-native plant species. The cover of Red-List species from the spontaneous vegetation was significantly higher in control plots. Our results indicate that not all aims can be reached on one restoration site. It has to be discussed if it is better to invest a restoration budget for measures aiming to increase acceptance of endangered pioneer plant species from the spontaneous vegetation or to introduce more attractive and more competitive species of later successional stages.
This study examined the relation between employees' perceived extent of change and adaptive performance, focusing on the roles of expressive suppression (i.e. the habit of suppressing overt expressions of emotion) at work and perceived strain. Analysing survey data of 153 employees in Germany with different occupational backgrounds via bootstrapping, the conceptual moderated indirect effect scheme was supported. As hypothesized, greater changes were associated with higher strain. Strain, in turn, was negatively related to adaptive performance. Although extent of change did not directly affect adaptive performance, the data supported the expected indirect relationship via strain. Finally, expressive suppression at work acted as a buffer of this indirect effect: extent of change was only negatively related to strain for employees low in suppression. In line with newer evidence, our results indicate that the suppression of overt emotional expressions at work can have positive effects under certain circumstances.