Issue | Example | Tip |
---|---|---|
: Errors of completeness happen when respondents cannot find an appropriate response category. | Any question about education must include “no education." | Ensure that your response categories are fully exhaustive, and include “don’t know,” “prefer not to reply,” and “other (specify)” wherever relevant. Extensive questionnaire piloting will also help identify incompleteness in response options. |
When categories overlap there may be multiple ways that a respondent can answer a question. | If categories run 0-5, 5-10, 10-15, etc., then respondents whose answer is 5 have two possible categories they should choose. | Ensure that all categories are mutually exclusive. |
Issue | Example of sub-optimal approach | Example of better approach |
---|---|---|
Individuals may vary in the accuracy or completeness of their recollections. One way around this is to ask respondents to record information in real time. | “What did you eat for dinner on Tuesday 3 weeks ago?” | “I asked you to keep a food diary to record what you eat every day. Could you show me your food diary for Tuesday 3 weeks ago?” |
Individuals tend to rely too heavily on the first (or sometimes most recent) piece of information they see and will be more likely to give an answer that is close to that information. Avoid adding anchors to questions wherever possible. | “Most people have 3 meals per day. How many meals per day do you think is normal?” | “People vary in the number of meals they consume per day. How many meals per day do you think is normal?” |
: Respondents may be biased if a question is framed to suggest a particular answer—especially if the question or answer implies approval of one response over others. Frame all questions as neutrally as possible | Candidate X has fantastic policies around health and education. Would you consider voting for Candidate X? | Would you consider voting for Candidate X? |
: Respondents will tend to answer questions in a manner that is favorable to others, i.e., emphasize strengths, hide flaws, or avoid stigma. They may be reluctant to admit to behavior that is not generally approved of. Try to ask questions indirectly, ensure that respondents have complete privacy (and remind them of it!) and try to make sensitive questions less specific. Additional information on asking sensitive questions can be found in the and sections as well as our . | Hitting your child to discipline them is illegal in your country. Have you ever hit your child to discipline them? | People have different strategies for teaching discipline to their children. Have you ever hit your child to discipline them? |
: People tend to perceive recent events as being more remote (backward telescoping) and distant events as being more recent (forward telescoping), which can lead to over- or under-reporting. | What big purchases have you made in the last year? | What big purchases have you made since the 20th January last year? Please don’t include any purchases you made before that. Note that If you asked the same question at baseline it is even better to say something like: When I visited you before you said you had bought X and Y in the last year. What big purchases have you made since then? |
Individuals may choose the first in a list of options that sounds acceptable without listening to the rest—or they may choose the last one as it’s the most recent and easiest to remember. Options to get around this include limiting the length of question/number of answers and randomizing response order for lists. | Always using the same question ordering or allowing respondents to pick an answer before hearing all of the options. | Randomize the question ordering and insist that the respondents hear the whole list before choosing an option; or instruct enumerators to not read out the answer options, and train them to select those options that best reflect the respondent’s answer. |
Respondents have incentive to misreport if their answers may determine qualification for a program or whether they meet certain requirements. | If a certain level of school attendance is required to qualify for a government program then respondents may overstate the amount that their children attended school. | Stress anonymity/privacy, or use proxy measures or direct observation (e.g., the school’s attendance records) rather than self-reported answers. |
The intervention itself may cause the treatment (or control) group to be more likely to record certain events, more likely to respond to a question, or more likely to appear in administrative records | An intervention aims to decrease the incidence of a disease in a population through an innovative treatment, and the intervention involves a campaign to increase the number of individuals who go to the doctor. At endline, it appears that the incidence of the disease has increased, when in reality the campaign to get people to the hospital was successful, so more cases of the disease were recorded. | Be sure that the ability to measure an outcome is not correlated with the treatment assignment (e.g., any campaign to increase hospital attendance in treatment villages should also take place in control villages) and identify how the intervention may affect the response process and choose variables that are less susceptible to bias/easier to verify. |
Administrative data may suffer from the same sources of bias as survey data. As the researcher does not have a say in the data collection and processing phase, additional work may be needed to assess data accuracy. Common types of bias in administrative data include:
Reporting bias: As with primary data collection, respondents may have incentive to over- or under-report. An individual may under-report income to qualify for a social welfare program, while an administrative organization such as a school may overreport attendance to meet requirements. While the incentives to misreport may be stronger than with survey data, the problem is mitigated by the fact that much administrative data is not self-reported. To address reporting bias:
Identify the context in which the data were collected. Were there incentives to misreport information?
Choose variables that are not susceptible to bias (e.g., hospital visit rather than value of insurance claim)
Differential coverage: In addition to the issues listed above, in administrative data there may be additional differential coverage between those in the treatment vs control groups: i) differential ability to link individuals to administrative records and ii) differential probability of appearing in administrative records (e.g., victimization as measured by calls to report a crime).
Selection bias in administrative data occurs when administrative records only exist for individuals or organizations in contact with the administration in question. This could occur with program recipients, applicants, partner schools and hospitals, and so on.
Ask: what is the reason for the organization to collect this data?
To address differential coverage and selection bias:
Identify the data universe
Which individuals are included in the data and which are excluded, and why?
Identify how the intervention may affect the reporting of outcomes
Determine the direction in which differential selection might occur and how this might bias effect estimates.
Collect a baseline survey with identifiers for linking
This will ensure that you are equally likely to link treatment and control individuals to their records and identify differential coverage.
The severity of measurement error depends on the type and extent of error, as well as whether the bias is correlated with the treatment.
Bias that is uncorrelated with the treatment affects both the treatment and control equally, and so will not bias the estimate of the difference between the two groups at endline.
Bias that is correlated with the treatment is more serious: it affects the treatment and control groups differently, meaning that the estimate of the difference between the groups at endline is biased on average. This might lead to an erroneous conclusion about the sign and magnitude of the treatment effect.
Administrative data resources:
Non-administrative data:
Last updated February 2022.
These resources are a collaborative effort. If you notice a bug or have a suggestion for additional content, please fill out this form .
We thank Liz Cao , Ben Morse and Katharina Kaeppel for helpful comments. All errors are our own.
The Questionnaire Design section of the World Bank’s DIME Wiki, including:
Grosh and Glewwe’s Designing Household Survey Questionnaires for Developing Countries: Lessons from 15 Years of the Living Standards Measurement Study
McKenzie’s Three New Papers Measuring Stuff that is Difficult to Measure and Using BDM and TIOLI to measure the demand for business training in Jamaica via The World Bank’s Development Impact Blog
J-PAL’s Practical Guide to Measuring Girls’ and Women’s Empowerment in Impact Evaluations
Bradburn, N. M., Sudman, S., & Wansink, B. (2004). Asking questions: The definitive guide to questionnaire design : for market research, political polls, and social and health questionnaires (Rev). San Francisco: Jossey-Bass.
Deaton, A., & Zaidi, S. (2002). Guidelines for constructing consumption aggregates for welfare analysis (No. 135). World Bank Publications.
Deaton, Angus S., Measuring Poverty (July 2004). Princeton Research Program in Development Studies Working Paper. Available at SSRN: https://ssrn.com/abstract=564001 or http://dx.doi.org/10.2139/ssrn.564001
Fowler, F. J. (op. 1995). Improving survey questions: Design and evaluation. Thousand Oaks [etc.]: Sage.
Marsden, P. V., & Wright, J. D. (2010). Handbook of survey research (2nd). Bingley, UK: Emerald.
Saris, W. E., & Gallhofer, I. N. (2007). Design, evaluation, and analysis of questionnaires for survey research. Wiley series in survey methodology. Hoboken, N.J: Wiley-Interscience.
Tourangeau, R., Rips, L. J., & Rasinski, K. A. (2000). The psychology of survey response. Cambridge, U.K, New York: Cambridge University Press.
Abay, Kibrom A., Leah EM Bevis, and Christopher B. Barrett. "Measurement Error Mechanisms Matter: Agricultural intensification with farmer misperceptions and misreporting." American Journal of Agricultural Economics (2019).
Bursztyn, L., M. Callen, B. Ferman, A. Hasanain, & A. Yuchtman. 2014. "A revealed preference approach to the elicitation of political attitudes: experimental evidence on anti-Americanism in Pakistan." NBER Working Paper No. 20153.
Feeney, Laura (with assistance from Sachsse, Clare). “ Measurement ." Lecture, Delivered in J-PAL North America 2019 Research Staff Training (J-PAL internal resource)
Glennerster , Rachel and Kudzai Takavarasha. 2013. Running Randomized Evaluations: A Practical Guide . Princeton University Press: Princeton, NJ.
Karlan, Dean. “ 3.2 Measuring Sensitive Topics ." (J-PAL internal resource)
Sachsse, Clare. “ Theory of Change and Outcomes Measurement [California Franchise Tax Board / CA FTB]”, Delivered in J-PAL’s May 2019 CA FTB training. (J-PAL internal resource)
Sadamand, Nomitha. “ Measuring Better: What to measure, and How?” Lecture, Delivered in J-PAL South Asia’s 2019 Measurement and Survey Design Course. (J-PAL internal resource)
Sautmann, Anja. “ Measurement .” Lecture, Delivered in J-PAL North America’s 2018 Evaluating Social Programs Exec Ed Training.
Sudman, S. & N. Bradburn. 1982. Asking Questions: a Practical Guide to Questionnaire Design . A Wiley Imprint.
Do you need support in running a pricing or product study? We can help you with agile consumer research and conjoint analysis.
Conjointly offers a great survey tool with multiple question types, randomisation blocks, and multilingual support. The Basic tier is always free.
Fully-functional online survey tool with various question types, logic, randomisation, and reporting for unlimited number of surveys.
Completely free for academics and students .
The level of measurement refers to the relationship among the values that are assigned to the attributes for a variable. What does that mean? Begin with the idea of the variable, in this example “party affiliation.”
That variable has a number of attributes. Let’s assume that in this particular election context the only relevant attributes are “republican”, “democrat”, and “independent”. For purposes of analyzing the results of this variable, we arbitrarily assign the values 1 , 2 and 3 to the three attributes. The level of measurement describes the relationship among these three values. In this case, we simply are using the numbers as shorter placeholders for the lengthier text terms. We don’t assume that higher values mean “more” of something and lower numbers signify “less”. We don’t assume the value of 2 means that democrats are twice something that republicans are. We don’t assume that republicans are in first place or have the highest priority just because they have the value of 1 . In this case, we only use the values as a shorter name for the attribute. Here, we would describe the level of measurement as “nominal”.
First, knowing the level of measurement helps you decide how to interpret the data from that variable. When you know that a measure is nominal (like the one just described), then you know that the numerical values are just short codes for the longer names. Second, knowing the level of measurement helps you decide what statistical analysis is appropriate on the values that were assigned. If a measure is nominal, then you know that you would never average the data values or do a t-test on the data.
There are typically four levels of measurement that are defined:
In nominal measurement the numerical values just “name” the attribute uniquely. No ordering of the cases is implied. For example, jersey numbers in basketball are measures at the nominal level. A player with number 30 is not more of anything than a player with number 15 , and is certainly not twice whatever number 15 is.
In ordinal measurement the attributes can be rank-ordered. Here, distances between attributes do not have any meaning. For example, on a survey you might code Educational Attainment as 0=less than high school; 1=some high school.; 2=high school degree; 3=some college; 4=college degree; 5=post college. In this measure, higher numbers mean more education. But is distance from 0 to 1 same as 3 to 4? Of course not. The interval between values is not interpretable in an ordinal measure.
In interval measurement the distance between attributes does have meaning. For example, when we measure temperature (in Fahrenheit), the distance from 30-40 is same as distance from 70-80. The interval between values is interpretable. Because of this, it makes sense to compute an average of an interval variable, where it doesn’t make sense to do so for ordinal scales. But note that in interval measurement ratios don’t make any sense - 80 degrees is not twice as hot as 40 degrees (although the attribute value is twice as large).
Finally, in ratio measurement there is always an absolute zero that is meaningful. This means that you can construct a meaningful fraction (or ratio) with a ratio variable. Weight is a ratio variable. In applied social research most “count” variables are ratio, for example, the number of clients in past six months. Why? Because you can have zero clients and because it is meaningful to say that “…we had twice as many clients in the past six months as we did in the previous six months.”
It’s important to recognize that there is a hierarchy implied in the level of measurement idea. At lower levels of measurement, assumptions tend to be less restrictive and data analyses tend to be less sensitive. At each level up the hierarchy, the current level includes all of the qualities of the one below it and adds something new. In general, it is desirable to have a higher level of measurement (e.g. interval or ratio) rather than a lower one (nominal or ordinal).
Conjointly uses essential cookies to make our site work. We also use additional cookies in order to understand the usage of the site, gather audience analytics, and for remarketing purposes.
For more information on Conjointly's use of cookies, please read our Cookie Policy .
I am new to conjointly, i am already using conjointly.
Discover the world's research
Accessibility Links
Click here to close this panel.
Purpose-led Publishing is a coalition of three not-for-profit publishers in the field of physical sciences: AIP Publishing, the American Physical Society and IOP Publishing.
Together, as publishers that will always put purpose above profit, we have defined a set of industry standards that underpin high-quality, ethical scholarly communications.
We are proudly declaring that science is our only shareholder.
Launched in 1923 Measurement Science and Technology was the world's first scientific instrumentation and measurement journal and the first research journal produced by the Institute of Physics. It covers all aspects of the theory, practice and application of measurement, instrumentation and sensing across science and engineering.
Open all abstracts , in this tab
Hamidreza Eivazi et al 2024 Meas. Sci. Technol. 35 075303
High-resolution reconstruction of flow-field data from low-resolution and noisy measurements is of interest due to the prevalence of such problems in experimental fluid mechanics, where the measurement data are in general sparse, incomplete and noisy. Deep-learning approaches have been shown suitable for such super-resolution tasks. However, a high number of high-resolution examples is needed, which may not be available for many cases. Moreover, the obtained predictions may lack in complying with the physical principles, e.g. mass and momentum conservation. Physics-informed deep learning provides frameworks for integrating data and physical laws for learning. In this study, we apply physics-informed neural networks (PINNs) for super-resolution of flow-field data both in time and space from a limited set of noisy measurements without having any high-resolution reference data. Our objective is to obtain a continuous solution of the problem, providing a physically-consistent prediction at any point in the solution domain. We demonstrate the applicability of PINNs for the super-resolution of flow-field data in time and space through three canonical cases: Burgers' equation, two-dimensional vortex shedding behind a circular cylinder and the minimal turbulent channel flow. The robustness of the models is also investigated by adding synthetic Gaussian noise. Furthermore, we show the capabilities of PINNs to improve the resolution and reduce the noise in a real experimental dataset consisting of hot-wire-anemometry measurements. Our results show the adequate capabilities of PINNs in the context of data augmentation for experiments in fluid mechanics.
Simon Laflamme et al 2023 Meas. Sci. Technol. 34 093001
Structural health monitoring (SHM) is the automation of the condition assessment process of an engineered system. When applied to geometrically large components or structures, such as those found in civil and aerospace infrastructure and systems, a critical challenge is in designing the sensing solution that could yield actionable information. This is a difficult task to conduct cost-effectively, because of the large surfaces under consideration and the localized nature of typical defects and damages. There have been significant research efforts in empowering conventional measurement technologies for applications to SHM in order to improve performance of the condition assessment process. Yet, the field implementation of these SHM solutions is still in its infancy, attributable to various economic and technical challenges. The objective of this Roadmap publication is to discuss modern measurement technologies that were developed for SHM purposes, along with their associated challenges and opportunities, and to provide a path to research and development efforts that could yield impactful field applications. The Roadmap is organized into four sections: distributed embedded sensing systems, distributed surface sensing systems, multifunctional materials, and remote sensing. Recognizing that many measurement technologies may overlap between sections, we define distributed sensing solutions as those that involve or imply the utilization of numbers of sensors geometrically organized within (embedded) or over (surface) the monitored component or system. Multi-functional materials are sensing solutions that combine multiple capabilities, for example those also serving structural functions. Remote sensing are solutions that are contactless, for example cell phones, drones, and satellites. It also includes the notion of remotely controlled robots.
Malcolm A Lawn et al 2024 Meas. Sci. Technol. 35 105018
Precise control of advanced materials relies on accurate dimensional metrology at the sub-nanometre scale. At this scale, the accuracy of scanning probe microscopy (SPM) has been limited by the lack of traceable transfer standard artefacts with calibration structures of suitable dimensions. With the adoption in 2019 of the silicon crystal lattice spacing as a secondary realization of the metre in the International System of Units (SI), SPM users have direct access to a realization of the SI metre at the sub-nanometre level by means of the step height of self-assembled monatomic lattice steps that can form on the surface of silicon crystals. A key challenge of successfully adopting this pathway is establishing protocols to minimize measurement errors and artefacts in routine laboratory use. In this study, step height measurements of monoatomic lattice steps in an ordinal/staircase structure on a Si(111) crystal surface have been derived from images acquired with a commercially available, research-level atomic force microscope (AFM). Measurement results derived from AFM images using three different SPM image processing and analysis software packages are compared. Significant sources of measurement uncertainty are identified, principally the contribution from the dependence on scan direction. The calibration of the AFM derived from this measurement was used to traceably measure the sub-nanometre lattice steps on a silicon carbide crystal surface to demonstrate the viability of this calibration pathway.
Luigi Ribotta et al 2024 Meas. Sci. Technol. 35 105014
Adam Thompson et al 2021 Meas. Sci. Technol. 32 105013
Maximum permissible errors (MPEs) are an important measurement system specification and form the basis of periodic verification of a measurement system's performance. However, there is no standard methodology for determining MPEs, so when they are not provided, or not suitable for the measurement procedure performed, it is unclear how to generate an appropriate value with which to verify the system. Whilst a simple approach might be to take many measurements of a calibrated artefact and then use the maximum observed error as the MPE, this method requires a large number of repeat measurements for high confidence in the calculated MPE. Here, we present a statistical method of MPE determination, capable of providing MPEs with high confidence and minimum data collection. The method is presented with 1000 synthetic experiments and is shown to determine an overestimated MPE within 10% of an analytically true value in 99.2% of experiments, while underestimating the MPE with respect to the analytically true value in 0.8% of experiments (overestimating the value, on average, by 1.24%). The method is then applied to a real test case (probing form error for a commercial fringe projection system), where the efficiently determined MPE is overestimated by 0.3% with respect to an MPE determined using an arbitrarily chosen large number of measurements.
Ahmad Satya Wicaksana et al 2024 Meas. Sci. Technol. 35 095016
Siqi Gong et al 2024 Meas. Sci. Technol. 35 106128
Data-driven intelligent fault diagnosis methods generally require a large amount of labeled data and considerable time to train network models. However, obtaining sufficient labeled data in practical industrial scenarios has always been a challenge, which hinders the practical application of data-driven methods. A digital twin (DT) model of rolling bearings can generate labeled training dataset for various bearing faults, supplementing the limited measured data. This paper proposes a novel DT-assisted approach to address the issue of limited measured data for bearing fault diagnosis. First, a dynamic model of bearing with damages is introduced to generate simulated bearing acceleration vibration signals. A DT model is constructed in Simulink, where the model parameters are updated based on the actual system behavior. Second, the structural parameters of the DT model are adaptively updated using least squares method with the measured data. Third, a Vision Transformer (ViT) -based network, integrated with short-time Fourier transform, is developed to achieve accurate fault diagnosis. By applying short-time Fourier transform at the input end of the ViT network, the model effectively extracts additional information from the vibration signals. Pre-training the network with an extensive dataset from miscellaneous tasks enables the acquisition of pre-trained weights, which are subsequently transferred to the bearing fault diagnosis task. Experiments results verify that the proposed approach can achieve higher diagnostic accuracy and better stability.
Gustavo Quino et al 2021 Meas. Sci. Technol. 32 015203
Digital image correlation (DIC) is a widely used technique in experimental mechanics for full field measurement of displacements and strains. The subset matching based DIC requires surfaces containing a random pattern. Even though there are several techniques to create random speckle patterns, their applicability is still limited. For instance, traditional methods such as airbrush painting are not suitable in the following challenging scenarios: (i) when time available to produce the speckle pattern is limited and (ii) when dynamic loading conditions trigger peeling of the pattern. The development and application of some novel techniques to address these situations is presented in this paper. The developed techniques make use of commercially available materials such as temporary tattoo paper, adhesives and stamp kits. The presented techniques are shown to be quick, repeatable, consistent and stable even under impact loads and large deformations. Additionally, they offer the possibility to optimise and customise the speckle pattern. The speckling techniques presented in the paper are also versatile and can be quickly applied in a variety of materials.
A Sciacchitano 2019 Meas. Sci. Technol. 30 092001
Particle image velocimetry (PIV) has become the chief experimental technique for velocity field measurements in fluid flows. The technique yields quantitative visualizations of the instantaneous flow patterns, which are typically used to support the development of phenomenological models for complex flows or for validation of numerical simulations. However, due to the complex relationship between measurement errors and experimental parameters, the quantification of the PIV uncertainty is far from being a trivial task and has often relied upon subjective considerations. Recognizing the importance of methodologies for the objective and reliable uncertainty quantification (UQ) of experimental data, several PIV-UQ approaches have been proposed in recent years that aim at the determination of objective uncertainty bounds in PIV measurements.
This topical review on PIV uncertainty quantification aims to provide the reader with an overview of error sources in PIV measurements and to inform them of the most up-to-date approaches for PIV uncertainty quantification and propagation. The paper first introduces the general definitions and classifications of measurement errors and uncertainties, following the guidelines of the International Organization for Standards (ISO) and of renowned books on the topic. Details on the main PIV error sources are given, considering the entire measurement chain from timing and synchronization of the data acquisition system, to illumination, mechanical properties of the tracer particles, imaging of those, analysis of the particle motion, data validation and reduction. The focus is on planar PIV experiments for the measurement of two- or three-component velocity fields.
Approaches for the quantification of the uncertainty of PIV data are discussed. Those are divided into a-priori UQ approaches , which provide a general figure for the uncertainty of PIV measurements, and a-posteriori UQ approaches , which are data-based and aim at quantifying the uncertainty of specific sets of data. The findings of a-priori PIV-UQ based on theoretical modelling of the measurement chain as well as on numerical or experimental assessments are discussed. The most up-to-date approaches for a-posteriori PIV-UQ are introduced, highlighting their capabilities and limitations.
As many PIV experiments aim at determining flow properties derived from the velocity fields (e.g. vorticity, time-average velocity, Reynolds stresses, pressure), the topic of PIV uncertainty propagation is tackled considering the recent investigations based on Taylor series and Monte Carlo methods. Finally, the uncertainty quantification of 3D velocity measurements by volumetric approaches (tomographic PIV and Lagrangian particle tracking) is discussed.
Martin Kögler and Bryan Heilala 2020 Meas. Sci. Technol. 32 012002
Time-gated (TG) Raman spectroscopy (RS) has been shown to be an effective technical solution for the major problem whereby sample-induced fluorescence masks the Raman signal during spectral detection. Technical methods of fluorescence rejection have come a long way since the early implementations of large and expensive laboratory equipment, such as the optical Kerr gate. Today, more affordable small sized options are available. These improvements are largely due to advances in the production of spectroscopic and electronic components, leading to the reduction of device complexity and costs. An integral part of TG Raman spectroscopy is the temporally precise synchronization (picosecond range) between the pulsed laser excitation source and the sensitive and fast detector. The detector is able to collect the Raman signal during the short laser pulses, while fluorescence emission, which has a longer delay, is rejected during the detector dead-time. TG Raman is also resistant against ambient light as well as thermal emissions, due to its short measurement duty cycle.
In recent years, the focus in the study of ultra-sensitive and fast detectors has been on gated and intensified charge coupled devices (ICCDs), or on CMOS single-photon avalanche diode (SPAD) arrays, which are also suitable for performing TG RS. SPAD arrays have the advantage of being even more sensitive, with better temporal resolution compared to gated CCDs, and without the requirement for excessive detector cooling. This review aims to provide an overview of TG Raman from early to recent developments, its applications and extensions.
Meixuan Su et al 2024 Meas. Sci. Technol. 35 115305
As an efficient and environment-friendly method, electrostatic separation has gradually replaced flotation methods in the separation of magnesite in recent years. In the process of triboelectrostatic separation, the mineral particles are tribocharged driven by the air flow, then the trajectory is shifted under the action of the electric field, so as to realize the separation. The useful mineral in magnesite is MgCO 3 , but the theoretical research related to the charge characteristics of MgCO 3 is not sufficient. Particle image velocimetry (PIV), as an indirect measurement technique, is able to obtain the velocity field of the fluids from images. However, the particles moving in the air have the issues such as excessive speed and small particle size, which make the traditional PIV has low accuracy in estimating the motion of particles. In this paper, a high-speed camera is used to capture the motion trajectory of tribocharged MgCO 3 particles in a parallel electric field. A new optical flow method LFN-en-A network based on LiteFlowNet-en network is proposed to compute the particle motion trajectory by combining the deep learning method with the traditional PIV, which realizes the displacement estimation of particles moving in the air. It ultimately realizes the calculation of the charge-to-mass ratio on single particles. Analyzing the accuracy of the LFN-en-A network's estimation in the experiments, the estimation of LiteFlowNet-en was compared. Changing the shooting frame rate analyzes the optimal one required by the LFN-en-A network. Combining the estimation results of LFN-en-A to calculate the particle charge-to-mass ratio ( Q/m ), the Q / m of MgCO 3 particle was analyzed by changing the experimental conditions in the process of particles' tribocharging, which provided a new method for particle-to-charge ratio measurement.
Yan Zhang et al 2024 Meas. Sci. Technol. 35 116202
Batch processes play an important role in modern chemical industrial and manufacturing production, while the control of product quality relies largely on online quality prediction. However, the complex nonlinearity of batch process and the dispersion of quality-related features may affect the quality prediction performance. In this paper, a deep quality-related stacked isomorphic autoencoder for batch process quality prediction is proposed. Firstly, the raw input data are reconstructed layer-by-layer by isomorphic autoencoder and the raw data features are obtained. Secondly, the quality-related information is enhanced by analyzing the correlation between the isomorphic feature of each layer of the network and the output target, and constructing a correlation loss function. Thirdly, a deep quality-related prediction model is constructed to predict the batch process quality variables. Finally, experimental validation was carried out in penicillin fermentation simulation platform and strip hot rolling process, and the experimental results demonstrated the feasibility and effectiveness of the model proposed in this paper for the quality prediction of the batch process.
Huachuan Zhao et al 2024 Meas. Sci. Technol. 35 116306
During ship operations at sea, the vessel's attitude undergoes continuous changes due to various factors such as wind, waves, and its own motion. These influences are challenging to mathematically describe, and the changes in attitude are also influenced by multiple interconnected factors. Consequently, accurately predicting the ship's attitude presents significant challenges. Previous studies have demonstrated that phenomena like wind speed and wave patterns exhibit chaotic characteristics when affecting attitude changes. However, research on predicting ship attitudes lacks an exploration of whether chaotic characteristics exist and how they can be described and applied. This paper initially identifies the chaotic characteristics of ship attitude data through phase space reconstruction analysis and provides mathematical representations for them. Based on these identified chaotic characteristics, a Transformer model incorporating feature embedding layers is employed for time series prediction. Finally, a comparison with traditional methods validates the superiority of our proposed approach.
Libing Du et al 2024 Meas. Sci. Technol. 35 115602
Particle morphology is an important factor affecting the mechanical properties of granular materials. However, it is difficult to quantify the morphology characteristics of the complex concave particle. Fortunately, complex particle can be segmented by convex decomposition, so a new shape index named convex decomposition coefficient (CDC) related to the number of segmentations is proposed. First, the pocket concavity was introduced to simplify the morphology hierarchically. Second, the cut weight linked to concavity was defined and convex decomposition was linearly optimised by maximizing the total cut weights. Third, the CDC was defined as the minimum block number where the block area ratio cumulatively exceeded 0.9 in descending order. Finally, the proposed index was used to quantify the particle morphology of coral sand. The results demonstrate that the CDC of coral sands mainly ranges from 2 to 6, with a positively skewed distribution. Furthermore, CDC correlates well with three shape indices: sphericity, particle size, and convexity. Larger CDC is associated with smaller sphericity, larger particle size, and smaller convexity. The index has certain scientific research value and practical significance.
Zhenfa Shao et al 2024 Meas. Sci. Technol. 35 116111
In practical scenarios, gearbox fault diagnosis faces the challenge of extremely scarce labeled data. Additionally, variations in operating conditions and differences in sensor installations exacerbate data distribution shifts, significantly increasing the difficulty of fault diagnosis. To address the above issues, this paper proposes a wavelet dynamic joint self-adaptive network guided by a pseudo-label alignment mechanism (MDJSN-DFL). First, the wavelet-efficient convolution module is designed based on wavelet convolution and efficient attention mechanisms. This module is used to construct a multi-wavelet convolution feature extractor to extract critical fault features at multiple levels. Secondly, to improve the classifier's discriminability in the target domain, a transitional clustering-guided DFL is developed. This mechanism can capture fuzzy classification samples and improve the pseudo-label quality of the target domain. Finally, a dynamic joint mean square difference algorithm (DJSD) is proposed, which is composed of joint maximum mean square discrepancy and joint maximum mean discrepancy. The algorithm can adaptively adjust according to the dynamic balance factor to minimize the domain distribution discrepancy. Experiments on two different gearbox datasets show that MDJSN-DFL performs better in diagnostic scenarios under varying load conditions and different sensor installation setups, validating the proposed method's effectiveness and superiority.
Jiashuai Huang et al 2024 Meas. Sci. Technol. 35 112001
With the continuous development of the aerospace, defense, and military industry, along with other high-end fields, the complexity of machined parts has gradually increased. Consequently, the demand for tool intelligence has also strengthened. However, traditional tools are prone to wear during cutting due to high cutting forces, high temperatures, and vibrations. Intelligent tools, in contrast to traditional ones, integrate sensors into their design, allowing for real-time monitoring of the cutting status and timely prediction of tool wear. The application of intelligent tools in machining significantly enhances machining quality, increases productivity, and reduces production costs. In this review, first, the tool wear monitoring methods were classified and discussed. Second, the intelligence and innovation of sensors in monitoring cutting force, temperature, and vibration were introduced, and the commonly used types of sensors for online monitoring of cutting force were detailed. Furthermore, different types of sensors in tool wear were discussed, and the advantages of multi-sensor monitoring were summarized. Some urgent issues and perspectives that need to be addressed were proposed, providing new ideas for the design and development of intelligent tools.
Dang Tuyet Minh and Nguyen Ba Dung 2024 Meas. Sci. Technol. 35 112002
Path planning for unmanned aerial vehicle (UAV) is the process of determining the path that travels through each location of interest within a particular area. There are numerous algorithms proposed and described in the publications to address UAV path planning problems. However, in order to handle the complex and dynamic environment with different obstacles, it is critical to utilize the proper fusion algorithms in planning the UAV path. This paper reviews some hybrid algorithms used in finding the optimal route of UAVs that developed in the last ten years as well as their advantages and disadvantages. The UAV path planning methods were classified into categories of hybrid algorithms based on traditional, heuristic, machine learning approaches. Criteria used to evaluate algorithms include execution time, total cost, energy consumption, robustness, data, computation, obstacle avoidance, and environment. The results of this study provide reference resources for researchers in finding the path for UAVs.
Qi Wang et al 2024 Meas. Sci. Technol. 35 102001
With the booming development of modern industrial technology, rotating machinery fault diagnosis is of great significance to improve the safety, efficiency and sustainable development of industrial production. Machine learning as an effective solution for fault identification, has advantages over traditional fault diagnosis solutions in processing complex data, achieving automation and intelligence, adapting to different fault types, and continuously optimizing. It has high application value and broad development prospects in the field of fault diagnosis of rotating machinery. Therefore, this article reviews machine learning and its applications in intelligent fault diagnosis technology and covers advanced topics in emerging deep learning techniques and optimization methods. Firstly, this article briefly introduces the theories of several main machine learning methods, including Extreme Learning Machines (ELM), Support Vector Machines (SVM), Convolutional Neural Networks (CNNs), Deep Belief Networks (DBNs) and related emerging deep learning technologies such as Transformer, adversarial neural network (GAN) and graph neural network (GNN) in recent years. The optimization techniques for diagnosing faults in rotating machinery are subsequently investigated. Then, a brief introduction is given to the papers on the application of these machine learning methods in the field of rotating machinery fault diagnosis, and the application characteristics of various methods are summarized. Finally, this survey discusses the problems to be solved by machine learning in fault diagnosis of rotating machinery and proposes an outlook.
Liuyang Song et al 2024 Meas. Sci. Technol. 35 092003
This paper presents a comprehensive review of the state-of-the-art techniques for predicting the remaining useful life (RUL) of rolling bearings. Four key aspects of bearing RUL prediction are considered: data acquiring, construction of health indicators, development of RUL prediction algorithms, and evaluation of prediction results. Additionally, publicly available datasets that can be used to validate bearing prediction algorithms are described. The existing RUL prediction algorithms are categorized into three types and have been comprehensively reviewed: physical-based, statistical-based, and data-driven. In particular, the progress made in data-driven prediction methods is summarized, and typical methods such as rerrent neural network, convolutional network, graph convolutional network, Transformer, and transfer learning-based methods are introduced in detail. Finally, the challenges faced by data-driven methods in RUL prediction for bearings are discussed.
Mohanraj T et al 2024 Meas. Sci. Technol. 35 092002
Milling is an extremely adaptable process that can be utilized to fabricate a wide range of shapes and intricate 3D geometries. The versatility of the milling process renders it useful for the production of a diverse range of components and products in several industries, including aerospace, automotive, electronics, and medical equipment. Monitoring tool conditions is essential for maintaining product quality, minimizing production downtime, and maximizing tool life. Advances in this field have been driven by the need for increased productivity, reduced tool wear, and improved process efficiency. Tool condition monitoring (TCM) in the milling process is a critical aspect of machining operations. TCM involves assessing the health and performance of cutting tools used in milling machines. As technology evolves, staying updated with the latest developments in this field is essential for manufacturers seeking to optimize their milling operations. However, addressing the challenges associated with sensor integration, data analysis, and cost-effectiveness remains crucial. To fill this research gap, this paper provides an overview of the extensive literature on monitoring milling tool conditions. It summarizes the key focus areas, including tool wear sensors and the application of various machine learning and deep learning algorithms. It also discusses the potential applications of TCM beyond wear detection, such as predicting tool breakage, tool wear, the cutting tool's remaining lifetime, and the challenges faced by TCMs. This review also provides suggestions for potential future research endeavors and is anticipated to offer valuable insights for the development of advanced TCMs in terms of tool wear monitoring and predicting remaining useful life.
Gu et al
Damage to the composite propeller blades could lead to rotational imbalance, which seriously affects the operational safety of unmanned aerial vehicles (UAVs), therefore, a novel method combining the Teager energy operator and bidirectional temporal convolutional network is proposed for detecting, localizing, and quantifying the damage-related imbalance in the blades. A flexible sensing system that contains MEMS accelerometers, signal conditioning, and wireless transmission is integrated with the composite propeller for in-situ signal acquisition of the propeller blades. Teager energy operator (TEO) is applied to demodulate and enhance the pulse compositions in vibration signals and singular value decomposition (SVD) is employed to suppress random noise, resulting in denoised Teager energy spectrums for model input. Temporal convolutional network (TCN) has been widely used in sequence signal modeling because the causal dilated convolution could learn the context information of sequence signals while maintaining the advantages of parallel computing. To fully extract the signal features, bidirectional temporal convolutional network (BiTCN) models are established to learn both the forward and backward signal features. Experimental verification results show that the proposed method detects the existence of imbalance with 100% accuracy, and the accuracies of localization and quantization are 99.65% and 98.61%, respectively, which are much higher than those of the models with the original signal as input. In addition, compared with the other four different algorithms, BiTCN is superior in terms of convergence speed and prediction accuracy.
Zhao et al
AbstractFlow visualization in harsh environments such as a scramjet combustor featuring highly turbulentsupersonic reactive flow with intensive luminescence emission is chanllenging and typically lack ofsufficient spatiotemporal resolution that is essential for resolving flow dynamics. This study presents adevelopment of a robust flow visualization technique with an exceptional spatiotemporal resolution in ascramjet combustor. By utilizing a customized LED light source, the short pulse duration along with ahigh peak power and a single-color emission ensures an instantaneous exposure with little backgroundluminescence interference. Focusing schlieren image measurements with mitigated path-integrationeffect are successfully demonstrated in a scramjet engine combustor at instant, frame-straddling, andsequential temporal resolutions of 100 ns, 500 ns, and 26 µs, respectively, along with a megapixelimaging resolution. Consequently, in addition to flow visualization, it is worth highlighting that theexceptional spatiotemporal correlation resolved by present measurements exhibits attractive potentialsof velocimetry for harsh high-speed flow environments.
Zhou et al
Laser absorption tomography (LAT) has been widely employed to capture two/three-dimensional reactive flow-field parameters with a penetrating spatiotemporal resolution. In industrial environments, LAT is generally implemented by measuring multiple, e.g. 30 to more than 100, wavelength modulated laser transmissions at high imaging rates, e.g. tens to thousands of frames per second (fps). A short-period LAT experiment can generate extensive load of data, which require massive computational source and time for data post-processing. In this work, a large-scale data processing platform is designed for industrial LAT. The platform significantly speeds up LAT signal processing by introducing a parallel computing architecture. By identifying the discrepancy between the measured and theoretical spectra, the new platform enables indexing of the laser-beam measurements that are disturbed by harsh-environment noise. Such a scheme facilitates effective removal of noise-distorted beams, which can lead to artefacts in the reconstructed images. The designed platform is validated by a lab-based LAT experiment, which is implemented by processing the laser transmissions of a 32-beam LAT sensor working at 250 fps. To process a 60-second LAT experimental dataset, the parallelism enabled by the platform saves computational time by 40.12% compared to the traditional single-thread approach. The error-detection scheme enables the successful accurate identification of noise-distorted measurements, i.e. 0.59% of overall laser-beam measurements that fall out of the physical model.
Lu et al
Arresters are one of the critical components of the power system.
However, due to the arrester's regular and uniform umbrella skirt, both traditional
manual detection methods and existing computer vision approaches exhibit limitations
in accuracy and efficiency. This paper proposes an automatic, robust, efficient arrester
point cloud registration method to address this problem. First, a robotic arm
maneuvers a depth camera to capture point cloud data from various perspectives.
Then, the fast global registration (FGR) point cloud coarse registration method
based on the signature of histograms of orientations (SHOT) descriptor to produce
preliminary registration results. This result is ultimately used as the initial value of
the improved iterative closest point (ICP) algorithm to refine the registration further.
Experimental results on various data sets collected from arrester and public data sets
show that the algorithm's root mean square error(RMSE) is less than 0.1mm, meeting
the requirements of the engineering application of arrester detection.
Xu et al
Because of the "soft-field" effect and ill-posed and ill-conditional inverse problem, it is difficult to obtain high quality images from an electrical capacitance tomography (ECT) system. To achieve high quality images and fast imaging speed with limited measurement data, an image reconstruction algorithm, which was initially proposed for compressive sensing, is modified for reconstructing ECT images. In this proposed algorithm, deep networks were inspired by the iterative shrinkage-thresholding algorithm (ISTA) to achieve a mathematically interpretable model with learnable parameters. On this basis, the traditional Landweber iteration is combined with ISTA-Net to optimize ECT image reconstruction. The training and test process is driven by the dynamic simulation coupling the gas-oil two-phase flow field and ECT electrostatic field. Test results demonstrate that this algorithm is superior to the traditional image reconstruction algorithms for ECT. Compared with the LBP algorithm, the averaged image error and gas fraction error have dropped 20.44% and 16.74% respectively, while the computational speed is similar to Landweber iteration. The reconstruction accuracy of two-phase interface and gas fraction in this new algorithm has been validated by static experimental test, showing that it is promising to be applied in real gas-oil two phase flow measurement.
More Accepted manuscripts
Kenneth M Peterson et al 2024 Meas. Sci. Technol. 35 115601
Surface characteristics are a major contributor to the in-service performance, particularly fatigue life, of additively manufactured (AM) components. Centrifugal disk finishing (CDF) is one of many rigid media, abrasive machining processes employed to smooth the surfaces and edges of AM components. Within the general family of abrasive machining processes currently applied to AM, CDF is moderate in terms of material removal rate and the inertial forces exerted. How CDF alters the underlying microstructure of the processed surface is currently unknown. Here, white light profilometry and high-energy x-ray diffraction are employed to characterize surface finish, crystallographic texture, and anisotropic distributions of residual microscale strain as a function of depth in CDF-finished Inconel 718 manufactured with laser powder bed fusion. Surfaces are finished using both unimodal and bimodal finishing media size distributions. The CDF processes employed are found to remove surface crystallographic textures (here a {111} fiber texture) from AM components, but generally not alter the bulk texture (here a cube texture). CDF is also found to impart significant amounts of residual microscale strain into the first 100 μm from the sample surface.
Eberhard Manske et al 2024 Meas. Sci. Technol. 35 110201
Minqiu Zhou et al 2024 Meas. Sci. Technol.
Dustin Witkowski et al 2024 Meas. Sci. Technol.
Lifetime-based phosphor thermometry has been applied in a wide variety of surface temperature measurement applications due to its relative ease of implementation and robustness to background interference when compared to other optical temperature measurement methods. It is often assumed that the technique is minimally intrusive if the thickness of the applied phosphor coating is < 20 µm. To evaluate this assumption, high-speed phosphor surface temperature and thermocouple measurements were performed on 4140 steel substrates installed in the cylinder head of an optically-accessible internal-combustion engine for four operating conditions. For phosphor thermometry measurements, four substrates were studied, each coated with a phosphor layer of different thickness ranging between 6 µm and 47 µm. The phosphor thermometry measured temperature swings during combustion were shown to be heavily impacted by the presence of the phosphor coatings, increasing by roughly a factor of 2 - 2.5 when increasing the thickness from 6 µm to 30 µm. A technique was implemented which utilizes the temperature data in combination with a heat conduction model to provide estimates of the temperature swing and heat transfer flux in the absence of the phosphor coating. It was shown that even the 6 µm phosphor coating could lead to an order of magnitude increase in the temperature swing relative to the uncoated 4140 steel substrate. Despite the intrusiveness of phosphor thermometry for the surface temperature measurements, reasonable agreement was demonstrated between heat flux estimates determined with the heat transfer modeling technique and those deduced from temperature-swing measurements using two different high-speed thermocouples. The results indicate that phosphor surface thermometry can be a reliable surface temperature and heat flux diagnostic for transient high heat flux environments, as long as proper care is taken to account for the impact of the phosphor layer on the measurement.
Ryan Thomas et al 2024 Meas. Sci. Technol. 35 116106
This paper presents machine learning classification on simulated data of permeable conducting spheres in air and seawater irradiated by low frequency electromagnetic pulses. Classification accuracy greater than 90% was achieved. The simulated data were generated using an analytical model of a magnetic dipole in air and seawater placed 1.5–3.5 m above the center of the sphere in 50 cm increments. The spheres had radii of 40 cm and 50 cm and were of permeable materials, such as steel, and non-permeable materials, such as aluminum. A series RL circuit was analytically modeled as the transmitter coil, and an RLC circuit as the receiver coil. Additive white Gaussian noise was added to the simulated data to test the robustness of the machine learning algorithms to noise. Multiple machine learning algorithms were used for classification including a perceptron and multiclass logistic regression, which are linear models, and a neural network, 1D convolutional neural network (CNN), and 2D CNN, which are nonlinear models. Feature maps are plotted for the CNNs and provide explainability of the salient parts of the time signature and spectrogram data used for classification. The pulses investigated, which expand the literature, include a two-sided decaying exponential, Heaviside step-off, triangular, Gaussian, rectangular, modulated Gaussian, raised cosine, and rectangular down-chirp. Propagation effects, including dispersion and frequency dependent attenuation, are encapsulated by the analytical model, which was verified using finite element modeling. The results in this paper show that machine learning methods are a viable alternative to inversion of electromagnetic induction (EMI) data for metallic sphere classification, with the advantage of real-time classification without the use of a physics-based model. The nonlinear machine learning algorithms used in this work were able to accurately classify metallic spheres in seawater even with significant pulse distortion caused by dispersion and frequency dependent attenuation. This paper presents the first effort towards the use of machine learning to classify metallic objects in seawater based on EMI sensing.
Atharva Hans et al 2024 Meas. Sci. Technol. 35 116002
Measuring particles' three-dimensional (3D) positions using multi-camera images in fluid dynamics is critical for resolving spatiotemporally complex flows like turbulence and mixing. However, current methods are prone to errors due to camera noise, optical configuration and experimental setup limitations, and high seeding density, which compound to create fake measurements (ghost particles) and add noise and error to velocity estimations. We introduce a Bayesian volumetric reconstruction (BVR) method, addressing these challenges by using probability theory to estimate uncertainties in particle position predictions. Our method assumes a uniform distribution of particles within the reconstruction volume and employs a model mapping particle positions to observed camera images. We utilize variational inference with a modified loss function to determine the posterior distribution over particle positions. Key features include a penalty term to reduce ghost particles, provision of uncertainty bounds, and scalability through subsampling. In tests with synthetic data and four cameras, BVR achieved 95% accuracy with less than 3% ghost particles and an RMS error under 0.3 pixels at a density of 0.1 particles per pixel. In an experimental Poiseuille flow measurement, our method closely matched the theoretical solution. Additionally, in a complex cerebral aneurysm basilar tip geometry flow experiment, our reconstructions were dense and consistent with observed flow patterns. Our BVR method effectively reconstructs particle positions in complex 3D flows, particularly in situations with high particle image densities and camera distortions. It distinguishes itself by providing quantifiable uncertainty estimates and scaling efficiently for larger image dimensions, making it applicable across a range of fluid flow scenarios.
Frank J van Kann and Alexey V Veryaskin 2024 Meas. Sci. Technol. 35 115101
A novel room temperature capacitive sensor interface circuit is proposed and successfully tested, which uses a modified all-pass filter (APF) architecture combined with a simple series resonant tank circuit with a moderate Q -factor. It is fashioned from a discrete inductor with small dissipation resonating with a grounded capacitor acting as the sensing element to obtain a resolution of Δ C ∼ 2 zF in a capacitance range of 10–30 pF. The circuit converts the change in capacitance to the change in the phase of a carrier signal in a frequency range with a central frequency set up by the tank circuit's resonant frequency and is configured to act as a close approximation of the ideal APF. This cancels out the effects of amplitude modulation when the carrier signal is imperfectly tuned to the resonance. The proposed capacitive sensor interface has been specifically developed for use as a front-end constituent in ultra-precision mechanical displacement measurement systems, such as accelerometers, seismometers, gravimeters and gravity gradiometers, where moving plate grounded air gap capacitors are frequently used. Some other applications of the proposed circuit are possible including the measurement of the electric field, where the sensing capacitor depends on the applied electric field, and cost effective capacitive gas sensors. In addition, the circuit can be easily adapted to function with very small capacitance values (1–2 pF) as is typical in MEMS-based transducers.
K F A Jorissen et al 2024 Meas. Sci. Technol. 35 115501
We present the study of millisecond-resolved polymer brush swelling dynamics using infrared spectroscopy with a home-built quantum cascade laser-based infrared spectrometer at a 1 kHz sampling rate after averaging. By cycling the humidity of the environment of the polymer brush, we are able to measure the swelling dynamics sequentially at different wavenumbers. The high sampling rate provides us with information on the reconformation of the brush at a higher temporal resolution than previously reported. Using spectroscopic ellipsometry, we study the brush swelling dynamics as a reference experiment and to correct artefacts of the infrared measurement approach. This technique informs on the changes in the brush thickness and refractive index. Our results indicate that the swelling dynamics of the polymer brush are poorly described by Fickian diffusion, pointing toward more complicated underlying transport.
Jing Guo et al 2024 Meas. Sci. Technol. 35 105119
Electrical impedance tomography (EIT) has become an integral component in the repertoire of medical imaging techniques, particularly due to its non-invasive nature and real-time imaging capabilities. Despite its potential, the application of EIT in minimally invasive surgery (MIS) has been hindered by a lack of specialized electrode probes. Existing designs often compromise between invasiveness and spatial sensitivity: probes small enough for MIS often fail to provide detailed imaging, while those offering greater sensitivity are impractically large for use through a surgical trocar. Addressing this challenge, our study presents a breakthrough in EIT probe design. The open electrode probe we have developed features a line of 16 electrodes, thoughtfully arrayed to balance the spatial demands of MIS with the need for precise imaging. Employing an advanced EIT reconstruction algorithm, our probe not only captures images that reflect the electrical characteristics of the tissues but also ensures the homogeneity of the test material is accurately represented. The versatility of our probe is demonstrated by its capacity to generate high-resolution images of subsurface anatomical structures, a feature particularly valuable during MIS where direct visual access is limited. Surgeons can rely on intraoperative EIT imaging to inform their navigation of complex anatomical landscapes, enhancing both the safety and efficacy of their procedures. Through rigorous experimental validation using ex vivo tissue phantoms, we have established the probe's proficiency. The experiments confirmed the system's high sensitivity and precision, particularly in the critical tasks of subsurface tissue detection and surgical margin delineation. These capabilities manifest the potential of our probe to revolutionize the field of surgical imaging, providing a previously unattainable level of detail and assurance in MIS procedures.
Antonella D'Alessandro et al 2024 Meas. Sci. Technol. 35 105116
Civil constructions significantly contribute to greenhouse gas emissions and entail extensive energy and resource consumption, leading to a substantial ecological footprint. Research into eco-friendly engineering solutions is therefore currently imperative, particularly to mitigate the impact of concrete technology. Among potential alternatives, shot-earth-concrete, which combines cement and earth as a binder matrix and is applied via spraying, emerges as a promising option. Furthermore, this composite material allows for the incorporation of nano and micro-fillers, thereby providing room for enhancing mechanical properties and providing multifunctional capabilities. This paper investigates the damage detection capabilities of a novel smart shot-earth concrete with carbon microfibers, by investigating the strain sensing performance of a full-scale vault with a span of 4 m, mechanically tested until failure. The material's strain and damage sensing capabilities involve its capacity to produce an electrical response (manifested as a relative change in resistance) corresponding to the applied strain in its uncracked state, as well as to exhibit a significant alteration in electrical resistance upon cracking. A detailed multiphysics numerical (i.e. mechanical and electrical) model is also developed to aid the interpretation of the experimental results. The experimental test was conducted by the application of an increasing vertical load at a quarter of the span, while modelling of the element was carried out by considering a piezoresistive material, with coupled mechanical and electrical constitutive properties, including a new law to reproduce the degradation of the electrical conductivity with tensile cracking. Another notable aspect of the simulation was the consideration of the effects of the electrical conduction through the rebars, which was found critical to accurately reproduce the full-scale electromechanical response of the vault. By correlating the outcomes from external displacement transducers with the self-monitoring features inherent in the proposed material, significant insights were gleaned. The findings indicated that the proposed smart-earth composite, besides being well suited for structural applications, also exhibits a distinctive electromechanical behavior that enables the early detection of damage initiation. The results of the paper represent an important step toward the real application of smart earth-concrete in the construction field, demonstrating the effectiveness and feasibility of full-scale strain and damage monitoring even in the presence of steel reinforcement.
More Open Access articles
Measurement is the process of assigning numbers or values to quantities or attributes of objects or events according to certain rules or standards. It is a fundamental aspect of science, mathematics, and everyday life, allowing us to quantify and compare various aspects of the physical world.
Table of Content
In simple words, measurement means using a yardstick to determine the characteristics of a physical object. In addition to physical objects, qualitative concepts, such as songs and paintings, or an abstract phenomenon, can also be measured. However, measuring qualitative concepts is a comparatively difficult task because numbers cannot be easily assigned to them. For example, it is easy to state that the weight of an object or a subject is 10 kg.
However, if a person is asked to measure a song for its good composition, then it becomes difficult to say that the song is 10 per cent good or so. Today, there exist standardised tools to measure abstract concepts such as intelligence, unity, honesty, bravery, success and stress. High accuracy and confidence can be expected while measuring quantitative characteristics of an object.
A measurement scale refers to a classification that defines the nature of information within the numerals assigned to variables. Measurement scales have been classified into four types as shown in Figure:
Let us now discuss the types of measurement scales in detail.
In this scale, the variables are named or labelled in no specific order. This is the measurement scale in which numbers are assigned to things, beings or events to classify or identify or label them. All of the nominal scales are mutually exclusive and bear no numerical significance. For example, the assignment of different numbers (1, 2, 3, …) to cricket players in a team, books in a library, and computers in the Internet café. These numbers cannot be used to perform mathematical operations.
If 11 players of a cricket team are assigned numbers from 1 to 11, finding the average of 1 to 11 does not signify any meaning. In this case, the only use of these numbers is to count the team members. The nominal scale represents the lowest level of measurement. However, it is helpful when there is a need to classify data. For example, for a question “What are your political views?”, we can have the following nominal scale:
S. No. | Objective |
---|---|
1 | Left Orientation |
2 | Right Orientation |
3 | Centre |
4 | Conservative |
This is the scale that only implies greater than or less than but does not answer how much greater or less. Only inequalities can be set up with respect to ordinal scale and other arithmetic operations cannot be performed. The ordinal scale can be used to make only comparisons. In ordinal scale, data is shown in order of magnitude. An example of the ordinal scale is as follows:
For example, the ordering of colour preferences of Mr. A are:
1 | Silver |
2 | White |
3 | Black |
4 | Red |
Thus, with the ordinal scale, researcher can use median or mode to determine the central tendency of a set of ordinal data.
This is the scale in which the interval between successive positions is equal. The positions are separated by equally spaced intervals or basis. For example, a person represents his/her level of happiness along a scale rated from 1 to 10.
With the interval scale, the following conclusions can be made:
The basic limitation of the interval scale is that it does not contain an absolute zero. A simple example of the interval scale is the scale of temperature in which absolute zero is unattainable theoretically. Therefore, the interval scale does not have the provision to measure the absence of any characteristic such as zero happiness (or absence of happiness). The interval scale contains features of nominal and ordinal scales. In addition, it involves the concept of equality of intervals. In the interval scale, more arithmetic operations, such as addition and subtraction can be performed. Mean can be calculated for interval scale.
This is the scale that contains absolute or true zero, which implies the absence of any trait. For example, on a centimetre scale, zero implies the absence of length or height. In the ratio scale, it is possible to take ratio of two observations. For example, it can be stated that the weight of Ram is twice that of Shyam. The ratio scale is the most powerful scale of measurement, as almost all statistical operations, which cannot be performed by other scales, can be performed by it.
Any tool used to measure or collect data is called measurement tool, which is also known as assessment tool. There are several types of measurement tools, such as observations, scales, questionnaires, surveys, interviews and indexes (or indices). There are four stages of developing measurement tools, which are explained as follows:
This is the first stage in the process of development of measuring tools. At this stage, the researcher develops a good understanding of the topic he/she wants related to his research study. For example, research on the pros and cons of the multiparty political system requires a proper understanding of the concept behind this system.
Without referring to the theories related to the multi-party system, a good understanding about this system cannot be developed. However, if the research is being done on a concept such as stress (which has already been researched extensively), no exclusive concept building is required.
After developing a concept at the first stage, the researcher is required to clearly identify the dimensions of the concept. For example, when a researcher wants to conduct a study related to image of a company, he/she may relate image of the company to dimensions or factors such as customer service level, customer treatment, product quality, employee treatment, social responsibility, corporate leadership, etc.
At this stage, indicators for the research subject are selected. Indicators help in measuring the elements of a concept such as knowledge, opinion, choices, expectations and feelings of respondents. Examples of indicators are variables.
For example, the effectiveness of a medicine (concept) used for treating a chronic disease is related to indicators such as changes in the mortality rate, recurrence of that disease, etc. The researcher may convert these indicators into variables that can be measured. For instance, number of deaths caused by that disease, number of patients who were again affected by that disease, can be the indicators of the said concept.
After determining multiple elements of a particular concept and selecting suitable indicators for the research, the researcher needs to combine all the indicators into a summated scale because separate indicators cannot give a certain measurement of the concept. For example, a price index is based on the weighted sum of prices.
What should be the characteristics of a good measurement tool? The answer to this question is that the tool should clearly and accurately indicate what the researcher intends to measure. In addition, the good measurement tool should be easy to use and should give reliable results.
The two major characteristics of a good measurement tool are explained as follows:
Reliability of a good measurement tool refers to the degree of confidence with which the measurement tool can be used to derive consistent results upon repeated application. A reliable instrument is not necessarily a valid instrument. However, a valid instrument must be reliable.
For example, a scale is used to measure the weight of objects. The scale consistently shows all objects to be overweight by 2 kg. In that case, the scale can said to be reliable because it is consistent, but it is not valid at all.
The reliability of the measurement tool can be affected because of factors such as subject-related factors, observer/interviewer reliability, instrument reliability and situational reliability. Reliability of physical instruments can be tested by using calibration. However, in case of non-physical instruments such as questionnaires, the reliability of instruments is based on their stability and internal consistency. Test-retest method is used to measure reliability of the instruments.
Validity refers to the degree to which a measurement tool succeeds in measuring what is expected to be measured. Reliability and validity are interdependent concepts. There can be reliability without validity; however, there can be no validity without reliability. There can be three types of validity based on which the validity of a measuring instrument is assessed.
These are content validity, criterion-related validity and construct validity. Content validity refers to the judgement of one or more subject matter experts regarding the measurement tool. At times, researchers use a well-established measurement procedure [for example, an established survey (say Survey A) for measuring stress level] that measures a variable of interest (say, V) as a base for developing another measurement procedure [for example, newly created survey (say Survey B) for measuring stress level] that measures the same variable of interest V.
Construct validity indicates how well the measurement tool measures different constructs. Validity is assessed by one of three methods: content validation, criterion-related validation and construct validation.
Business Ethics
( Click on Topic to Read )
Corporate social responsibility (CSR)
Lean Six Sigma
Operations Research
Operation Management
Service Operations Management
Procurement Management
Strategic Management
Supply Chain
What is research design features, components, what is research design types, what is hypothesis definition, meaning, characteristics, sources, what is measure of central tendency, what is sample size determination, formula, determining,, what is research problem components, identifying, formulating,, what is causal research advantages, disadvantages, how to perform, steps in questionnaire design, what is parametric tests types: z-test, t-test, f-test, what is research types, purpose, characteristics, process, leave a reply cancel reply.
You must be logged in to post a comment.
We’ve spent the time in finding, so you can spend your time in learning
Tests and Measures
The term "tests and measures" refers to tools of measurement for analytical and diagnostic purposes. In the field of social work this might refer to a survey, instrument, diagnostic test, exam, questionnaire, survey, and/or measure.
Engagement with Tests and Measures
As a student, course work for classes in the Social Work program may have references to the tests and measures researchers used in their analysis. As you continue to engage in research and your own work as a practitioner, you might use tests and measures to begin to identify your own conclusions.
Tests and measures can be a tool of evaluation. For example, if researching social health of older adults in a population, you might engage and assign a score based on responses to the Geriatric Depression Scale. Names for these tests and measures often are very direct and clear, for example:
Resources for finding the tests and measures relevant to your topic of research are available below.
1st Edition
The last two decades have seen a marked growth in comparative research within the field of housing studies. This reflects the increasing globalisation of housing finance and therefore the interconnectedness of housing markets, growing interest among researchers and policy makers in learning from developments in other countries and the availability of more funding and better comparative data to support their endeavours. Concurrently, comparative housing research has become more sophisticated, as research training has improved, the number of journals publishing this research has increased and researchers have become what one might call more ‘methodologically aware’. However, despite these developments, there is no single volume book that deals with the distinct challenges that arise from comparative housing research, compared to other fields of comparative policy analysis. These challenges relate to spatial fixity of housing, its dual role as a consumption and investment good, and as the "wobbly pillar" of the welfare state, which is delivered using a complex mix of government and market supports. This volume reflects on the significant methodological strides made in the comparative housing research field during this period. The book also considers the considerable challenges that remain if comparative housing research is to match the methodological and theoretical sophistication evident in other comparative social science fields and maps a route for this journey. This book was published as a special issue of the International Journal of Housing Policy .
Mark Stephens is Professor of Public Policy at the Institute for Housing, Urban and Real Estate Research, Heriot-Watt University, Edinburgh. Michelle Norris is a senior lecturer in social policy at the School of Applied Social Science, University College Dublin, Ireland. The editors jointly convene the European Network for Housing Research Working Group on Comparative Housing Policy.
"Meaning and Measurement in Comparative Housing Research reflects the ongoing interest in comparative housing research, as well as ongoing efforts to reflect on and improve methodological approaches, and to draw on wider social science disciplines to work out new ways forward. After a decade’s neglect of method in comparative housing research, it is a welcome contribution." – Urban Studies, Sean McNelis, Swinburne Institute for Social Research, Australia " When reading the articles included in this interesting edited volume, it becomes evident that the field of comparative housing research has indeed witnessed farreaching developments and innovations over the past years... this volume provides a comprehensive overview of recent studies and recommendations in the field of comparative housing research, which will undoubtedly prove useful for students and practitioners alike." – Elise de Vuijst, Delft University of Technology, Delft, The Netherlands
VitalSource is a leading provider of eBooks.
This eBook is already in your shopping cart. If you would like to replace it with a different purchasing option please remove the current eBook option from your cart.
Implementation Science volume 19 , Article number: 58 ( 2024 ) Cite this article
332 Accesses
9 Altmetric
Metrics details
Implementation support practitioners (ISPs) are professionals that support others to implement evidence-informed practices, programs, and policies in various service delivery settings to achieve population outcomes. Measuring the use of competencies by ISPs provides a unique opportunity to assess an understudied facet of implementation science—how knowledge, attitudes, and skills used by ISPs affects sustainable change in complicated and complex service systems. This study describes the development and validation of a measure—the Implementation Support Competencies Assessment (ISCA)—that assesses implementation support competencies, with versatile applications across service contexts.
Recently developed practice guide materials included operationalizations of core competencies for ISPs across three domains: co-creation and engagement, ongoing improvement, and sustaining change. These operationalizations, in combination with recent empirical and conceptual work, provided an initial item pool and foundation on which to advance measurement development, largely from a confirmatory perspective (as opposed to exploratory). The measure was further refined through modified cognitive interviewing with three highly experienced ISPs and pilot-testing with 39 individuals enrolled in a university-based certificate program in implementation practice. To recruit a sample for validation analyses, we leveraged a listserv of nearly 4,000 individuals who have registered for or expressed interest in various events and trainings focused on implementation practice offered by an implementation science collaborative housed within a research-intensive university in the Southeast region of the United States. Our final analytic sample included 357 participants who self-identified as ISPs.
Assessments of internal consistency reliability for each competency-specific item set yielded evidence of strong reliability. Results from confirmatory factor analyses provided evidence for the factorial and construct validity of all three domains and associated competencies in the ISCA.
The findings suggest that one’s possession of high levels of competence across each of the three competency domains is strongly associated with theorized outcomes that can promote successful and sustainable implementation efforts among those who receive implementation support from an ISP. The ISCA serves as a foundational tool for workforce development to formally measure and assess improvement in the skills that are required to tailor a package of implementation strategies situated in context.
Peer Review reports
This study describes the development and validation of a measure—the Implementation Support Competencies Assessment (ISCA)—that assesses implementation support competencies. Measuring the use of competencies by implementation support practitioners (ISPs) provides an opportunity to assess an understudied facet of implementation science—how knowledge, attitudes, and skills used by ISPs affects implementation.
Results of the validation study offer evidence of the reliability, factorial validity, and construct validity of the ISCA.
The ISCA serves as a foundational tool for workforce development to measure and improve the skills required to build implementation capacity and to tailor multi-faceted implementation strategies situated in context.
Implementation support practitioners (ISPs) are professionals that support others to implement evidence-informed practices, programs, and policies to achieve population outcomes [ 1 , 2 , 3 , 4 , 5 ]. Several influences have contributed to the increasing attention given to describing and understanding the role of ISPs in building implementation capacity. These factors include: 1) interest in building a competent workforce for supporting implementation and evidence use [ 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 ]; 2) recent publications describing the competencies needed to be effective in an implementation support role [ 1 , 2 , 3 , 4 , 5 , 9 ]; 3) growing calls from the field of implementation science to address the emerging gap between implementation research and implementation practice—referred to as the paradoxical, ironic, or secondary gap [ 10 , 11 , 12 ]; and 4) emerging evidence that the use of multi-faceted implementation strategies to support innovations in health and social care has had limited effects on population outcomes [ 13 ].
Combined, these factors point to a need to understand how other aspects of implementation processes, beyond the use of specific implementation strategies , can contribute to improved implementation and population outcomes. Implementation support competencies could be a promising focal point on this front, which can be conceptualized as mechanisms that support professionals in providing high-quality implementation support and capacity-building; competencies represent an integration of an ISP’s values, knowledge, attitudes, and skills [ 14 ]. Measuring the use of competencies by ISPs provides a unique opportunity to assess an understudied facet of implementation science—how values, knowledge, attitudes, and skills used by ISPs affects sustainable change in service systems.
ISPs rely on technical and relational skills to identify, tailor, and improve evidence-based implementation strategies in different service contexts to ensure high-quality, consistent implementation of evidence-informed practices. To understand how ISPs do this work, it is important to systematically gather data from ISPs on (a) the skills they use to support change and (b) how confident and competent they are in using these skills to build implementation capacity [ 15 ]. Previous research foregrounded the critical question of “what it takes” to build sustainable implementation outcomes that contribute to improved and more equitable outcomes for people and communities [ 15 ]. The identification and explication of competencies for ISPs represents progress in the field toward understanding “what it takes;” however, there has remained a gap in how to measure these competencies well. This study describes the development and validation of a measure—the Implementation Support Competencies Assessment (ISCA) [ 16 ]—that assesses implementation support competencies, with versatile applications across service contexts.
The work of ISPs must account for the dynamic and highly relational nature of implementation that involves the integration of multiple stakeholder perspectives, the identification of crucial barriers to implementation that are often invisible to observers, and the assessment of available resources to address challenges and enhance facilitators [ 1 ]. Developing a workforce that can provide implementation support will require the field of implementation science to look beyond theories, models, and frameworks, and to more deeply understand how to assess and cultivate the competencies required by professionals working to promote and sustain evidence use in human service systems. Studying implementation capacity-building approaches, such as those used by ISPs, that are situated within contexts and emphasize the relational support needed to build organizational capability for service change might be a promising method for understanding how we can achieve improved implementation and population outcomes [ 13 ]. The ISCA is a tool that could support these efforts, and the specific aims of the current study are (a) to test whether the items for each ISCA competency offer a consistent and accurate representation of that competency (i.e., reliability); (b) confirm the hypothesized factor-structure of the competencies and domains within which they are nested (i.e., factorial validity), and (c) assess whether the measures are significantly associated with hypothesized outcomes of implementation support (i.e., construct validity).
Our process of developing the ISCA was informed by DeVellis [ 17 ], whereby we engaged in a systematic and rigorous process of measurement development. To begin, we leveraged recent scholarship that offers clear and rich descriptions of the constructs intended for measurement—the 15 core competencies posited to undergird effective implementation support [ 1 , 2 , 3 , 4 , 5 ]. Recently developed practice guide materials intended to inform the work of ISPs also include operationalizations or core activities for each core competency [ 18 ]. These operationalizations, in combination with recent empirical and conceptual work noted above, provided an initial item pool (116 items across the 15 competencies) and foundation on which to advance measurement development, largely from a confirmatory perspective (as opposed to exploratory).
Next, we sought to identify an optimal format for measurement. This process was informed by other extant competency measures and our desire to balance parsimony (low respondent burden) with informativeness. Ultimately, we selected an ordinal-level response-option set whereby individuals could self-report their level of perceived competence with respect to each item. Consistent with other existing competency self-assessments [ 19 ], we selected the following response-option set: 1 = not at all competent, 2 = slightly competent, 3 = moderately competent, 4 = very competent, and 5 = extremely competent. The research team then initiated a three-stage process for item review and refinement. The first stage involved members of the research team identifying opportunities to simplify and consolidate possible items in the item pool. This led to a slight reduction in items (now 113) and item simplification.
The second stage involved use of modified cognitive interviewing with three experienced ISPs. The three participants were invited to review the assessment items in preparation for their interview, and during their interview (about 60 minutes) they were asked the following questions for each competency item set: (a) how clear are the items for this competency? (b) how accessible do the items feel for potential users? (c) what changes, if any, would you recommend for these items? Feedback from respondents led to several minor edits, shifts in terminology (e.g., use of “partner” instead of “stakeholder”), and opportunities to further clarify language used in some items (e.g., defining “champions”). All potential item revisions were reviewed and accepted by two research team members with extensive implementation research and practice experience.
The third stage involved pilot-testing the assessment with a group of professionals who were enrolled in a university-based certificate program focused on cultivating ISP core competencies. Prior to the delivery of certificate program content, participants were asked to complete the ISCA. Following the completion of each competency-specific item set, participants were given the following open-ended prompts: (a) please identity any items that felt unclear or confusing; (b) please identify any language used in these items that was difficult to understand; and (c) please provide any other thoughts or insights you would like to share about these items. The assessment was completed by 39 individuals, enabling us to tentatively assess internal consistency reliability for each competency item set (Cronbach’s alpha values ranged from .70 to .94; McDonald’s omega values ranged from .70 to .95), as well as the distributional properties of item responses (results indicated the items were not burdened significantly by skewness or kurtosis). We were also able to leverage open-ended feedback to incorporate several minor item edits, which were again reviewed and approved by the same two members of the research team.
Our next step was to prepare the assessment for validation analyses. In addition to the assessment items, we developed a set of items intended to measure two core constructs posited to be associated with the ISP core competencies [ 2 ]. One construct represented ISP gains, or the extent to which ISPs report receiving recognition, credibility, and respect from those who receive their implementation support. The second construct represented recipient benefits, or the extent to which ISPs perceive the recipients of their support experiencing increases in (a) relational capacities with the ISP, (b) implementation capability, (c) implementation opportunities, and (d) implementation motivation [ 2 ]. More details about the specific items used to measure these constructs and the ISCA are provided in the Final Measures subsection.
To recruit a sample for validation analyses, we leveraged a listserv of nearly 4,000 individuals who have registered for or expressed interest in various events and trainings focused on implementation practice offered by an implementation science collaborative housed within a research-intensive university in the Southeast region of the United States. A series of emails were sent to members of this listserv describing our efforts to validate the ISCA, with an invitation to participate. Voluntary responses (no participation incentives were offered) were collected between June and November 2023 using Qualtrics, a web-based survey platform. The survey included informed consent materials, items to collect information about respondent sociodemographic and professional characteristics, the ISCA items, and validation items. The median completion time for the survey was 22.7 minutes among the 357 participants in our final analytic sample.
Table 1 features an overview of participant characteristics. The majority of participants identified as women (84%), with 15% identifying as men, 1% identifying as gender nonconforming, and 1% preferring not to provide information about their gender identity (percentages are rounded, resulting in the possibility that the total exceeds 100%). Participants could select all racial and ethnic identifies that applied to them; 76% identified as White, 11% identified as Black, 9% identified as Asian, 7% identified as Hispanic, 1% identified as Native American/American Indian/Alaska Native, 0.3% identified as Pacific Islander, 3% identified as other, and 2% preferred not to provide information about their racial/ethnic identity. Six continents of residence were represented among participants, with 78% of participants residing in North America, 7% in Europe, 6% in Australia, 4% in Asia, 4% in Africa, and 2% in South America. Thirty-eight percent indicated having more than 15 years of professional experience, 23% indicated having one-to-five years of experience, 22% indicated have six-to-ten years of experience, and the remaining 17% indicated having between 11 and 15 years of experience. The following service types were well represented among participants (more than one type could be indicated by participants): public health (32%), health (31%), mental and behavioral health (26%), child welfare (22%), and K-12 education (18%), among others. The three most common work settings were non-profit organizations (36%), higher education (27%), and state government (20%; more than one setting could be indicated by participants). See Table 1 for more details.
Implementation support competencies assessment (isca).
Rooted in recent scholarship and foundational steps of measurement development described earlier, the ISCA included item sets (ranging from 5 to 15 items and totaling 113 items) intended to measure each of 15 core competencies posited to undergird effective implementation support, with competencies nested within one of three overarching domains: co-creation and engagement, ongoing improvement, and sustaining change. The co-creation and engagement domain included items designed to measure the following five competencies: co-learning (6 items), brokering (6 items), address power differentials (7 items), co-design (6 items), and tailoring support (7 items). See Appendix 1 for a list of all items associated with this domain. The ongoing improvement domain included items designed to measure the following six competencies: assess needs and assets (6 items); understand context (6 items); apply and integrate implementation frameworks, strategies, and approaches (5 items); facilitation (9 items); communication (6 items); and conduct improvement cycles (6 items). See Appendix 2 for a list of all items associated with this domain. The sustaining change domain included items designed to measure the following four competencies: grow and sustain relationships (11 items), develop teams (15 items), build capacity (8 items), and cultivate leaders and champions (9 items). See Appendix 3 for a list of all items associated with this domain. Information about internal consistency reliability for each item set is featured in the Results section as a key component of the psychometric evaluation of the ISCA.
When completing the ISCA, participants were instructed to reflect on their experiences supporting implementation in various settings, review each item, and assess their level of competence by selecting one of the following response options: not at all competent (1), slightly competent (2), moderately competent (3), very competent (4), or extremely competent (5). If participants did not have direct experience with a particular item, they were instructed to indicate how competent they would expect themselves to be if they were to conduct the activity described in the item.
Consistent with the mechanisms of implementation support articulated by Albers et al. [ 2 ], we developed and refined multi-item scales intended to measure two constructs theorized to be byproducts of ISPs possessing proficiency across the 15 core competencies of implementation support provision. Specifically, we developed three items intended to measure ISP gains ; or the extent to which ISPs receive recognition, credibility, and respect from those who receive their implementation support. Specifically, participants were asked to indicate their level of agreement (ranging from 1 = Strongly Disagree to 5 = Strongly Agree) with the following three statements: “I have credibility among those who receive my implementation support,” “I am respected by those who receive my implementation support,” and “My expertise is recognized by those who receive my implementation support.”
We also developed ten items intended to measure recipient benefits , or the extent to which ISPs perceive the recipients of their support experiencing increases in (a) relational capacities with the ISP, (b) implementation capability, (c) implementation opportunities, and (d) implementation motivation [ 2 ]. Specifically, participants were asked to indicate their level of agreement (ranging from 1 = Strongly Disagree to 5 = Strongly Agree) with the following ten statements: “I am trusted by those who receive my implementation support;” “Those who receive my implementation support feel safe trying new things, making mistakes, and asking questions;” “Those who receive my implementation support increase their ability to address implementation challenges;” “Those who receive my implementation support gain competence in implementing evidence-informed interventions in their local settings;” “I provide opportunities for continued learning to those who receive my implementation support;” “I promote implementation friendly environments for those who receive my implementation support;” “Those who receive my implementation support strengthen commitment to their implementation work;” “Those who receive my implementation support feel empowered to engage in their implementation work;” “Those who receive my implementation support demonstrate accountability in their implementation work;” and “Those who receive my implementation support develop an interest in regularly reflecting on their own implementation work.” Information about internal consistency reliability for item sets related to the two validation constructs is featured in the Results section.
To generate evidence of the internal consistency reliability of competency-specific item sets, we estimated Cronbach’s alpha, McDonald’s omega, and Raykov’s rho coefficients for each of the 15 competencies [ 20 , 21 ]. To generate evidence of the factorial and construct validity of the ISCA, we then employed confirmatory factor analysis (CFA) in Mplus 8.6 [ 22 ]. Consistent with our hypothesized model, we estimated three separate second-order CFA models, one for each of the three competency domains: co-creation and engagement, ongoing improvement, and sustaining change. The first CFA model specified the co-creation and engagement domain as a second-order latent factor with the following five competencies specified as first-order latent factors: co-learning, brokering, address power differentials, co-design, and tailoring support. The second CFA model focused on the ongoing improvement domain as a second-order latent factor with the following six competencies specified as first-order latent factors: assess needs and assets; understand context; apply and integrate implementation frameworks, strategies, and approaches; facilitation; communication; and conduct improvement cycles. The third CFA model focused on the sustaining change domain as a second-order latent factor with the following four competencies specified as first-order latent factors: grow and sustain relationships, develop teams, build capacity, and cultivate leaders and champions. In all three models, ISP gains and recipient benefits were regressed on the second-order domain factor, and the error terms for the validation constructs were allowed to covary.
For purposes of model identification and calibrating the latent-factor metrics, we fixed first- and second-order factor means to a value of 0 and variances to a value of 1. To accommodate the ordinal-level nature of the ISCA items (and items used to measure the validation constructs), we employed the means- and variance-adjusted weighted least squares (WLSMV) estimator and incorporated a polychoric correlation input matrix [ 23 ]. Some missing values were present in the data, generally reflecting a steady rate of attrition as participants progressed through the ISCA. Consequently, the analytic sample for each second-order factor model varied, such that the model for the co-creation and engagement domain possessed all 357 participants, the model for the ongoing improvement domain possessed 316 participants, and the model for the sustaining change domain possessed 296 participants. Within each model, pairwise deletion was used to handle missing data, which enables the flexible use of partial responses across model variables to estimate model parameters. Missing values were shown to meet the assumption of Missing Completely at Random (MCAR) per Little’s multivariate test of MCAR ( \({\chi }^{2}\) [94] = 83.47, p = 0.77), a condition under which pairwise deletion performs well [ 24 , 25 ].
To assess model fit, the following indices and associated values were prespecified as being indicative of good model fit: Comparative Fit Index (CFI) and Tucker-Lewis Index (TLI) values greater than 0.95, standardized root mean square residual (SRMR) values less than .08, and root mean square error of approximation (RMSEA) values less than or equal to 0.06 (including the upper-level 90% confidence interval) [ 26 , 27 ]. Each factor-analytic model was over-identified and sufficiently powered to detect not-close model fit [ 28 ].
We submitted our study proposal (study #: 23-0958) to our university’s Office of Human Research Ethics, whereby our study was approved and determined to be exempt from further review.
Assessments of internal consistency reliability for each competency-specific item set yielded evidence of strong reliability. Specifically, Cronbach’s alpha, McDonald’s omega, and Raykov’s rho ranged from 0.82 to 0.96 across competencies. Internal consistency reliability was also strong for the two validation constructs. For ISP gains, Cronbach’s alpha and Raykov’s rho were 0.86; McDonald’s omega was 0.87. For recipient benefits, Cronbach’s alpha and Raykov’s rho were 0.91; McDonald’s omega was 0.92. See Table 2 for a detailed overview of reliability estimates across competencies and validation constructs.
Co-creation and engagement domain.
Figure 1 features the second-order factor model with co-creation and engagement specified as a second-order factor and the five corresponding competencies specified as first-order factors. ISP gains and recipient benefits were also regressed on the co-creation and engagement factor. This model yielded good model fit ( \({\chi }^{2}\) [937] = 1857.16, p < .001; CFI = 0.95; TLI = 0.95; SRMR = 0.05; RMSEA = 0.052 [upper-level 90% confidence interval: 0.056]). All first-order standardized factor loadings were statistically significant and valued between 0.66 and 0.87. All standardized second-order factor loadings were statistically significant and valued between 0.89 and 0.93. Per standardized regression coefficients, the co-creation and engagement domain also was significantly and positively associated with ISP gains ( \(\beta\) = 0.62, p < .001; R 2 = 0.38) and recipient benefits ( \(\beta\) = 0.66, p < .001; R 2 = 0.44).
Second-Order Confirmatory Factor Analysis of Domain 1 and Construct Validation (Standardized Parameters). Note: Error terms for observed indicators and full measurement models for the two focal endogenous constructs are omitted to retain visual parsimony. All parameter estimates are standardized. *** p < .001. All first-order and second-order factor loadings are significant at p < .001 level. ISP = Implementation support practitioner
Figure 2 features the second-order factor model with ongoing improvement specified as a second-order factor and the six corresponding competencies specified as first-order factors. ISP gains and recipient benefits were also regressed on the ongoing improvement factor. This model yielded good model fit, overall ( \({\chi }^{2}\) [1215] = 2707.55, p < .001; CFI = 0.95; TLI = 0.95; SRMR = 0.06; RMSEA = 0.062 [upper-level 90% confidence interval: 0.065]). All first-order standardized factor loadings were statistically significant and valued between 0.68 and 0.96. All second-order standardized factor loadings were statistically significant and valued between 0.80 and 0.95. Per standardized regression coefficients, the ongoing improvement domain also was significantly and positively associated with ISP gains ( \(\beta\) = 0.61, p < .001; ; R 2 = 0.37) and recipient benefits ( \(\beta\) = 0.64, p < .001; R 2 = 0.41).
Second-Order Confirmatory Factor Analysis of Domain 2 and Construct Validation (Standardized Parameters). Note: Error terms for observed indicators and full measurement models for the two focal endogenous constructs are omitted to retain visual parsimony. All parameter estimates are standardized. *** p < .001. All first-order and second-order factor loadings are significant at p < .001 level. FSA = frameworks, strategies, and approaches; ISP = implementation support practitioner
Figure 3 features the second-order factor model with sustaining change specified as a second-order factor and the four corresponding competencies specified as first-order factors. ISP gains and recipient benefits were also regressed on the sustaining change factor. This model yielded good model fit ( \({\chi }^{2}\) [1477] = 2927.16, p < .001; CFI = 0.96; TLI = 0.96; SRMR = 0.05; RMSEA = 0.058 [upper-level 90% confidence interval: 0.061]). All first-order standardized factor loadings were statistically significant and valued between 0.79 and 0.94. All second-order standardized factor loadings were statistically significant and valued between 0.88 and 0.94. Per standardized regression coefficients, the sustaining change domain also was significantly and positively associated with ISP gains ( \(\beta\) = 0.69, p < .001; R 2 = 0.48) and recipient benefits ( \(\beta\) = 0.75, p < .001; R 2 = 0.57).
Second-Order Confirmatory Factor Analysis of Domain 3 and Construct Validation (Standardized Parameters). Note: Error terms for observed indicators and full measurement models for the two focal endogenous constructs are omitted to retain visual parsimony. All parameter estimates are standardized. *** p < .001. All first-order and second-order factor loadings are significant at p < .001 level. ISP = Implementation support practitioner
Across all three models, standardized factor loadings associated with the validation constructs were statistically significant and ranged between 0.72 and 0.95. These details were omitted from figures to preserve visual parsimony. Taken together, results from the three models provided evidence for the factorial and construct validity of all three domains and associated competencies in the ISCA. See Appendices 1, 2, and 3 for summaries of standardized factor loadings, item communalities (i.e., proportion of item variance attributable to its corresponding latent factor), and item response frequencies. A correlation matrix of all study variables is available upon request.
With respect to alternative models, we compared the second-order factor specification for each domain with models in which only first-order factors were specified (and allowed to correlate). We then assessed differences in model fit between the first-order and second-order factor specifications. Leveraging the guidelines provided by Cheung and Rensvold [ 29 ], we specifically assessed differences in CFI values to determine whether alternative models differed significantly. Decreases in CFI values of more than 0.01-units between an original model and alternative model would indicate a significant worsening of model fit. For all three domains, the first-order specification and second-order specification did not differ significantly (i.e., changes in CFI did not exceed 0.003-units in any case; more details about these analyses are available upon request). When alternative models yield statistically negligible differences in model fit, it is good practice to favor the more parsimonious specification (i.e., the model with fewer parameter estimates and more degrees of freedom). Because second-order factor structures are more parsimonious than first-order factor structures (with all possible first-order factor correlations), we retained the second-order factor models as optimal.
As noted earlier, response rates steadily declined as participants progressed through the ISCA. As reported in Table 2 , the number of responses provided for the items associated with each competency ranged from a high of 357 (the first competency) and decreased linearly to a low of 290 (the fifteenth and final competency). The average attrition rate from competency-to-competency was 1.5%. Moreover, we did not observe any anomalous are unexpected levels of data missingness for any particular item.
Open-ended feedback from pilot-test participants also provided evidence of the acceptability of the ISCA. Pilot-test participants described the ISCA as thorough, clear, easy to understand, and applicable to their work. The ISCA also was described as a tool that could support self-reflection and guide professional development efforts. One pilot-test participant even stated that they “really enjoyed” completing the ISCA.
The purpose of the current study was to psychometrically evaluate the ISCA, a promising assessment instrument intended to measure levels of competence across 15 core competencies posited to undergird the effective provision of implementation support in various service delivery settings. Our results offer evidence of the internal consistency reliability and factorial validity of the ISCA, including its three specific domains and associated competencies. The strength of relationships between each domain and the specified validation constructs—ISP gains and recipient benefits—also provide notable evidence of the construct validity of the ISCA. In alignment with the mechanisms of implementation support articulated by Albers and colleagues [ 2 ], our findings suggest that one’s possession of high levels of competence across each of the three competency domains is strongly associated with theorized outcomes that can promote successful and sustainable implementation efforts among those who receive implementation support from an ISP.
It is important to highlight that previous research undergirding the identification and operationalization of the ISCA competencies included an integrated systematic review of strategies used by ISPs and the skills needed to use these strategies in practice, along with survey research and interviews that centered the experiences of professionals providing implementation support in diverse service sectors and geographic regions [ 1 , 2 , 3 , 4 , 5 ]. This previous research on ISPs identified the high level of skill required by those providing implementation support, leading to questions about how to select, recruit, and build the capacity of these professionals.
The ISCA serves as a foundational tool for workforce development to measure and improve the skills that are required to both engage in the relational and complex processes involved in building implementation capacity and to tailor a package of implementation strategies situated in context. As we seek to understand how the strategies and skills used by ISPs bring about change in service systems, the ISCA can be used to answer key questions posed by Albers and colleagues [ 3 ] including (a) how these competencies can be built and maintained in service settings in ways that activate mechanisms for change, (b) how different skills may be needed in different settings and contexts, and (c) how the roles of ISPs can be supported to establish cultures of learning and evidence use.
As we seek to understand how ISPs activate mechanisms of change to support implementation and evidence use, the ISCA can be used to support self-assessments that identify areas of strength and professional development opportunities for growing the skills needed to build implementation capacity. Supervisors can use the ISCA to inform professional development and decisions around recruitment and hiring. Taken together, the ISCA can be used to further define the role of key actors in the field of implementation science who represent the implementation support system [ 30 ].
The ISCA is foundational for future research on the role of implementation support and can be used for evidence-building related to implementation practice. For example, the ISCA can be used to assess whether trainings and other implementation support capacity-building activities promote gains in core competencies. The ISCA also can be used for basic implementation research including assessing the extent to which possession and use of particular competencies is associated with implementation progress across implementation stages [ 31 ] and long-term implementation outcomes in real-world settings [ 32 , 33 ].
As we seek to understand the characteristics of effective implementation teams and champions, the ISCA also can be used to identify “clusters” of competencies that appear to bolster specific support roles in various implementation efforts. Moreover, research leveraging the ISCA might be well positioned to identify the presence of specific competency portfolios possessed by members of implementation teams, highlighting the potential for teams to assemble a group of ISPs who, when brought together, provide coverage of the various competencies that undergird effective implementation support. Research on this front seems promising, as it is unlikely that any single ISP would possess high levels of competence across all 15 competencies reflected in the ISCA.
The current study possesses some limitations that should shape interpretations and conclusions. First, although there was notable diversity in the analytic sample with respect to sociodemographic and professional characteristics, study findings likely generalize best to ISPs who reside in North American and identify as White women. Second, the total analytic sample size was insufficiently large to support multiple group comparison analyses (i.e., tests of measurement invariance), whereby the psychometric properties of the ISCA could be compared across meaningful subgroups (e.g., continent of residence, gender identity, racial/ethnic identity, service type, service setting). Future research should seek to recruit very large samples that would support such analyses, which could highlight whether the ISCA performs equivalently across various respondent characteristics. Third, as a self-assessment tool, the ISCA is potentially subject to the common biases inherent in self-report data. Moreover, study participants provided both their competency assessments and responses to the outcome measures used for validation purposes. Consequently, associations between the ISCA constructs and validation constructs could be inflated due to common method variance. Future research should endeavor to validate the ISCA using outcome measures collected from the recipients of implementation support. Indeed, we view the current study as a launching point for a larger body of work intended to robustly validate the ISCA.
This study brings together several years of theory development and research on the role of ISPs and the competencies that are needed for them to be successful in their role. To date, a psychometrically validated measure of implementation support competencies has not been available. Results from the current study showcase a promising, psychometrically robust assessment tool on this front—the Implementation Support Competencies Assessment (ISCA). As a whole, results from this study also provide compelling evidence of reliability and validity with respect to the implementation support competencies identified by Metz and colleagues. Using the ISCA can shed light on the black box of many current implementation studies that fail to show positive effects of specific implementation strategies on implementation outcomes [ 13 ]. The ISCA enables understanding of the level of competency with which implementation strategies are selected, tailored, and delivered, which may be as important as the specific strategy or package of strategies selected. At the very least, the ISCA can support efforts to understand the impact that competent (or less competent) implementation support has on the outcomes of a particular implementation effort.
The datasets used and/or analyzed during the current study are available from the corresponding author upon reasonable request.
Implementation Support Practitioner
Implementation Support Competencies Assessment
Confirmatory Factor Analysis
Comparative Fit Index
Tucker-Lewis Index
Root Mean Error Square of Approximation
Weighted least squares, means- and variance-adjusted
Missing Completely at Random
Albers B, Metz A, Burke K. Implementation support practitioners- A proposal for consolidating a diverse evidence base. BMC Health Serv Res. 2020;20(1):368.
Article PubMed PubMed Central Google Scholar
Albers B, Metz A, Burke K, Bührmann L, Bartley L, Driessen P, et al. The Mechanisms of Implementation Support - Findings from a Systematic Integrative Review. Res Soc Work Pract. 2022;32(3):259–80.
Article Google Scholar
Albers B, Metz A, Burke K, Bührmann L, Bartley L, Driessen P, et al. Implementation Support Skills: Findings From a Systematic Integrative Review. Res Soc Work Pract. 2021;31(2):147–70.
Metz A, Albers B, Burke K, Bartley L, Louison L, Ward C, et al. Implementation Practice in Human Service Systems: Understanding the Principles and Competencies of Professionals Who Support Implementation. Hum Serv Organ Manag Leadersh Gov. 2021;45(3):238–59.
Google Scholar
Bührmann L, Driessen P, Metz A, Burke K, Bartley L, Varsi C, et al. Knowledge and attitudes of Implementation Support Practitioners—Findings from a systematic integrative review. Ochodo E, editor. PLoS One. 2022;17(5):e0267533.
Moore JE, Rashid S, Park JS, Khan S, Straus SE. Longitudinal evaluation of a course to build core competencies in implementation practice. Implement Sci. 2018;13(1):1–13.
Mosson R, Augustsson H, Bäck A, Åhström M, Von Thiele Schwarz U, Richter A, et al. Building implementation capacity (BIC): A longitudinal mixed methods evaluation of a team intervention. BMC Health Serv Res. 2019;19(1):1–12.
Park JS, Moore JE, Sayal R, Holmes BJ, Scarrow G, Graham ID, et al. Evaluation of the “Foundations in Knowledge Translation” training initiative: preparing end users to practice KT. Implement Sci. 2018;13(1):63.
William A, Aldridge I, Roppolo RH, Brown J, Bumbarger BK, Boothroyd RI. Mechanisms of change in external implementation support: A conceptual model and case examples to guide research and practice. 2023 Jun 21. https://doi.org/10.1177/26334895231179761 .
Westerlund A, Sundberg L, Nilsen P. Implementation of Implementation Science Knowledge: The Research-Practice Gap Paradox. Worldviews Evid Based Nurs. 2019;16(5):332–4.
Juckett LA, Bunger AC, McNett MM, Robinson ML, Tucker SJ. Leveraging academic initiatives to advance implementation practice: a scoping review of capacity building interventions. Implement Sci. 2022;17(1):1–14.
Jensen TM, Metz AJ, Disbennett ME, Farley AB. Developing a practice-driven research agenda in implementation science: Perspectives from experienced implementation support practitioners. Implement Res Pract. 2023;3:4.
Boaz A, Baeza J, Fraser A, Persson E. ‘It depends’: what 86 systematic reviews tell us about what strategies to use to support the use of research in clinical practice. Implement Sci. 2024;19(1):1–30.
Mazurek Melnyk B, Gallagher-Ford L, English Long L, Fineout-Overholt E. The establishment of evidence-based practice competencies for practicing registered nurses and advanced practice nurses in real-world clinical settings: Proficiencies to improve healthcare quality, reliability, patient outcomes, and costs. Worldviews Evid Based Nurs. 2014;11(1):5–15.
Metz A, Jensen T, Farley A, Boaz A. Is implementation research out of step with implementation practice? Pathways to effective implementation support over the last decade. Implement Res Pract. 2022;3:263348952211055.
Metz A, Albers B, Jensen T. Implementation support competencies assessment (ISCA). https://doi.org/10.17605/OSF.IO/ZH82F . 2024.
DeVellis R. Scale development: theory and applications. 3rd ed. Thousand Oaks, CA: Sage; 2012.
Metz A, Louison L, Burke K, Albers B, Ward C. Implementation support practitioner profile: Guiding principles and core competencies for implementation practice. University of North Carolina at Chapel Hill; 2020.
Wilson J, Ward C, Fetvadjiev VH, Bethel A. Measuring Cultural Competencies: The Development and Validation of a Revised Measure of Sociocultural Adaptation. J Cross Cult Psychol. 2017;48(10):1475–506.
McNeish D. Thanks coefficient alpha, we’ll take it from here. Psychol Methods. 2018;23(3):412–33.
Article PubMed Google Scholar
Padilla MA, Divers J. A Comparison of Composite Reliability Estimators: Coefficient Omega Confidence Intervals in the Current Literature. Educ Psychol Meas. 2016;76(3):436–53.
Muthén L, Muthén B. Mplus User’s Guide. 8th ed. Los Angeles, CA: Muthén & Muthén; 2017.
Flora DB, Curran PJ. An Empirical Evaluation of Alternative Methods of Estimation for Confirmatory Factor Analysis With Ordinal Data. Psychol Methods. 2004;9(4):466–91.
Shi D, Lee T, Fairchild AJ, Maydeu-Olivares A. Fitting Ordinal Factor Analysis Models With Missing Data: A Comparison Between Pairwise Deletion and Multiple Imputation. Educ Psychol Meas. 2020;80(1):41–66.
Li C. Little’s test of missing completely at random. Stata J. 2013;13(4):795–809.
Browne MW, Cudeck R. Alternative Ways of Assessing Model Fit. Sociol Methods Res. 1992;21(2):230–58.
Hu LT, Bentler PM. Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Struct Equ Modeling. 1999;6(1):1–55.
MacCallum RC, Browne MW, Sugawara HM. Power analysis and determination of sample size for covariance structure modeling. Psychol Methods. 1996;1(2):130–49.
Cheung GW, Rensvold RB. Evaluating Goodness-of-Fit Indexes for Testing Measurement Invariance. Struct Equ Modeling. 2002;9(2):233–55.
Wandersman A, Duffy J, Flaspohler P, Noonan R, Lubell K, Stillman L, et al. Bridging the Gap Between Prevention Research and Practice: The Interactive Systems Framework for Dissemination and Implementation. Am J Comm Psychol. 2008;41(3–4):171–81.
McGuier EA, Kolko DJ, Stadnick NA, Brookman-Frazee L, Wolk CB, Yuan CT, et al. Advancing research on teams and team effectiveness in implementation science: An application of the Exploration, Preparation, Implementation, Sustainment (EPIS) framework. Implement Res Pract. 2023;1:4.
Proctor E, Silmere H, Raghavan R, Hovmand P, Aarons G, Bunger A, et al. Outcomes for implementation research: Conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health. 2011;38(2):65–76.
Proctor EK, Bunger AC, Lengnick-Hall R, Gerke DR, Martin JK, Phillips RJ, et al. Ten years of implementation outcomes research: a scoping review. Implement Sci. 2023;18(1):1–19.
Download references
The authors wish to thank Amanda Farley for her support in preparation of the measure and data collection. The authors also wish to thank Mackensie Disbennett for her support in the process of reviewing and refining measurement items.
No external funding sources supported this study.
Authors and affiliations.
School of Education, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
Todd M. Jensen
School of Social Work, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
Allison J. Metz
Institute for Implementation Science in Healthcare, University of Zurich, Zurich, Switzerland
Bianca Albers
You can also search for this author in PubMed Google Scholar
The first author led analysis and drafted the methods and results sections and reviewed and edited all sections. The second author co-led conceptualization of items and drafted the background and discussion sections and reviewed and edited all sections. The third author co-led conceptualization of items and reviewed and edited all sections.
Correspondence to Todd M. Jensen .
Ethics approval and consent to participate.
The Institutional Review Board of the primary authors’ university reviewed this study and designated it as exempt (IRB #23-0958).
Not applicable.
The authors declare that they have no competing interests.
Publisher's note.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Implementation Support Competencies Assessment (ISCA) (Metz, Albers, & Jensen, 2024) – Items for the Co-Creation and Engagement Domain, Standardized Factor Loadings, Item Communalities, and Item Response Frequencies
Response Frequencies | ||||||||
---|---|---|---|---|---|---|---|---|
# | Competency and Item | FL | Com. | 1 | 2 | 3 | 4 | 5 |
| ||||||||
C1.1 | Obtain clear understanding of the system, organizational context, and culture in which implementation will take place | 0.78 | 0.61 | 6% | 33% | 47% | 14% | 0% |
C1.2 | Create opportunities for new ideas to emerge | 0.73 | 0.53 | 0% | 9% | 28% | 47% | 16% |
C1.3 | Build trust and respect for all perspectives involved in supporting implementation | 0.73 | 0.54 | 5% | 21% | 54% | 20% | 0% |
C1.4 | Communicate and listen so that you can integrate different perspectives and types of knowledge | 0.69 | 0.48 | 2% | 20% | 52% | 27% | 0% |
C1.5 | Provide interactive and educational trainings on implementation science | 0.66 | 0.44 | 7% | 23% | 28% | 28% | 14% |
C1.6 | Tailor approaches to enhance implementation readiness at individual, organizational, and system levels | 0.84 | 0.70 | 2% | 16% | 37% | 33% | 13% |
| ||||||||
C2.1 | Identify individuals or groups that should be involved in implementation and seek to understand why they were not yet included | 0.74 | 0.55 | 1% | 14% | 34% | 37% | 14% |
C2.2 | Connect individuals or groups who have been disconnected in the system by serving as a relational resource | 0.72 | 0.51 | 4% | 17% | 36% | 34% | 10% |
C2.3 | Develop and regularly convene implementation groups and teams with diverse partners | 0.80 | 0.64 | 2% | 13% | 29% | 41% | 16% |
C2.4 | Connect people strategically in a variety of ways when there is a potential for mutual benefit | 0.79 | 0.62 | 1% | 12% | 32% | 41% | 15% |
C2.5 | Support the use of evidence and data with implementation partners to support implementation | 0.77 | 0.59 | 1% | 7% | 26% | 39% | 28% |
C2.6 | Promote opportunities for implementation partners to engage with each other in the use of evidence and data | 0.83 | 0.69 | 1% | 12% | 31% | 41% | 16% |
| ||||||||
C3.1 | Put the experiences of end users (e.g., service recipients) at the center of decisions about implementation | 0.72 | 0.51 | 1% | 10% | 30% | 41% | 19% |
C3.2 | Identify how different partners can influence implementation | 0.85 | 0.72 | 0% | 9% | 33% | 42% | 15% |
C3.3 | Identify existing power structures in the implementation setting | 0.80 | 0.64 | 1% | 15% | 31% | 39% | 14% |
C3.4 | Use facilitation techniques to honor all voices involved in implementation | 0.77 | 0.59 | 2% | 13% | 25% | 39% | 20% |
C3.5 | Carefully attend to which partners hold the most and least power to influence implementation | 0.85 | 0.72 | 2% | 16% | 32% | 39% | 11% |
C3.6 | Seek and gain buy-in from formal and informal leaders (e.g., champions, opinion leaders, or others potentially influencing the implementation because of their reputation or credibility) to include diverse expertise in team discussions | 0.84 | 0.70 | 2% | 13% | 35% | 37% | 13% |
C3.7 | Support partners in developing an authentic and evolving shared understanding about implementation | 0.87 | 0.76 | 1% | 14% | 35% | 36% | 14% |
| ||||||||
C4.1 | Work with partners to build a strong fit between a selected intervention and its implementation | 0.85 | 0.72 | 3% | 8% | 34% | 39% | 16% |
C4.2 | Support collaborative implementation planning involving all partners | 0.87 | 0.76 | 1% | 8% | 30% | 43% | 18% |
C4.3 | To the extent possible, enable implementation partners to co-design any implementation tools, products, processes, governance structures, service models, strategies, and policies | 0.80 | 0.65 | 4% | 15% | 31% | 36% | 14% |
C4.4 | Promote ongoing testing of implementation tools, products, and processes to improve them | 0.77 | 0.59 | 3% | 15% | 30% | 37% | 17% |
C4.5 | Support the modification of specific implementation strategies based on local context | 0.85 | 0.71 | 2% | 12% | 27% | 40% | 19% |
C4.6 | Facilitate activities that prioritize the needs of people who are intended to benefit from the intervention being implemented | 0.83 | 0.70 | 2% | 12% | 30% | 36% | 20% |
| ||||||||
C5.1 | Regularly assess the implementation support needs and assets of different partner groups | 0.84 | 0.71 | 2% | 15% | 41% | 29% | 13% |
C5.2 | Facilitate agreement on the implementation supports that will be offered to different partner groups | 0.84 | 0.71 | 2% | 18% | 37% | 33% | 10% |
C5.3 | Develop a plan for meetings and activities (virtual or onsite) based on the goals of implementation partners | 0.80 | 0.64 | 2% | 7% | 23% | 44% | 25% |
C5.4 | Be responsive to “ad hoc”/ “just in time” support needs of implementation partners | 0.76 | 0.58 | 1% | 8% | 25% | 44% | 23% |
C5.5 | Regularly assess whether your level of support matches the needs, goals, and context of implementation | 0.78 | 0.60 | 1% | 15% | 35% | 36% | 13% |
C5.6 | Work with partners to tailor implementation strategies to meet local needs and assets | 0.87 | 0.76 | 1% | 10% | 34% | 40% | 15% |
C5.7 | Continuously promote the adaptability of implementation strategies used by partners | 0.86 | 0.75 | 1% | 14% | 33% | 37% | 16% |
Implementation Support Competencies Assessment (ISCA) (Metz, Albers, & Jensen, 2024) – Items for the Ongoing Improvement Domain, Standardized Factor Loadings, Item Communalities, and Item Response Frequencies
Response Frequencies | ||||||||
---|---|---|---|---|---|---|---|---|
# | Competency and Item | FL | Com. | 1 | 2 | 3 | 4 | 5 |
| ||||||||
C6.1 | Collaborate with partners to identify the needs and assets of different individuals and groups involved in implementation | 0.83 | 0.68 | 1% | 8% | 35% | 41% | 15% |
C6.2 | Engage people with lived experience to discover needs and assets | 0.68 | 0.47 | 3% | 16% | 28% | 35% | 18% |
C6.3 | Facilitate the identification of relevant resources to be used in implementation | 0.81 | 0.66 | 1% | 10% | 29% | 44% | 17% |
C6.4 | Support implementation partners to understand each other’s perspectives on the need for change | 0.88 | 0.77 | 2% | 10% | 36% | 39% | 12% |
C6.5 | Use a variety of data sources to highlight needs and assets related to implementation | 0.85 | 0.71 | 2% | 11% | 28% | 37% | 22% |
C6.6 | Use data to explore the unique needs of specific populations (e.g., race, ethnicity, gender, socioeconomic status, geography, ability status) | 0.78 | 0.60 | 3% | 13% | 32% | 32% | 20% |
| ||||||||
C7.1 | Involve diverse partners from throughout the system to identify and understand the implications of implementation | 0.87 | 0.76 | 1% | 15% | 35% | 35% | 14% |
C7.2 | Review available evidence to determine the relevance and fit of the proposed intervention to be implemented | 0.80 | 0.64 | 1% | 10% | 27% | 41% | 20% |
C7.3 | Assess the fit of the proposed intervention with the values, needs, and resources of the service setting | 0.89 | 0.79 | 2% | 11% | 32% | 38% | 18% |
C7.4 | Assess the fit of the proposed intervention with the current political, financial, and organizational contexts | 0.83 | 0.69 | 5% | 14% | 37% | 32% | 12% |
C7.5 | Continuously identify and respond to changes in the systems which affect implementation | 0.88 | 0.77 | 3% | 13% | 34% | 37% | 13% |
C7.6 | Identify and support actions that manage risks and assumptions for implementation | 0.88 | 0.78 | 5% | 16% | 40% | 30% | 8% |
| ||||||||
C8.1 | Remain up to date on evidence developed through implementation research and practice | 0.91 | 0.82 | 4% | 15% | 31% | 35% | 16% |
C8.2 | Remain up to date on knowledge about implementation frameworks, models, theories, and strategies | 0.90 | 0.81 | 3% | 18% | 33% | 32% | 14% |
C8.3 | Educate partners about the best available evidence on implementation frameworks, strategies, and approaches that could be used to support implementation | 0.86 | 0.74 | 7% | 19% | 32% | 31% | 10% |
C8.4 | Include all relevant partners in the selection, combination, and co-design of implementation strategies and approaches | 0.93 | 0.86 | 6% | 13% | 37% | 35% | 9% |
C8.5 | In collaboration with partners, support the use of implementation frameworks, approaches, and strategies that are best suited for the specific service setting | 0.93 | 0.87 | 6% | 13% | 35% | 34% | 12% |
| ||||||||
C9.1 | Ensure that meetings and convenings to support implementation are welcoming and engaging for all participants | 0.75 | 0.57 | 1% | 7% | 20% | 51% | 21% |
C9.2 | Support relevant partners in identifying barriers to implementation | 0.88 | 0.77 | 1% | 7% | 24% | 48% | 21% |
C9.3 | Facilitate the identification of partners needed to develop and execute strategies for addressing barriers to implementation | 0.89 | 0.80 | 2% | 8% | 31% | 41% | 18% |
C9.4 | Serve as a formal and informal facilitator as determined by an analysis of the implementation challenge and context | 0.86 | 0.74 | 4% | 12% | 23% | 41% | 20% |
C9.5 | Support implementation partners to generate and prioritize ideas to address barriers to implementation | 0.87 | 0.76 | 1% | 9% | 27% | 44% | 19% |
C9.6 | Support partners to evaluate alternatives, summarize key points, sort ideas, and exercise judgment in the face of simple challenges with easy solutions | 0.88 | 0.77 | 2% | 12% | 29% | 41% | 16% |
C9.7 | Support partners to generate alternatives, facilitate open discussion, gather different points of view, and delay quick decision-making in the face of complex challenges with no easy solutions | 0.87 | 0.76 | 3% | 14% | 32% | 36% | 15% |
C9.8 | Use facilitation methods (e.g., action planning, brainstorming, role playing, ranking, scenario development) that match the implementation challenge | 0.80 | 0.63 | 3% | 17% | 25% | 38% | 17% |
C9.9 | Respond to emergent implementation challenges with flexibility and adaptability | 0.87 | 0.76 | 2% | 10% | 27% | 42% | 19% |
| ||||||||
C10.1 | Work with partners to develop communication protocols that facilitate engagement with each other. | 0.95 | 0.90 | 4% | 17% | 32% | 35% | 12% |
C10.2 | Work with partners to develop communication protocols that communicate and celebrate implementation progress | 0.94 | 0.89 | 4% | 14% | 35% | 35% | 12% |
C10.3 | Work with partners to develop communication protocols that report barriers hindering implementation | 0.96 | 0.93 | 5% | 19% | 33% | 32% | 12% |
C10.4 | Work with partners to develop communication protocols that periodically review past decisions to continually assess their appropriateness | 0.92 | 0.84 | 7% | 21% | 35% | 26% | 11% |
C10.5 | Support the development of tailored communication protocols for different audiences | 0.88 | 0.77 | 3% | 18% | 31% | 34% | 14% |
C10.6 | Encourage partners to regularly communicate with and gather feedback from individuals inside and outside the implementing system | 0.89 | 0.79 | 3% | 16% | 30% | 35% | 17% |
| ||||||||
C11.1 | Facilitate the identification of relevant quantitative and qualitative data about implementation activities and outcomes | 0.88 | 0.77 | 3% | 12% | 37% | 26% | 22% |
C11.2 | Support the development of processes and structures for the routine collection, analysis, and interpretation of implementation data | 0.89 | 0.79 | 3% | 17% | 32% | 31% | 18% |
C11.3 | Ensure that different partners have access to relevant, valid, and reliable data to help guide implementation decision-making | 0.90 | 0.80 | 3% | 17% | 30% | 36% | 15% |
C11.4 | Encourage the collection and use of data to explore the impact of implementation on different subgroups | 0.88 | 0.77 | 3% | 12% | 31% | 35% | 20% |
C11.5 | Develop partners’ capacity to continuously use data for implementation decision-making through modeling, instruction, and coaching | 0.90 | 0.81 | 4% | 17% | 31% | 33% | 15% |
C11.6 | Help create structures that ensure that crucial information about implementation and improvement is circulated among all partners | 0.95 | 0.89 | 5% | 19% | 33% | 30% | 14% |
Implementation Support Competencies Assessment (ISCA) (Metz, Albers, & Jensen, 2024) – Items for the Sustaining Change Domain, Standardized Factor Loadings, Item Communalities, and Item Response Frequencies
Response Frequencies | ||||||||
---|---|---|---|---|---|---|---|---|
# | Competency and Item | FL | Com. | 1 | 2 | 3 | 4 | 5 |
| ||||||||
C12.1 | Build trust with implementation partners by being transparent and accountable in all actions | 0.87 | 0.75 | 1% | 2% | 19% | 48% | 29% |
C12.2 | Build relationships with implementation partners from all parts of the implementation setting | 0.87 | 0.76 | 1% | 5% | 25% | 46% | 23% |
C12.3 | Continuously evaluate the strengths and weaknesses of your relationships with implementation partners | 0.86 | 0.74 | 2% | 14% | 31% | 39% | 14% |
C12.4 | Seek and incorporate feedback from implementation partners about the strengths and weaknesses of your relationships with them | 0.80 | 0.64 | 3% | 21% | 26% | 36% | 15% |
C12.5 | Facilitate open communication that enables difficult conversations with implementation partners, when needed, to regulate distress in your relationships with them | 0.81 | 0.66 | 3% | 12% | 33% | 36% | 16% |
C12.6 | Demonstrate your competency to implementation partners | 0.87 | 0.76 | 2% | 9% | 32% | 43% | 15% |
C12.7 | Enter the implementation setting with humility as a learner | 0.73 | 0.53 | 5% | 17% | 46% | 32% | 0% |
C12.8 | Demonstrate commitment and persistence in the face of complex challenges | 0.86 | 0.74 | 0% | 4% | 17% | 47% | 32% |
C12.9 | Encourage and enable implementation partners to share their perspectives openly and honestly | 0.87 | 0.75 | 1% | 4% | 17% | 51% | 27% |
C12.10 | Normalize implementation challenges; ask questions; ask for support from implementation partners | 0.89 | 0.78 | 1% | 6% | 19% | 43% | 32% |
C12.11 | Support implementation partners to understand each other’s perspective; highlight areas of shared understanding and common goals | 0.90 | 0.81 | 1% | 10% | 23% | 45% | 21% |
| ||||||||
C13.1 | Guide efforts to assemble implementation teams | 0.83 | 0.70 | 2% | 11% | 36% | 37% | 15% |
C13.2 | Facilitate the development of clear governance structures for implementation teams | 0.84 | 0.71 | 7% | 21% | 36% | 27% | 10% |
C13.3 | Support teams to select, operationalize, tailor, and adapt interventions | 0.88 | 0.77 | 2% | 14% | 35% | 37% | 12% |
C13.4 | Support teams to develop operational processes and resources for building staff competency | 0.87 | 0.76 | 3% | 16% | 37% | 32% | 13% |
C13.5 | Support teams to identify, collect, analyze, and monitor meaningful data | 0.79 | 0.62 | 1% | 14% | 30% | 36% | 18% |
C13.6 | Support teams to engage leadership, staff, and partners in using data for improvement | 0.85 | 0.72 | 2% | 14% | 28% | 38% | 18% |
C13.7 | Support teams to build capacity for sustained implementation | 0.90 | 0.80 | 1% | 14% | 34% | 38% | 13% |
C13.8 | Support teams to build cross-sector collaborations that are aligned with new ways of work | 0.80 | 0.64 | 3% | 21% | 34% | 31% | 11% |
C13.9 | Support teams to develop effective team meeting processes, including the establishment of consistent meeting schedules and standing agendas | 0.78 | 0.61 | 1% | 12% | 31% | 37% | 19% |
C13.10 | Ensure implementation teams have sufficient support from organizational leadership to promote successful implementation | 0.82 | 0.67 | 2% | 18% | 40% | 31% | 10% |
C13.11 | Help to develop communication protocols that ensure relevant information about implementation is circulated among implementation teams and their members | 0.82 | 0.67 | 2% | 18% | 34% | 36% | 10% |
C13.12 | Develop processes for ongoing assessment and improvement of implementation team functioning | 0.84 | 0.70 | 4% | 21% | 34% | 32% | 9% |
C13.13 | Support implementation teams in providing opportunities for learning and professional development to its members | 0.85 | 0.73 | 2% | 17% | 32% | 36% | 13% |
C13.14 | Work to enhance cohesion and trust among implementation team members | 0.84 | 0.71 | 1% | 12% | 31% | 40% | 16% |
C13.15 | Help manage and resolve conflict among implementation team members | 0.81 | 0.65 | 4% | 19% | 35% | 34% | 9% |
| ||||||||
C14.1 | At the outset of implementation, model with implementation partners the changes that will be implemented | 0.86 | 0.74 | 5% | 19% | 33% | 34% | 10% |
C14.2 | Work with implementation partners to assess capacity for sustained implementation, including budget considerations | 0.83 | 0.69 | 4% | 23% | 38% | 25% | 10% |
C14.3 | Facilitate implementation partners’ access to capacity-building training, modeling, or coaching for implementation | 0.84 | 0.70 | 3% | 16% | 35% | 32% | 13% |
C14.4 | Model with implementation partners relevant knowledge, skills, behaviors, and practices | 0.91 | 0.82 | 4% | 11% | 29% | 41% | 15% |
C14.5 | Coach implementation partners in their use of relevant knowledge, skills, behaviors, and practices | 0.89 | 0.80 | 5% | 14% | 32% | 34% | 16% |
C14.6 | Help identify and shape organizational processes needed to build capacity for implementation | 0.86 | 0.74 | 3% | 19% | 34% | 32% | 13% |
C14.7 | Support implementation partners in identifying and addressing future challenges or barriers to sustained implementation | 0.87 | 0.76 | 3% | 13% | 36% | 35% | 13% |
C14.8 | Promote collaboration and new partnerships that will build the capacity of implementation partners | 0.84 | 0.70 | 3% | 11% | 38% | 34% | 14% |
| ||||||||
C15.1 | Identify existing leaders who can support implementation | 0.85 | 0.72 | 1% | 11% | 27% | 43% | 18% |
C15.2 | Help build the capacity of leaders to lead implementation | 0.92 | 0.85 | 4% | 17% | 30% | 36% | 14% |
C15.3 | Support partners in developing processes for regular coordination meetings with leaders related to implementation | 0.92 | 0.84 | 3% | 13% | 32% | 38% | 13% |
C15.4 | Help identify and involve emerging leaders who can support implementation | 0.88 | 0.78 | 3% | 16% | 31% | 34% | 16% |
C15.5 | Build the capacity of emerging leaders to support implementation | 0.92 | 0.84 | 4% | 16% | 32% | 34% | 14% |
C15.6 | Support implementation partners in navigating any transitions in organizational leadership | 0.85 | 0.72 | 7% | 23% | 32% | 30% | 7% |
C15.7 | Support implementation partners in identifying champions who can support implementation (champions are professionals or lay persons who volunteer or are appointed to enthusiastically promote and support the implementation of an innovation) | 0.91 | 0.83 | 3% | 15% | 29% | 39% | 15% |
C15.8 | Support implementation partners in involving champions throughout the course of implementation | 0.94 | 0.89 | 3% | 15% | 28% | 41% | 13% |
C15.9 | Support implementation partners in reviewing and strengthening champion roles | 0.93 | 0.86 | 6% | 17% | 29% | 38% | 10% |
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/ .
Reprints and permissions
Cite this article.
Jensen, T.M., Metz, A.J. & Albers, B. Development and psychometric evaluation of the Implementation Support Competencies Assessment. Implementation Sci 19 , 58 (2024). https://doi.org/10.1186/s13012-024-01390-8
Download citation
Received : 11 April 2024
Accepted : 01 August 2024
Published : 06 August 2024
DOI : https://doi.org/10.1186/s13012-024-01390-8
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
ISSN: 1748-5908
The FDA has approved a new targeted drug specifically for brain tumors called low-grade gliomas. The drug, vorasidenib, was shown in clinical trials to delay progression of low-grade gliomas that had mutations in the IDH1 or IDH2 genes.
“Although there have been other targeted therapies for the treatment of brain tumors with the IDH mutation, [this one] has been one of the most successful in survival prolongation of brain tumor patients,” said Darell Bigner, MD, PhD, the E. L. and Lucille F. Jones Cancer Distinguished Research Professor and founding director of the Preston Robert Tisch Brain Tumor Center at Duke.
In clinical trials , progression-free survival was estimated to be 27.7 months for people in the vorasidenib group versus 11.1 months for those in the placebo group.
Bigner, Katherine Peters, MD, PhD, professor of neurology and neurosurgery, and others at the Duke Brain Tumor Center played pivotal roles in the development and approval of the drug: Bigner in the early collaborations with Johns Hopkins University that led to the discovery of the IDH mutation, and Peters, more recently, as lead investigator in the clinical trials.
Patents developed from the early collaborations were licensed to industry through the Duke University Office for Translation & Commercialization, making this the seventh drug currently on the market with Duke intellectual property roots.
Drs. Bigner and Peters as well as Hai Yan, MD, PhD (formerly the Henry S. Friedman Distinguished Professor of Neuro-Oncology at Duke) answered questions about the work that led to vorasidenib.
Bigner: The discovery of the mutant IDH gene is one of the most important discoveries in neuro-oncology. The IDH mutation has been incorporated by the World Health Organization into the rapid and accurate diagnosis and classification of astrocytic, oligodendroglial, and glioblastoma multiforme brain tumors. Never before has there been a single gene mutation that contributed so greatly to classification. Most importantly, it was immediately recognized that the IDH mutation could be targeted with drugs to treat the group of patients that had malignant brain tumors that expressed the IDH mutation.
Bigner: Perhaps the most important collaboration between Johns Hopkins and Duke came in the work that led to the discovery of the IDH mutation. The National Cancer Institute had established a program in which genome sequencing of all the major cancers was to be done and decided that glioblastoma would be the first cancer that they investigated. The Johns Hopkins and Duke group decided to also perform complete genome sequencing of glioblastoma. The NCI did not do complete genome sequencing. Using the Duke material, the Johns Hopkins group sequenced the entire genome that could be done at that time, in 2008. The sequencing at that time was very laborious, rather than in the automated manner that can be done now. By doing almost complete genome sequencing, the Johns Hopkins and Duke group discovered the IDH mutation. The collaboration with Johns Hopkins was strengthened when [we] recruited Dr. Hai Yan [to Duke] in 2003. Dr. Yan had just completed a 5-year period as a post-doctoral research fellow at Johns Hopkins with Dr. Bert Vogelstein in Cancer Molecular Genetics.
Yan: Subsequent research from these teams produced numerous publications that further elucidated the pathological roles of IDH mutations, leading to the reclassification of gliomas in the WHO CNS classification. This body of work ultimately paved the way for the development of targeted therapies, culminating in the approval of [vorasidenib]. This collaboration … exemplifies the power of interdisciplinary and inter-institutional cooperation in driving scientific discovery and innovation in cancer treatment.
Yan: Mutations in the IDH1 or IDH2 genes result in elevated levels of the oncometabolite D-2HG, disrupting normal cellular functions and contributing to tumorigenesis. Vorasidenib selectively binds to the mutated IDH1 and IDH2 enzymes, inhibiting their activity and thereby reducing the production of D-2HG. This inhibition helps to restore normal cellular processes, reduce tumor cell proliferation, and promote the differentiation of cancer cells.
Yan: The development and approval of vorasidenib represent a significant milestone in the field of oncology, particularly in the treatment of brain cancers. It validates the approach of targeting specific genetic mutations with precision therapies and reinforces the importance of personalized medicine in oncology. This success is likely to inspire further research into targeting other genetic mutations and metabolic pathways in various cancers.
Bigner: There are indeed plans to explore the potential of vorasidenib beyond its current indications. Researchers are investigating its use in combination with other therapies, such as immune checkpoint inhibitors, to enhance therapeutic efficacy. Additionally, studies are being planned or are already under way to assess the effectiveness of vorasidenib in treating other types of brain cancer s, solid tumors and leukemia with IDH mutations. The ongoing research aims to expand the therapeutic applications of vorasidenib and optimize its use in various clinical settings, potentially benefiting a broader spectrum of cancer patients.
What were the outcomes of the clinical trial for vorasidenib.
Peters: The INDIGO clinical trial was a phase 3 trial of vorasidenib, an oral inhibitor of mutant IDH1/2 that can readily cross the blood-brain barrier, versus placebo in patients with mutant IDH1/2 glioma. Treatment with vorasidenib significantly improved progression-free survival (27.7 months vorasidenib vs. 11.1 for placebo).
The key secondary endpoint was time to next intervention, which means the time to needing chemotherapy, radiation therapy, or more surgery. For patients receiving placebo, the median time to next intervention was 17.8 months, but for patients receiving vorasidenib, the median time to next intervention has not yet been reached. Thus, patients on vorasidenib could significantly delay chemotherapy, radiation therapy, or more surgery. Most importantly, the vorasidenib was well tolerated with only 3.6% of patients needing to stop the drug because of an adverse event.
Peters: Results showed that throughout the study, patients with IDH mutant low grade glioma had a good quality of life, and it was preserved throughout the study. Patients on vorasidenib were able to maintain their cognitive abilities and did not have any decline in their quality of life or cognition.
Peters: At Duke, we are conducting studies of vorasidenib on patients with high-grade tumors and enhancing disease. Most of these studies look at combining vorasidenib with immunotherapy. It will be exciting to see what will happen with the INDIGO study's long-term outcomes.
Peters : It is exciting to have a drug specifically targeted for these patients by inhibiting the mutant IDH enzyme. With vorasidenib being orally available, well-tolerated, and does not impair quality of life or cognition, we can extend people’s lives and delay the use of treatments such as radiation therapy and chemotherapy. I am so thankful to all the patients who participated in the groundbreaking study and for paving the way for future patients.
IMAGES
COMMENTS
Measurement is the process of observing and recording the observations that are collected as part of a research effort. There are two major issues that will be considered here. First, you have to understand the fundamental ideas involved in measuring. Here we consider two of major measurement concepts. In Levels of Measurement, I explain the ...
be done to measure something, whether measuring brain activity, attitude toward an object, organizational emphasis on research and development, or stock market performance. Therefore, these rules include a range of things that occur during the data collection process, such as how questions are worded and how a measure is administered.
Most of the measurements in Psychology a re on the interval scale. e.g. the Likert scale, RATIO MEASUREMENT. This is a further refinement in the measurement levels in that it provides us with ...
The importance of measurement in research and technology is indisputable. Measurement is the fundamental mechanism of scientific study and development, and it allows to describe the different phenomena of the universe through the exact and general language of mathematics, without which it would be challenging to define practical or theoretical approaches from scientific investigation.
Ratio level data provide the final and most robust level of measurement. Ratio level data are measured continuously, with equal spacing between intervals and with a true zero. Examples include height, weight, heart rate, and serum laboratory values. A zero value is interpreted as the absence of the characteristic.
In research, measurement is a systematic procedure for assigning scores, meanings, and descriptions to concepts so that those scores represent the characteristic of interest. Social scientists can and do measure just about anything you can imagine observing or wanting to study. Of course, some things are easier to observe or measure than others.
Liu and coauthors review the major data sources, measures and analysis methods in the science of science, discussing how recent developments in these fields can help researchers to better predict ...
Measurement in science begins with the activity of distinguishing groups or phenomena from one another. This process, which is generally termed classification, implies that we can place units of scientific study—such as victims, offenders, crimes, or crime places—in clearly defined categories or along some continuum.
RESEARCH SCENARIOS AND INSTRUMENT DEVELOPMENT OR ADAPTATION. Epidemiological studies require well-defined and socially relevant research questions, which, in turn, demand reliable and accurate measurements of the phenomena and concepts needed to answer them 8.Berry et al. 9 discuss three perspectives that are particularly relevant for the issues at hand.
Measurement is central to empirical research whether observational or experimental. A study of a novel, well-defined research question can fall apart due to inappropriate measurement. Measurement is defined in a variety of ways (Last 2001; Thorndike 2007; Manoj and Lingyak 2014 ), yet common to all definitions is the systematic application of ...
Reliability is consistency across time (test-retest reliability), across items (internal consistency), and across researchers (interrater reliability). Validity is the extent to which the scores actually represent the variable they are intended to. Validity is a judgment based on various types of evidence.
Define measurement and give several examples of measurement in psychology. ... This is why the research literature often includes different conceptual definitions of the same construct. In some cases, an older conceptual definition has been replaced by a newer one that works better. In others, researchers are still in the process of deciding ...
The dedicated Measurement journal special issue is featuring selected papers from the IMEKO TC-4 2023 Symposium in Pordenone, Italy. This curated collection offers a comprehensive glimpse into the latest research and innovations in measurement …. Submission deadline: 03 March 2024.
In scientific research, a variable is anything that can take on different values across your data set (e.g., height or test scores). There are 4 levels of measurement: Nominal:the data can only be categorized. Ordinal:the data can be categorized and ranked. Interval:the data can be categorized, ranked, and evenly spaced.
Reliability is consistency across time (test-retest reliability), across items (internal consistency), and across researchers (interrater reliability). Validity is the extent to which the scores actually represent the variable they are intended to. Validity is a judgment based on various types of evidence.
Measurement is a cornerstone of trade, science, technology and quantitative research in many disciplines. Historically, many measurement systems existed for the varied fields of human existence to facilitate comparisons in these fields.
Quantitative research is based on measurement and is conducted in a systematic, controlled manner. These measures enable researchers to perform statistical tests, analyze differences between groups, and determine the effectiveness of treatments. If something is not measurable, it cannot be tested. Keywords: measurements; quantitative research ...
Measures exist to numerically represent degrees of attributes. Quantitative research is based on measurement and is conducted in a systematic, controlled manner. These measures enable researchers to perform statistical tests, analyze differences between groups, and determine the effectiveness of treatments. If something is not measurable, it ...
Indicators: A way to measure and monitor a given milestone, outcome, or construct and help determine if our assumptions are correct. Examples include math test scores, reported burglaries, or daily wages. The relation between constructs, indicators and data is illustrated below: 1. Constructs, indicators, and data.
Construct validation—collecting evidence that the instruments scientists build actually measure the constructs scientists claim they measure—is a difficult and necessary part of the research process (Cronbach & Meehl, 1955). It has taken many programs of research, thousands of studies, and decades of work to identify, define, and measure ...
There are typically four levels of measurement that are defined: Nominal. Ordinal. Interval. Ratio. In nominal measurement the numerical values just "name" the attribute uniquely. No ordering of the cases is implied. For example, jersey numbers in basketball are measures at the nominal level.
In order to analyse data, the variables have to be defined and categorised using. different scales of measurements. There are four scales of measurements- nominal scale, ordinal scale, interval ...
SUPPORTS OPEN ACCESS. Launched in 1923 Measurement Science and Technology was the world's first scientific instrumentation and measurement journal and the first research journal produced by the Institute of Physics. It covers all aspects of the theory, practice and application of measurement, instrumentation and sensing across science and engineering.
3.4 Formation of Index. 4 Criteria of Good Measurement Tool. 4.1 Reliability. 4.2 Validity. In simple words, measurement means using a yardstick to determine the characteristics of a physical object. In addition to physical objects, qualitative concepts, such as songs and paintings, or an abstract phenomenon, can also be measured.
The term "tests and measures" refers to tools of measurement for analytical and diagnostic purposes. In the field of social work this might refer to a survey, instrument, diagnostic test, exam, questionnaire, survey, and/or measure. ... Resources for finding the tests and measures relevant to your topic of research are available below. Books ...
The last two decades have seen a marked growth in comparative research within the field of housing studies. This reflects the increasing globalisation of housing finance and therefore the interconnectedness of housing markets, growing interest among researchers and policy makers in learning from developments in other countries and the availability of more funding and better comparative data to ...
Measurement development process. Our process of developing the ISCA was informed by DeVellis [], whereby we engaged in a systematic and rigorous process of measurement development.To begin, we leveraged recent scholarship that offers clear and rich descriptions of the constructs intended for measurement—the 15 core competencies posited to undergird effective implementation support [1,2,3,4,5].
The Next Generation Anritsu Site Master represents a fusion of cutting-edge technology, customer-driven innovation, and decades of expertise in test and measurement solutions. It seamlessly integrates cable and antenna analysis with spectrum analysis and monitoring functionalities, offering a comprehensive testing solution for professionals ...
Duke brain tumor researchers are part of earliest collaborations that led to the development of the drug, shown to more than double progression-free survival The FDA has approved a new targeted drug specifically for brain tumors called low-grade gliomas. The drug, vorasidenib, was shown in clinical trials to delay progression of low-grade gliomas that had mutations in the IDH1 or IDH2 genes.