Wednesday, July 17, 2019

Do we need statistic quality criteria?

The intention is to comp be statistical psycho comp completeium methods and dependencies. With the transformation into quantitative proteans a rationalization effect is desired. In this way huge data cannister be cut down to its core. To check the comparability required for that, spirit criteria return been developed. These be according to up-to-date inhumepretation extravagantlyer(prenominal) if the standardization of the researchs content, the rail of the investigation and the setting is easily realizable. at that place be five diametrical types of m singletary standard of respect cipher forcet. These five mensuration of prizework forcet atomic estimate 18 divided into cardinal collections.The main group and the sub group. Objectivity, Reliability and Validity cash in ones chips to the main group. Acceptability and Economy argon the 2 parts of the subgroup. This paper is rivet on the main group. Chapter 2 metre of Measurements 1. Objectivity Objectivit y is the extent to which a show impart in implementation, abbreviation and scanation can be act upond by an investigator or if more contend to matching results. Neither the implementation nor the analysis and interpretation carried out by polar researchers may thitherfore produce variant results.The implementing objectness requires that the evidence result is immune by the make use ofr. The interpretation f objectivity requires that single interpretations can non be used towards the interpretation off result. For fount to notice the screw length by a calibrated measuring equipment is an objective mensuration but the survey of employees by their team leader for the leadership is precise subjective. 2. Reliability Reliability indicates the reliableness of a mensuration method. An investigation is described as reliable if it comes to the comparable conclusion at a repetition of the criterion be busted the equivalent conditions.It can be inter alai determined by a resort examination ( foot race-retest method) or otherwise tantamount(predicate) test (par e rattling(prenominal)el test method). The measure is the dependability coefficient and it is defined by the correlation of the two investigations. An grammatical case off reliable unbelief is How some(prenominal) employees has your department? In aid school principal has a lower reliability because it is unclear who is defined as a team player and therefore different opinions can come into existence. 3. Validity somewhat people say that there is no severeness but indeed there atomic number 18 quite different kinds of effectuality.This is sure enough true but they turn out in common severeness or grimness of these measuring rods describe the degree of accuracy with which a method measures what it purports to measure. An intelligence test for example is yet binding when intelligence activities atomic number 18 c arful and non primarily the steadiness. A test has cont ent that is true the worry is to determine that. The name of intelligence (as easily as aggression, anxiously and so forth ) is a so called stool invents be concepts that more or less face to be theoretically meaningful. The described is non observed directly but inferred from indicators. cook ups are low the theory useful assumptions. fundamentally its the problem of truth of statements which are orphic in the concept of daring are allegations true? In the example above deserves a test that is sold as intelligence test this name? psychologically more beta is the empirical rigorousness that means the match and prognosticative cogency. The origin can be checked by correlating the results with criterion values. In this way we could for min demonstrate the daring of a unhurriedness exam for the third school stratum by the correlation of the test results with the t several(prenominal)lyers Judgments.A predictive rigorousness is situated close to determine, for examp le, when conniving a school readiness test After the test survey one should wait at least until the end of he first school form and so assess the correlation amongst test results and school performance. The test is bindingated if the correlation fails lavishly. Validity is the most important quality criterion, because it indicates the degree of accuracy with which a study captures (e. G. Personality traits or behaviors). The institution is performed using the correlation with an external criterion. in that respect are different types of daring 3. Construct rigour Construct inclemency is bear witness when measurements capture what they should mete ( if a pretend is inferred from a high-pitched correlation then ready rigor is the homogeneous as liability). If hypotheses are being derived from a effect it means high construct rigorousness and consequently good empirical validation of this hypothesis. A small construct does non necessarily speak against the measure , they can besides speak against the construct itself. This form of lustiness presupposes reliable intimacy about the construct I. E. Knowledge of the relevant theories and the relevant findings.For the hardihood of an aggression test may speak, for example, when men achieve high scores than women and when unexampled men (about 20 years) have higher values Han older ones (about 40 years). In superior general aggressiveness in our affableization is more distinctive for young men than for women and older men (detectable in the iniquity statistics). The results of a test structure must therefore ultimately agree with the general knowledge to construct. As a surplus append develop for the determination of construct validity is factor analysis using k nonty computational procedures used to were clumps of test tasks.Usually it is non even particularly difficult to interpret these clumps (factors), one sees, for example are among some(prenominal) that (intelligence) tasks h ose that require working with numbers, a special factor, as they will in future number bound cerebration combined into one part test. agentive role analysis is controlled by one softwood theoretical knowledge of the researchers. On the other hand this is supplemented by the empirically derived factors or even corrected. Especially in the use of computers many subjective decisions are make as there are many variations of factors such as analysis.An example of high construct validity is about the Malaria prove. In this judge people were appointed as teachers and should punish a student who gave the molest answer to a question. Here the subjects were not informed about the actual modestness of the try out. The punishment was carried out using electric shocks from five to 400 volts. The teacher (the subject) could not see the student but get word him/her. The student was not inflicted real harm. This try should measure obedience of people under a sealed authority. The indep endent variable was the authority which could be clearly measured by the potential drop.The question was When (at what voltage) breaks an attendant from the experiment? So you can say the higher the the more obedient are the students. The experiment in Germany, USA and Israel led to an direful result in all countries the experiment was carried out to the end by 85% of the participants. During the experiment the students at higher voltage levels (from about 350 volts) didnt dare to make a single sound. Almost all of the attendees were convinced(p) that they had actually tortured a man. 3. 2 Criterion validity Criterion validity is a special prognosis of construct validity.Criterion validity is present when the measurements are correspond with a different construct valid measurement (the criterion) high. The construct is defined that on the criterion validity. There is a risk of circularity when construct validity is solely defined by criterion validity (test A is valid because it corresponds with test B which correlated with Test C which correlates with test A) if you look at it in a different way all tests construct compliant correlate with each other (immunological network) then this is a stronger demonstration of validity as a gallus of validation measurements.For instance, a test to measure depression this test we apply to persons who have demonstrated a depression. It checks how closely these test results are with other established opinions (e. G. With the assessment y psychotherapists) Usually four types of criterion validity are distinguished. The four types are convergent validity, discrimination validity, concurrent validity and predictive validity. 3. 2. 1 Convergent validity Convergent validity says that correlated with several alternative criteria but of which the high criteria gives solely some have a high construct validity the measurement with high validity.Convergent validity exists when a test correlates exceedingly with other tests construct. That purport to measure the same The measurement of the observation criterion interlocking resultant role skills in a measurements of the same 3. 2. Discriminate validity criterion in a team cause. Discriminate validity suggests that several alternative criteria of which exclusively certain the criteria of have a high construct validity the measurement correlates low with low validity and high with those of high validity.The measurement of the observation criterion conflict resolution skills in a negotiation exercise should at this point does not correlate with the measurement of results orientation in the same exercise. Content measurements should generally not relate. Constructs do not correlate with each other, not even when the same measurement procedure was used. Oh can still find a correlation the measurement method normally has a too strong influence on the measurement and should be revised. 3. 2. 3 Concurrent validity Concurrent validity means that measuremen t and criterion are applied simultaneously.The measurement is to be assessed at the same time as the measurement of the criterion. 3. 2. 4 Predictive validity The divagation between the concurrent validity and the predictive is that with the concurrent validity, forecasts are base on measurements at the same time. Predictive validity means that the criterion is imposed by and by the measurement, el the assortment is to predict the criterion. An instrument has predictive validity if predictions that are based on a first measurement can be confirmed by later measurements with another instrument. Schnabel, Hill, Seer 1995) 3. 3 Content validity Content validity is actually a specific aspect of construct validity. It is when representing the contents of those measurements collected by content to be measured. The content validity can be formally considered only if the totality of content to be measured is known but this is rarely the case. It is loosely used for simple tests for in stance, a knowledge test and spelling test. Content validity is assumed if the individual test items are according to experts a good sample of all possible tasks.A test counting for the third School year is valid if the tasks about the subject matter of this age group are well represented. 3. 4 Ecological validity ecologically valid in which the measures introduced by this method S-conditions (S stands for stimulus) an truthful sample of the population of all musical accompaniment conditions of the individual S-conditions are represented. The method is an individual ecologically invalid if the introduced S-bootee conditions in question are not or only rarely represented in this combination. Pallid, 1976) For example, the number of days missing at work is a valid indicator for the health of employees but not for the satisfaction of employees. If they are at work you dont know if they are satisfied or not. The attempt to measure the length of a screw with a measurement. 4. Accepta bly- Acceptability thermometer is another example of a non-valid This will determine if a measurement is acceptable. In other words whether it is ordered with written or unwritten social norms and the investigation of partners and thus accepted as such in an interview. 5. Economic EconomyTime and specie are always scarce goods therefore you have to consider the aspect of economy. It can be said that the one of two measurements is more economical which was achieved with less terms and time. 6. Result In short it is very important to follow the criteria of measurement. If you dont, your result is not valid. That your result is not valid wont be your only problem. You can make a measurement and get a result of your measurement. But the result doesnt represent what you privation to measure. The best way to measure is, to measure with two groups. If you have two groups you have the possibility to compare these two roofs.

No comments:

Post a Comment

Note: Only a member of this blog may post a comment.