The problem is no longer whether warning signs exist. The problem is whether departments will adopt a noncompensatory, tiered, evidence-based screening architecture that forces those warning signs to matter.
Core Thesis
The next phase of police hiring reform cannot stop at proving that the current model fails. It must replace that model with a new architecture. The evidence already discussed shows that agencies possess predictive warning signs before appointment, yet use them weakly, inconsistently, and too often through subjective, late-stage review. The solution is therefore structural: police hiring must move from a discretionary, clinician-mediated suitability model to a noncompensatory, behavior-specific screening framework that begins earlier, uses validated indicators, separates common screening signals from severe red flags, constrains decision-makers through structured rules, and reserves clinical review for a bounded post-offer role rather than treating it as the system’s primary predictive engine. Pages 11 through 13 of the current deck frame that transition directly: from “current flawed architecture” to “proposed objective architecture,” from “generalized impression” to “validated prediction,” and from decentralized discretion to criterion-linked, actuarial decision rules.
The reason this follow-up matters is that reform often collapses after diagnosis. Institutions acknowledge that the existing process is weak, subjective, and too permissive, yet preserve its basic logic by adding more review, more language of professionalism, or more procedural complexity. That is not reform. That is continuity under new branding. A serious solution requires the opposite. It requires moving prediction upstream through standardized background questionnaires and empirical indicator screening; differentiating applicant risk through a tiered framework that separates population-level indicators from critical red flags; reducing discretionary latitude through structured cutoffs and auditable reasoning; and building a defensible record showing why a candidate was advanced or screened out. The point is not to eliminate judgment. It is to bind judgment to evidence. The point is not to create a harsher version of the same model. It is to create a different one.
This follow-up therefore advances a direct institutional claim: the real reform question is no longer whether warning signs exist, but whether departments will adopt a hiring system designed to force those warning signs to matter. A profession that delegates coercive public authority cannot continue treating validated risk as one more topic for discussion inside an unstructured “whole person” narrative. It needs predictive gates, not generalized impressions. It needs noncompensatory rules for severe indicators, systematic treatment of high-prevalence predictors, and a transparent audit trail that makes hiring logic visible to the institution and defensible to the public. Anything less leaves the current failure intact: warning signs gathered, warning signs softened, and warning signs rediscovered later as complaint, lawsuit, scandal, or termination.
Executive Summary
The first thought-piece established the diagnosis. Police departments do not primarily suffer from an absence of warning signs. They suffer from the institutional misuse of them. Predictive prehire indicators exist, many significantly relate to later misconduct, and agencies have often used those indicators only minimally in hiring decisions. The natural next question is therefore not whether the current model is inadequate. It is what should replace it. This follow-up answers that question directly. Its focus is the proposed solution: a transition from discretionary, subjective, clinician-mediated hiring to an objective, tiered, and behavior-specific screening system built around predictive discipline. The deck accompanying the original piece already sketches that solution in concrete terms. On page 11, it contrasts the “current flawed architecture,” which over-relies on a late, expensive clinical psychological evaluation, with a “proposed objective architecture” built from standardized background questionnaires, empirical indicator screening, background verification, and bounded clinical review. On page 13, it sharpens that contrast further by opposing the “clinical paradigm” of generalized impression and subjective interviews to an “actuarial paradigm” of validated prediction, actuarial decision rules, and criterion-linked screening.
The central premise of the follow-up is that no serious reform can remain compensatory. A system that allows severe red flags to be explained away by “whole person” balancing is the same system that created the current problem. That is why the new model must be noncompensatory. Some indicators should operate as systematic screening triggers because they are common and stable enough to support population-level exclusion pressure. Others should function as critical red flags because their severity and liability implications are too serious to be diluted by narrative context. Page 6 of the deck frames this point clearly in “The Predictive Risk Matrix,” distinguishing Tier 1 “Population Screening” from Tier 2 “Critical Red Flags,” and warning that a system which treats all adverse facts as merely “something to discuss” dissolves meaningful distinctions back into unstructured discretion. The follow-up should build directly on that insight. It should argue that the proposed solution is not simply more information or more caution. It is the adoption of a decision structure in which some facts must carry exclusionary force.
That new structure has several components, all already visible in the existing materials. First, prediction must move upstream. Page 11 makes the sequencing point: validated behavioral signals should be collected and assessed before the process reaches the high-cost, late-stage clinical screen. Second, the system must operate at the item level rather than through broad suitability narratives. Page 12’s “Blueprint for an Objective Model” identifies item-level specificity, tiering and calibration, and structured verification as core features. Third, discretion must be constrained rather than celebrated. The same page calls for reduced discretionary latitude, transparency in reasoning, and a defensible audit trail. Fourth, professional roles must be reallocated. The predictive work should be done by an evidence-based personnel-selection structure, while clinical evaluators retain only a bounded post-offer role. Finally, the system must become publicly legible. A screening decision should be capable of explanation in concrete terms: which indicator triggered concern, how that indicator is classified, and why the candidate advanced or was screened out. That is what makes the model objective rather than merely more elaborate.
The broader significance of the solution is civic, not merely administrative. The problem in police hiring has never been just an HR problem. Page 14 states this directly by framing the issue as “A Civic Crisis, Not Just an HR Issue,” and by emphasizing that delegating the state’s monopoly on force demands the highest level of predictive discipline, not discretionary leeway. That is the governing principle of the follow-up. The public does not bear the cost of weak police hiring in abstract organizational terms. It bears that cost through avoidable force, constitutional violations, sexual misconduct, false arrests, taxpayer-funded settlements, and degraded institutional legitimacy. A solution piece must therefore make clear that rebuilding police screening is not a matter of managerial optimization. It is a condition of lawful and responsible public power. The transition from impression to infrastructure is not optional modernization. It is the minimum reform required once the profession can no longer deny that the warning signs are visible, measurable, and predictive.
I. The Reform Mistake After Diagnosis
The first institutional mistake in police hiring is the one your earlier thought-piece already identified: departments continue to rely on a model that mistakes subjective clinical impression for predictive discipline. The second mistake is what typically happens after that critique lands. Institutions acknowledge that the current system is weak, imprecise, discretionary, or outdated, and then proceed to preserve its basic architecture with new language, new training materials, and new assurances that the process will now be administered more carefully. That is not a reform of structure. It is a reform of description. The underlying decision model remains the same: broad evaluative discretion, late-stage reliance on generalized suitability review, and an institutional preference for discussing warning signs rather than forcing them to carry predetermined consequences. If that architecture survives, the outcome will survive with it. The public may hear more disciplined terminology. It will not receive more disciplined screening. That is the central reform mistake after diagnosis, and it is already embedded in the logic of the slide deck itself. Pages 3 and 4, taken together, establish the precise problem. Page 3 identifies “The Evidence Deficit: The Data We Have vs. The Data We Use,” while page 4 labels the current model “The Category Error of Modern Police Hiring.” Those two slides do not merely criticize isolated practices. They expose a system that has information, understands in broad terms that the information matters, and nevertheless routes it through the wrong mechanism.
That distinction matters because institutions are often far more comfortable with critique than with redesign. A department can admit that hiring standards vary. It can admit that screening tools are imperfect. It can even admit that more evidence-based practices would be helpful. What it resists is the conclusion that the existing architecture must be displaced. Yet the deck makes clear that the current architecture is not failing because it lacks sincerity. It is failing because it relies on the wrong decision structure. Page 4 presents the contrast directly. On the left, “Clinical Impression” is described as the current state, a subjective approach that relies on late-stage clinical psychology to form generalized “whole person” narratives and flattens critical red flags into broad suitability judgments. On the right, “Actuarial Prediction” is presented as the required state, an objective approach that relies on personnel psychology to identify behavior-specific, empirically validated misconduct risks and bind human judgment to structured evidence. That is not a modest refinement. It is a change in governing logic. The institutional failure begins when departments read that contrast and assume they can keep the left-hand structure while borrowing some of the language from the right.
The slide deck does not support that softer reading. Page 3 shows why. The problem is not that agencies lack data about applicant risk. The slide states that 8,539 candidates were screened across more than 150 agencies, that 6,075 were hired and tracked over five years, and that 15 of 19 pre-hire misbehavior indicators significantly predicted later misconduct, with hazard ratios reaching up to 14.59. Yet the same slide identifies the institutional failure in stark terms: disclosing prior misconduct reduced a candidate’s hiring chances by only about 5%, and the graphic explicitly portrays that result as “95% Hired Despite Red Flags” and “5% Screened Out.” The bottom line on the slide is blunt: the profession does not suffer from an absence of information; it suffers from the institutional misuse of it. That sentence matters because it rules out the most common bureaucratic evasion. Departments can no longer plausibly say the problem is uncertainty, incomplete knowledge, or the inherent difficulty of prediction. The deck’s own diagnosis is that the information exists and is not being operationalized. Once that is established, the reform question changes. The issue is no longer how to improve the tone of the review process. It is how to rebuild the process so that known warning signs stop being administratively negotiable.
That is precisely where institutions often retreat. They accept the criticism in abstract but preserve the same discretionary format in practice. They leave red flags inside a qualitative “whole person” assessment. They continue to treat severe signals as something to be explained rather than something that should alter the candidate’s status. They keep the late-stage evaluative interview as the prestige center of the process. They promise better calibration while refusing to adopt structured exclusion rules. In effect, they convert diagnosis into improved rhetoric instead of improved design. The deck’s “category error” slide is especially useful here because it captures the difference between what institutions say they are doing and what they are actually doing. The left side of page 4 lists words like “maturity,” “judgment,” “concern,” and “fit” floating inside the cloud of clinical impression. That image is not accidental. It visualizes a model in which risks are translated into diffuse evaluative vocabulary. The problem with that vocabulary is not simply that it is soft. The problem is that it is structurally compensatory. Once every adverse fact is transformed into a generalized concern about overall suitability, every fact becomes easier to soften, contextualize, offset, or ignore. The very act of translating behavior-specific evidence into broad clinical language drains the predictive force from the warning sign.
This is why the follow-up cannot merely say that police hiring needs “better standards.” That phrase is too easy for institutions to absorb without real change. Better standards can mean more guidance layered onto the same discretionary process. Better standards can mean more careful documentation of the same clinical interview. Better standards can mean stronger admonitions to evaluators while leaving the candidate’s status dependent on generalized professional impression. None of that is enough if the architecture remains intact. The slide deck is stronger than that. It argues, by structure and by contrast, that the reform required is a transition from one paradigm to another. Page 13 makes this explicit by naming the shift. It contrasts “The Clinical Paradigm (Current Model)” with “The Actuarial Paradigm (Objective Model).” The clinical model’s core mechanism is “Generalized Impression,” its primary tool is “Subjective Clinical Interviews,” and its treatment of risk is “Flattened into ‘Overall Suitability.’” The objective model replaces those with “Validated Prediction,” “Actuarial Decision Rules,” and risk treatment “Tiered by Prevalence & Severity.” The significance of that slide is that it denies the idea that the current model can be rescued through incremental adjustment alone. It identifies two different governing systems. The reform mistake after diagnosis is failing to understand that difference and assuming that one can preserve the old system by speaking in the vocabulary of the new one.
The practical consequence of that mistake is predictable. If the process remains discretionary, late-stage, and compensatory, then severe indicators will continue to be absorbed into narrative balancing. If the process remains grounded in “overall suitability,” then departments will continue to treat evidence as advisory rather than binding. If clinical review remains the central prestige mechanism, then predictive information will continue to be subordinated to generalized professional judgment. The result is not mysterious. The slide deck already shows it. Red flags are gathered, discussed, and ultimately diluted. Candidates with meaningful warning signs continue to enter the institution. The department later responds to the downstream scandal as though the real failure began after the appointment. In reality, the architecture of that later failure was already preserved at the point of hire. The problem is not simply that institutions are slow to act. It is that they preserve a system designed to make early action easy to avoid.
That is why the first section of this follow-up must press a harder point than the original diagnosis. The original piece established that the current model is weak, subjective, and under-predictive. This follow-up must establish that critique alone is institutionally safe unless it is paired with structural displacement. A department can admit almost anything about the current model so long as it is not required to surrender discretion. It can tolerate the language of evidence. It resists the architecture of evidence. The slide deck, read honestly, leaves no room for that halfway position. Page 3 says the profession misuses the data it already has. Page 4 says the profession is using the wrong paradigm. Page 13 says the answer is not a more careful version of the current model, but a different one. The reform mistake after diagnosis, then, is preserving discretionary infrastructure after conceding that discretion is the mechanism through which predictive evidence keeps getting neutralized. Until that changes, every institutional promise of “improvement” will remain largely linguistic. The form of the process may look more sophisticated. The function of the process will remain the same.
II. Move Prediction Upstream
If the first reform mistake is preserving the old architecture after admitting its defects, the second is temporal: institutions continue to place the decisive predictive burden too late in the hiring sequence. The slide deck addresses this point most directly on page 11, “Moving Prediction Upstream,” and that page is one of the most important in the entire follow-up because it converts the critique from abstraction into process design. What it shows is that the current system over-relies on late-stage, high-cost clinical interviews to make generalized “suitability” decisions. The slide’s “Current Flawed Architecture” begins with a conditional offer, proceeds to background, and then places enormous weight on an “Expensive Clinical Psych Eval,” accompanied by a callout explicitly stating that the system over-relies on late-stage, high-cost clinical interviews to make generalized suitability decisions. The proposed architecture, by contrast, does not eliminate review; it reorders it. It begins with standardized background questionnaires, then empirical indicator screening using data-driven cutoffs, followed by background verification, and only then bounded clinical review. That sequence is not a cosmetic administrative preference. It is the heart of the proposed solution. The institution must stop asking the most expensive and least behavior-specific stage to do work that should already have been done earlier, more cheaply, and more objectively.
The sequencing question matters because timing shapes what kind of decision the institution is capable of making. A process that begins with standardized questionnaires and empirical indicator screening starts by asking what is known, what can be measured, and what the evidence says about the significance of that information. A process that reserves its main predictive energy for a late-stage clinical review starts by asking how the applicant presents, how a professional interprets the applicant’s narrative, and whether broad concerns can be reconciled with eventual clearance. Those are fundamentally different orientations. The first treats prediction as a screening task. The second treats prediction as an evaluative conversation. Page 11 sides decisively with the first. It explicitly states that the proposed objective architecture moves validated behavioral signals before the conditional offer and binds clinical review strictly to post-offer pathology checks. That sentence is critical because it tells us not only when the predictive work should happen, but also what clinical review should and should not be doing. Clinical review is not abolished. It is limited. The architecture reallocates its function. That is the difference between reforming the hiring process and merely adding new stages to the old one.
There is a practical reason for this upstream shift, and the deck does not hide it. The current model is expensive and inefficient because it carries too many candidates too far into the process before forcing the institution to confront what should already have been apparent. If validated indicators can be collected early, then high-risk candidates can be identified before the department invests in a stage that is both costlier and less well suited to item-level predictive discrimination. This is one of the places where the slide deck’s solution is especially elegant. It does not argue only from fairness or public safety. It also argues from institutional design. A process that puts prediction first is cheaper, clearer, and harder to manipulate. A process that puts generalized evaluation first is costlier, murkier, and more hospitable to discretionary drift. Departments often defend late-stage review as carefulness. Page 11 reveals that it is often the opposite: an inefficient architecture that postpones the hardest decisions until the point where subjective discretion is greatest.
Moving prediction upstream also changes the meaning of background investigation. In the current model, background can become merely preparatory—a file-building step before the more prestigious clinical or suitability review. In the proposed model, background becomes central because it is one of the earliest sites where validated indicators are collected and organized. The slide’s architecture makes that plain. Standardized background questionnaires come first. Then empirical indicator screening using data-driven cutoffs. Then background verification. The sequencing implies a hierarchy of function. Background is no longer just evidence gathering for a later holistic judgment. It becomes part of the predictive gate itself. That matters because the earlier thought-piece established that the profession does not lack warning signs; it underuses them. Moving prediction upstream is therefore not simply about speed. It is about ensuring that the warning signs matter before they are swallowed by narrative. The earlier the institution applies structured risk logic, the harder it becomes for later-stage professional impression to erase what the data already say.
The architecture on page 11 is strengthened further by page 12’s “Blueprint for an Objective Model.” The first principle listed there is “Temporal Sequencing & Early Collection,” which directs departments to move validated prehire screening upstream to gather predictive indicators early and at lower cost, before conditional offers. That principle is not ancillary. It is the opening feature of the objective model for a reason. Without proper sequencing, every other improvement remains vulnerable to dilution. Departments can have tiered indicators, structured verification, and transparency in reasoning, but if those features are introduced only after the institution has already committed itself to a candidate and made clinical review the prestige center of the process, the design remains unstable. The process must begin where the evidence is strongest and where the institution still has practical freedom to exclude without having to rationalize why a candidate already favored should now be removed. Moving prediction upstream therefore protects the integrity of the rest of the model. It places evidence at the point of maximum preventive utility.
This shift also answers one of the most persistent institutional objections: that hiring requires holistic assessment and that hard judgments cannot responsibly be made early. The deck’s architecture responds by distinguishing between predictive screening and bounded clinical review. It does not deny that some later-stage assessment may be warranted. It denies that such assessment should carry the main predictive burden. A process that starts with validated behavioral indicators is not prematurely rigid. It is properly sequenced. It recognizes that there are different kinds of questions in police hiring and that they belong in different stages. The question of whether an applicant presents behavior-specific indicators of future misconduct risk should come first because it can be asked through standardized instruments and verified through investigation. The question of whether an applicant presents a distinct clinical or pathological concern belongs later and in a narrower role. Page 11 makes that division explicit. The current model confuses the two. The proposed model disentangles them.
There is also a deeper institutional virtue in moving prediction upstream: it weakens the system’s dependence on performative rigor. Departments like processes that look serious. Late-stage clinical interviews, expensive evaluations, and extended review stages create the appearance of caution and professionalism. But the deck’s solution insists that appearance cannot substitute for design. A process can look intensive and still be structurally unserious if it postpones the actual predictive work until after discretion has had multiple opportunities to soften the evidence. By contrast, a process that begins with standardized background questionnaires and empirical screening may look less glamorous, but it is more faithful to the actual problem. The issue is not whether a candidate feels suitable after prolonged review. The issue is whether the institution can identify risk signals that should affect access to the badge before those signals are narratively reframed. Moving prediction upstream is therefore both a technical and symbolic correction. It reorders the process around substance instead of spectacle.
The same lesson appears again on page 13, where the “Actuarial Paradigm” is described as using validated prediction and actuarial decision rules, while the current clinical paradigm relies on subjective interviews and generalized impression. Although page 13 is more comparative than sequential, it reinforces the upstream logic of page 11. Actuarial systems work best when they are positioned early enough to serve as gates rather than late enough to become advisory overlays. If they come too late, they risk becoming one more input into the same generalized balancing process the current model already performs. If they come early, they structure the field on which all later review occurs. That is why the temporal issue is inseparable from the institutional one. Moving prediction upstream is not just about efficiency. It is about authority. It determines whether evidence governs the process or merely participates in it.
For that reason, the solution piece must be clear: prediction belongs at the front end. The current architecture waits too long, spends too much, and asks the wrong stage to do the wrong work. The proposed architecture begins with standardized collection, empirical indicator screening, and verification, and only afterward allows a bounded clinical role. That sequence reflects a more honest understanding of what police hiring actually is. It is not principally an exercise in professional impression. It is a screening function in which certain known signals must be identified and acted upon before the institution hands over authority. The moment departments understand that, the logic of page 11 becomes unavoidable. The task is not to make late-stage review more refined. The task is to stop relying on late-stage review to perform predictive work that should have already been completed upstream.
III. Make the Model Noncompensatory
If the predictive work must move earlier, the next requirement is even harder for institutions to accept: the model must become noncompensatory. This is where the follow-up becomes most forceful, because it is here that the deck’s proposed solution most directly collides with the culture of police hiring as it currently operates. A compensatory model is one in which a serious red flag can be offset by other favorable impressions or by an evaluator’s belief that the candidate is otherwise strong. A noncompensatory model is one in which certain indicators are so severe, so probative, or so institutionally dangerous that they cannot be washed away by “whole person” balancing. Page 6 of the deck, “The Predictive Risk Matrix,” is the clearest statement of this principle. It separates applicant facts into categories based on prevalence and severity and identifies two decisive zones: Tier 1 “Population Screening” and Tier 2 “Critical Red Flags.” The slide’s key teaching is that not all warning signs are equal and that evidence demands a calibrated, tiered response. But the most important line is its “Takeaway”: a system that treats all adverse facts as just “something to discuss” dissolves meaningful distinctions back into unstructured discretion. That sentence is the noncompensatory argument in concentrated form. If everything remains discussable, nothing remains decisive.
The culture of police hiring has long favored the opposite instinct. Departments are habituated to the language of judgment, maturity, context, rehabilitation, and overall suitability. Those concepts are not meaningless, but they become dangerous when used to absorb indicators that should operate as stopping points. Page 6 rejects that culture directly. It identifies Tier 2 “Critical Red Flags” as including domestic violence citations, unjustified use of force, and racially offensive behavior complaints. The action rule attached to that box is uncompromising: “Absolute red flags requiring hard exclusions despite smaller data counts.” That is the language of a noncompensatory system. It means the institution does not get to say that a domestic violence citation is concerning but outweighed by later performance, impressive demeanor, or a favorable interview. It means confirmed unjustified force does not become one factor among many. It means racially offensive conduct complaints are not simply contextualized into a broader character narrative. It means some indicators must stop the process.
This is probably the strongest part of the solution because it addresses the precise mechanism through which the current model fails. The earlier critique established that the profession underuses the warning signs it already has. The reason it underuses them is not only poor timing or inconsistent standards. It is also compensatory logic. Severe facts enter the system, but the system is designed to metabolize them into discussion. By the time the candidate exits the process, the red flag has been softened into an issue of “concern,” “fit,” or “overall judgment.” The predictive force of the fact is thereby neutralized. A noncompensatory model breaks that pattern by refusing to let certain indicators be converted into narrative. It preserves the indicator as an exclusionary event. That is why the “Tier 2” category matters so much. It creates a class of signals whose institutional meaning is fixed enough that they cannot be bargained away through discretionary interpretation.
There is a reason institutions resist this model. Noncompensatory rules reduce flexibility. They prevent departments from hiring a candidate they like despite a fact pattern they would rather minimize. They constrain the evaluator who believes he can see beyond the record. They limit the hiring manager who wants to solve a staffing problem by taking a calculated risk. And they weaken the prestige of the late-stage interview by making clear that some decisions have already been functionally determined by earlier evidence. All of this is precisely why the model is necessary. The problem in police hiring has never been the absence of discretion. It has been the overabundance of it. A system that gives institutions too many opportunities to reinterpret red flags will produce exactly the outcome the earlier thought-piece described: warning signs gathered, acknowledged, and nevertheless hired through. Noncompensatory design is the mechanism that interrupts that cycle.
The slide deck’s phrasing is especially helpful because it does not argue that every adverse fact deserves hard exclusion. Page 6 preserves nuance by distinguishing “Background Noise” from Tier 1 and Tier 2. Minor unrelated indiscretions are explicitly placed in their own zone. That distinction is important because it prevents institutions from caricaturing the noncompensatory argument as a demand for rigid exclusion based on any negative fact. The deck is more sophisticated than that. It says some facts are background noise. Some facts support population screening. Some facts require absolute exclusion. The problem is not that the model is too hard. The problem is that the current model refuses to distinguish among categories with enough discipline. It compensates across domains that should not be treated as morally or predictively interchangeable. Noncompensatory design, then, is not the abandonment of calibration. It is calibration taken seriously.
The force of the Tier 2 category is reinforced by page 5, “Dissecting Pre-Hire Risk,” which breaks applicant risk into four validated domains of behavioral signal: prior occupational trouble, law-enforcement trouble, temper and violence, and irresponsible behaviors. Page 5 ends with a synthesis note: prior behavior is the strongest signal of future behavior when the predictor closely resembles the institutional criterion. That principle strengthens the case for noncompensatory exclusion because some Tier 2 indicators are not just severe; they are highly commensurate with the risks the institution claims to be screening against. A domestic violence citation is not merely bad optics. It is an indicator tied to violence and abuse. Unjustified use of force in a prior law-enforcement setting is not merely adverse history. It is behavior sampled from the very context into which the institution is considering re-entry. A racially offensive behavior complaint is not simply awkward conduct. It is a warning sign about the candidate’s likely interaction with a public he will police under state authority. The closer the match between the indicator and the institutional risk, the stronger the argument that the signal cannot responsibly be compensated away.
This is where the follow-up should state the normative principle clearly: some facts must actually mean stop. That sounds obvious, but police hiring has not consistently behaved as if it were true. The slide deck is implicitly repudiating the institution’s longstanding habit of turning every severe fact into a “conversation.” The result of that culture is not mercy or nuance. It is loss of distinction. It is the collapse of different categories of risk into a single “whole person” pool where the evaluator’s subjective narrative becomes the ultimate authority. Page 6 expressly rejects that collapse. The hard-exclusion language for Tier 2 exists because the model recognizes that some indicators are too consequential to be entrusted to soft balancing. A candidate with such a signal is not simply higher risk in the abstract. The candidate presents a type of risk the institution cannot honestly say it failed to foresee if it proceeds with the hire anyway.
The civic dimension of the noncompensatory model should also be emphasized. This is not ordinary employment. The institution is deciding who will exercise force, restraint, discretionary command, and public credibility. In that context, the refusal to make some red flags decisive is not just an HR choice. It is an allocation of public exposure to foreseeable harm. Page 14 of the deck, though outside the pages the user highlighted for this section, states that delegating the state’s monopoly on force demands the highest level of predictive discipline, not discretionary leeway. That line captures why noncompensatory design is necessary in policing even if other sectors tolerate more compensatory balancing. The public is not protected by a model that says every dangerous fact can still be reconsidered through generalized narrative. The public is protected when the institution is willing to say that some predictors disqualify, period.
So Section III should be unambiguous. The proposed model must be noncompensatory because the current system’s biggest vice is its ability to launder severe risk through broad suitability language. Tier 2 indicators such as domestic violence citations, unjustified use of force, and racially offensive conduct complaints cannot be reduced to “something to discuss.” The system must preserve their force by treating them as hard stops. Without that feature, every other reform will remain vulnerable to the same old pattern: better forms, better interviews, better rhetoric—and the same institutional willingness to hire through what should have ended the candidacy.
IV. Separate Tier 1 from Tier 2
One of the most important reasons police hiring continues to fail is that it has not disciplined itself to distinguish among types of adverse information with sufficient rigor. It has treated applicant risk too often as a single undifferentiated field, to be summarized in a final judgment about whether the person seems suitable overall. Pages 5 and 6 of the slide deck provide the corrective. They show that the institution must separate Tier 1 from Tier 2 because not all warning signs do the same work, carry the same frequency, or require the same institutional response. Page 5 supplies the domains: prior occupational trouble, law-enforcement trouble, temper and violence, and irresponsible behaviors. Page 6 then translates those domains into a risk matrix based on prevalence and severity, distinguishing “Background Noise,” Tier 1 “Population Screening,” and Tier 2 “Critical Red Flags.” The point of reading these pages together is to show that the proposed objective model is not simply more data-driven. It is more discriminating. It insists that departments stop flattening qualitatively different signals into one vague narrative of overall suitability.
Tier 1 exists because some indicators are common enough, stable enough, and broad enough in their predictive value to justify systematic use at the population level. Page 6 identifies prior written reprimands, unfavorable terminations, and bad credit as Tier 1 indicators and pairs them with an action rule stating that they “must trigger systematic, population-level screening exclusions.” That is a powerful formulation because it shows the deck’s model is not limited to severe one-off facts. It also includes higher-frequency indicators that, taken seriously, should materially shape how agencies process applicant pools. These are not necessarily absolute disqualifiers in the same sense as Tier 2 red flags, but they are not casual discussion points either. They are the backbone of the institution’s screening architecture. They enable an agency to impose structured exclusion pressure across a broad applicant population rather than waiting only for spectacular red flags to emerge. In that sense, Tier 1 is how the model becomes administratively serious. It treats common predictive indicators as part of a repeatable and rule-bound system rather than leaving them to variable evaluator instinct.
Tier 2 performs a different function. It captures indicators that may appear less frequently but carry such severe liability implications that they cannot responsibly be handled as supplemental context. Page 6 places domestic violence citations, unjustified use of force, and racially offensive behavior complaints in this category and states that they require “hard exclusions despite smaller data counts.” The importance of this distinction cannot be overstated. A department that does not separate Tier 2 from Tier 1 will either overreact to ordinary indicators or underreact to serious ones. If everything is treated the same, then nothing is calibrated. If everything becomes “one more factor,” then high-severity indicators lose their force. The deck’s matrix prevents that collapse by making the institution specify what kind of problem it is dealing with. Is the indicator prevalent and useful for population-level screening? Or is it rarer but severe enough to require a hard stop? That is the kind of structured differentiation the current model lacks.
Page 5 helps explain why the distinction is necessary. It breaks the applicant’s history into validated domains rather than allowing the institution to process the file as one amorphous character narrative. Prior occupational trouble includes negligence warnings, unfavorable terminations, and repeated job-hopping. Law-enforcement trouble includes prior reprimands, demotions, and unjustified use of force. Temper and violence include documented physical altercations and domestic violence citations. Irresponsible behaviors include bad credit, moving violations, and support arrears. This structure matters because it reveals that adverse facts are not interchangeable. A negligence warning does not function the same way as a domestic violence citation. A prior written reprimand does not function the same way as an unjustified use of force. Yet the current hiring model too often treats them all as pieces of a generalized impression. The Tier 1/Tier 2 distinction is the corrective to that habit. It restores categorical differences that the “whole person” model tends to erase.
The mistake of flattening risk is not only analytical. It is institutional. Once all adverse facts are pooled together under overall suitability, the evaluator gains enormous power to decide how much any particular indicator should matter. That discretion is exactly what the slide deck is trying to discipline. Tier 1 reduces discretion by making some common indicators systematically relevant. Tier 2 reduces discretion even further by making some severe indicators effectively non-negotiable. The distinction therefore serves a dual function. It improves prediction, and it constrains the institution’s habitual tendency to turn structured evidence back into narrative judgment. Without this separation, departments can continue doing what they have long done: treating every file as unique in a way that obscures the existence of repeatable patterns. The more the institution insists every case is singular, the easier it becomes to deny that common decision rules should exist at all. The deck rejects that posture by insisting that prevalence and severity are knowable and should structure response.
This is also why Tier 1 and Tier 2 should not be confused with “minor” and “major” in a purely rhetorical sense. Tier 1 indicators are not trivial. They are the material from which serious screening architecture is built. Because they occur with enough frequency and show stable relations, they allow the institution to create population-level filters rather than improvising case by case. A department that ignores Tier 1 because the indicators are not sensational will miss the broader function of the model. Conversely, a department that treats Tier 2 as simply “more serious Tier 1” will fail to understand why the deck assigns hard exclusions to that category. The distinction is not only one of degree. It is one of institutional use. Tier 1 organizes screening across the applicant population. Tier 2 marks the points at which discretion must stop. Together, they create a model that is both scalable and morally serious.
The current model’s deepest flaw is that it dissolves both categories into a single vague story of overall suitability. That phrase sounds responsible, but it is analytically destructive because it treats all predictive information as ultimately commensurable. Under that logic, a candidate with repeated prior reprimands might be cleared because he interviews well. A candidate with a domestic violence citation might be retained in consideration because other aspects of the file seem favorable. A candidate with bad credit and an unfavorable termination might be viewed as complicated but promising. None of those outcomes is impossible under a subjective model because the model has no durable internal distinctions strong enough to resist discretionary smoothing. Pages 5 and 6 are an answer to that entire way of thinking. They tell the institution to stop asking only whether the person seems suitable overall and start asking what type of indicator is present, how often such indicators appear, how severe they are, and what category of institutional response they trigger.
There is also a broader legitimacy benefit to keeping Tier 1 and Tier 2 separate. A system that cannot explain why some facts generate systematic screening consequences while others generate hard exclusions is difficult to audit and easy to manipulate. By contrast, a system that explicitly distinguishes population-level indicators from critical red flags becomes more transparent both internally and publicly. Page 12’s “Blueprint for an Objective Model” later builds on this by calling for transparency in reasoning and a defensible audit trail, but those features depend on the Tier 1/Tier 2 distinction already being in place. Without it, every explanation collapses back into “the evaluator considered everything.” With it, the institution can say what kind of signal was present and why the signal required the response it did. That is the beginning of accountable hiring.
So Section IV should conclude with a clear proposition: the current model fails because it flattens distinct categories of risk into one vague suitability judgment. Pages 5 and 6 provide the alternative. Tier 1 supports population screening because some indicators are common and stable enough to justify systematic exclusions. Tier 2 supports critical red-flag exclusion because some indicators are so severe that they require hard stops. Not all adverse facts are equal, and a hiring model that refuses to act like that is not nuanced. It is undisciplined. The purpose of separating Tier 1 from Tier 2 is therefore not bureaucratic neatness. It is to force the institution to preserve distinctions that predictive evidence has already established and that public safety can no longer afford to lose.
