Patterns & Themes
Cross-cutting findings from provider type and NQS quality data
These patterns emerge from cross-referencing two data dimensions: provider type (how a service is structured and owned) and NQS regulatory quality rating (set by government assessors). Each pattern includes a signal strength indicator and sample size - use these to judge how much weight to place on each finding.
The proportion of services achieving "Exceeding NQS" - the highest regulatory rating - varies dramatically by provider type. School council and community-run services consistently outperform commercial operators. The pattern is monotonic: not-for-profit types sit at the top, for-profit types at the bottom, with no crossover.
Commercial chains account for 66% of all OSHC services in metro Melbourne - the dominant provider type also has the second-lowest Exceeding rate. The market is structurally skewed: the operators who run most services are the least likely to achieve the highest quality outcome.
Commercial chains operate 803 of 1,212 OSHC services in metro Melbourne - a 66% market share. Despite this dominance, they have a 6.4% Exceeding NQS rate, compared to 21.8% for school council services.
(803 / 1,212 services)
Exceeding NQS rate
Exceeding NQS rate
to exceed than commercial chain
For-profit market share in Australian OSHC grew from ~35% (2014) to ~53% (2020) nationally. In metro Melbourne the concentration is higher still. The structural implication: as commercial chains have grown to dominate, the share of services achieving the highest quality rating has not kept pace.
Before-and-after school care operates in split shifts (~7–9am and ~2:30–6pm), making full-time staffing structurally impractical. As a result, every major OSHC operator runs a casual educator workforce - this is a sector-wide feature, not a differentiating quality signal in itself. What does vary is staff consistency: how often the same educators show up, and how much the provider invests in retaining them.
| Provider | Staffing model | Consistency signal | Key evidence |
|---|---|---|---|
| TeamKids | Direct-employed; retention focus | Strong | 60% internal promotions; never lost a school contract since 2017; $13k retention bonus for coordinators |
| Camp Australia | Direct-employed casual pool | Moderate | 5,000+ staff; casuals may rotate across multiple sites; some staff report inconsistent shift allocation |
| OSHClub (JAG) | Direct-employed mixed (perm + casual) | Moderate | 2,500+ staff across JAG brands; active ongoing casual recruitment; federal wage increase absorbed |
| TheirCare | Geographic casual pool | Concerning | 2.7/5 staff rating (Glassdoor, 65 reviews); high manager turnover; some reviews allege casuals placed in RP role at unfamiliar services |
| Extend | Direct-employed casual | Unknown | Insufficient public data to assess meaningfully |
Importantly, using third-party agency staff for gap-filling does not necessarily indicate lower quality. ANZUK Education - a specialist OSHC recruiter - has published analysis suggesting that services using quality agency staff to manage peak demand can perform at or above the sector average on NQS ratings (ANZUK Education blog, 2024). The risk arises when agency or casual staff become the primary workforce rather than a planned supplement to a stable team.
The NQS overall rating - Exceeding, Meeting, or Working Towards - is a single composite determined by assessors weighing performance across seven quality areas (QA1–QA7). Two services sharing the same headline rating can have strikingly different profiles underneath. A service rated "Meeting NQS" may be Exceeding on educational programming (QA1) while Working Towards on child safety (QA2) - a combination that is invisible in the top-level label but has material implications for families.
This is not a flaw in the NQS framework - the composite rating is designed to give services credit for areas where they excel. But for parents making a targeted decision (e.g., prioritising child safety or educational quality), the overall rating alone provides an incomplete picture. The lens-based quality scores on this site weight quality areas to surface these differences.
QA4 - Staffing arrangements - covers educator-to-child ratios, qualifications, and the continuity of supervision. Unlike other quality areas, QA4 does not describe a discrete domain of practice: it is a foundation that every other quality area depends on. A service cannot reliably deliver high-quality educational programming (QA1), safe supervision (QA2), or meaningful relationships (QA5) without adequate, stable, well-qualified staffing underneath.
The scoring model on this site treats QA4 as cross-cutting: it contributes half-weight to both the Child Safety & Relationships group (QA2, QA4, QA5) and the Learning & Environment group (QA1, QA3, QA4). This design reflects the structural reality: QA4 weakness does not belong exclusively to one domain - it degrades quality across the board. Under the "Safety First" and "Relationships" lenses, QA4 is weighted 1.5× - meaning services with staffing shortfalls are more clearly distinguished from better-staffed peers.