AI & Gig-Work Fairness Policy
Be Good • Be Fair • Be Kind
1. Purpose
To ensure Able's AI systems and marketplace practices embody our core values: Be Good • Be Fair • Be Kind.
2. AI governance
| Transparency |
Plain-English summary of matching algorithm factors is published (weights table). Users can request additional explanation of any automated recommendation. |
| Contestability |
Workers or buyers may trigger human review of a match or decision within 72 hours. |
| Bias audit |
Independent audit of matching scores every quarter from Oct 2025; headline metrics published on our blog. |
| Data minimisation |
Only shift-relevant attributes (skills, availability, location radius < 15 mi) are passed into the model. Sensitive traits (age, gender, ethnicity) are not captured by Able matching agent. |
| Safety & security |
Model inputs/outputs logged, salted and retained 12 months for incident response; penetration testing monthly. |
3. Fair-work standards
| Earnings |
Minimum rate ≥ statutory National Minimum/Living Wage for the role; same-day payout (Stripe Express) at 100% of agreed hours. |
| Autonomy |
Freelancers set their own rates, availability and max travel distance; no exclusivity clauses. |
| Transparency |
All fees displayed to both parties before booking; Able's 8% + VAT buyer fee is fixed and published. |
| Representation |
Community channel plus quarterly round-tables with the CEO & Community Manager. |
| Well-being |
System flags burnout risk (≥40 hrs in rolling 7 days); prompts rest and will throttle offers. |
| Dispute resolution |
Escalation path: Chat → Community Manager → Independent Arbitrator (CEDR model rules) within 28 days. |
4. Reputation & Feedback (No Ratings)
| No public star-ratings |
Able does not display 1-to-5 stars, streaks or productification reputation metrics for workers or buyers. We consider such numerical scoring systems to reinforce confirmation bias, cause stress and can entrench discrimination. |
| Context-rich feedback instead |
After each gig both parties are invited to give structured feedback via short free-text prompts ("What went well?" / "What could be improved?"). This is visible only to the counter-party unless they chose to make it public and Able's Trust & Safety Team. |
| AI-assisted summarisation – human-reviewed |
Where many gigs occur, a natural-language summary is produced to help users surface common themes; a human moderator checks for tone and inadvertent disclosure of personal data before release. These summaries can be contested via our discord channel and community manager. |
| Use of feedback in matching |
Qualitative insights may be converted into a binary 'trusted partner' flag (visible only to Able's internal systems) after manual verification. The flag can increase the weighting of future matches but never blocks access to gigs; any negative pattern triggers human review. |
| Right to challenge |
Workers and buyers can request deletion or redaction of feedback they consider unfair or inaccurate. Disputes follow the escalation path set out in §3 "Dispute resolution". |
5. Monitoring & reporting
Key indicators (fill-rate, cancellation rate, wage compliance, bias metrics) are reported on a public Fairness Dashboard updated monthly.