Editorial Methodology
How ContractorHQ Reviews Software
We don't test software in a marketing demo. We test it the way it actually gets used — in the field, under pressure, by technicians who need it to work when the signal drops, the customer is watching, and the next job is already overdue.
Bernard Guido — Lead Reviewer
Master tradesman · CRM developer · 25+ years in field operations
Why most software reviews are useless for contractors
The typical contractor software review is written by a tech journalist who signed up for a free trial, clicked through the onboarding wizard, and listed the features from the vendor's own marketing page. They have never dispatched an emergency call. They have never had a tech call in sick at 6am and needed to reassign eight jobs before the first customer woke up. They have never watched an invoice app freeze in a client's driveway while the customer is standing there waiting.
ContractorHQ reviews are written by someone who has been on all three sides of this problem: the contractor who needed the software to work, the dispatcher who relied on it under pressure, and the developer who built a replacement from scratch when the existing tools failed. That is not a credential you can fake, and it is not a perspective any aggregator site can replicate.
1,000+ jobs dispatched
through proprietary CRM
5+ years
building FSM software from scratch
50+ partner contractors
coordinated simultaneously
The Framework
8 Field-Test Criteria
Every software reviewed on ContractorHQ is evaluated against these eight criteria. Each one was developed from real operational failure points — things that actually cost contractors time or money in the field.
Offline Reliability
CriticalField techs work in basements, tunnels, and rural sites with zero signal. We test every app's offline mode under real disconnection — not just by switching airplane mode in a coffee shop. If invoices, job notes, or dispatch updates are lost after reconnection, that is a hard fail.
Fat-Finger UI Test
CriticalCan a technician complete a job closure, capture a signature, and issue an invoice while wearing work gloves in a poorly lit van? We evaluate every key workflow for tap target size, error recovery, and one-hand usability. Software that requires precise tapping fails real field conditions.
Dispatch Latency
HighWe simulate emergency rescheduling: reassigning 10+ jobs simultaneously under time pressure. We measure how many taps it takes, whether conflicts are flagged automatically, and whether notification to the field tech is instant. This is the #1 differentiator between tools that help dispatchers and tools that slow them down.
Pricebook Depth & Flat-Rate Accuracy
HighGeneric pricebooks are useless. We verify whether each tool includes trade-specific flat-rate codes for HVAC, plumbing, electrical, and roofing — not just hourly billing templates. We test whether custom pricebook entries survive app updates and sync correctly across the team.
Multi-Crew Coordination
HighRunning 3 HVAC crews simultaneously is fundamentally different from running 1. We evaluate how each tool handles crew assignment conflicts, overlapping job windows, and real-time status updates when three crews are moving between jobs at the same time.
Data Portability & Exit Rights
CriticalCan you actually get your own data back if you decide to leave? We review the full export options for customer records, job history, invoices, and financial data. We flag any tool that buries export in support tickets, charges for data access, or omits key fields from exports.
True Cost of Ownership
HighWe calculate the full 12-month cost including: base plan, per-user fees, required add-ons (GPS tracking, marketing, pricebook modules), onboarding fees, and typical implementation timeline. The advertised starting price is rarely what a functional deployment actually costs.
Onboarding ROI Timeline
MediumMany contractors pay full monthly fees for 6–12 months before their team is actually using the software effectively. We document the typical 'dead zone' between signup and actual ROI, including training hours required, data migration complexity, and typical time-to-first-invoice on the new platform.
How a review is produced
Trial account + full setup
We sign up for a real trial account and complete the full onboarding process — including data import, team setup, and pricebook configuration. We do not use vendor-provided demo environments.
Field scenario simulation
We run each software through a scripted set of real-world scenarios: emergency rescheduling, offline job completion, on-site payment, and customer dispute handling. We document every friction point.
Practitioner interview corroboration
We cross-reference our findings with feedback from active contractors in Reddit communities (r/ProHVACR, r/FieldService) and direct interviews. If 40 plumbers say a specific feature is broken, that is in the review.
Pricing verification
We verify all pricing directly from the vendor's pricing page and note every add-on that is required for the features described. We calculate the true 12-month cost for a 3-person, 5-person, and 10-person team.
Architecture assessment
As a CRM developer, Bernard evaluates the software's API availability, data export completeness, integration reliability, and mobile architecture. These are things only a developer can meaningfully assess.
Verdict and update cycle
Every review receives a final 'Contractor's Verdict' written in plain English. Reviews are updated when pricing changes, major features are added or removed, or when community feedback indicates the review is no longer accurate. Target: every 30 days.
Our relationship with software vendors
ContractorHQ earns commissions when you purchase software through our links. This does not affect our rankings or the content of our reviews. Software is never ranked higher because a vendor pays us more — rankings are determined by the 8 field-test criteria above.
We have declined to continue promoting software that we found to be unreliable in field testing, even when those programs offer higher commissions. We will not recommend software that we would not use ourselves.
See the methodology in action
Read how these criteria play out in our most detailed reviews.