Understanding backoffLimit in Kubernetes Job

I’ve created a Cronjob in kubernetes, with job’s backoffLimit defaulting to 6 and pod’s RestartPolicy to Never , the pods are deliberately configured to FAIL. As I understand, (for podSpec with restartPolicy : Never) Job controller will try to create backoffLimit number of pods and then it marks the job as Failed, so, I expected that there would be 6 pods in Error state.

This is the actual Job’s status:

status:
  conditions:
  - lastProbeTime: 2019-02-20T05:11:58Z
    lastTransitionTime: 2019-02-20T05:11:58Z
    message: Job has reached the specified backoff limit
    reason: BackoffLimitExceeded
    status: "True"
    type: Failed
  failed: 5

Why were there only 5 failed pods instead of 6 ? Or is my understanding about backoffLimit in-correct?