In my lab at work, all bought this way. I got them for between $899 and $1050. Right now they are $1100
Now some caveats. This is if you are doing a lot of training/fine-tuning of the model. This card has 24GB, which is rare for the consumer-grade cards. It also has the ampere engine with can do the mixed-fp16 computations, which allows you to fit bigger models and bigger data batches on the card.
Because the tutorial uses YOLO and not a network with memory (like a photo RNN or LSTM), it means you can feed the pictures in in any order, even during inference, without a problem. This is where a card with a lot of memory could shine. You would push over a lot of frames, let it run, then pull back the result.
The other thing you have to worry about during a training/fine-tuning is that the k-means for the auto generation of the training set is likely is done by CPU and not by the GPU. That could be a bottleneck for a large dataset. So even if it trains well and the inference/deployment is fast on the GPU, generating the training set could take a very long time.