Join 1M+ other developers and:
- Get help and share knowledge in Q&A
- Subscribe to topics of interest
- Get courses & tools that help you grow as a developer or small business owner
Does the linux "time" command measure user time (not real time or system time) correctly when using vCPUs (ie. on DigitalOcean)?
We’re deploying a CMS instance (contest management system, Github repo link). To accurately measure whether or not a given solution to a task is fast enough, the system would need to accurately measure time even when the CPU might be used by another droplet for a short time in the middle of the execution of a solution. So, I’d like to know whether or not the fact that droplets run on virtual CPU cores affects the linux “time” command.