This question evaluates a candidate's ability to perform asymptotic performance analysis and reason about scalability trade-offs between parallel computation and aggregate merge costs.
You are splitting a dataset of size n across k identical machines.
Total runtime: T(k) = O(n/k + k^2)
Which value of k approximately minimizes the total runtime? Pick ONE option:
Login required