Hi evreybody can anyone help me to solve this problem in introduction of algorithms ( Suppose program A has time complexity θ(f(n)) and program B has time complexity θ(g(n)) where f = o(g). What can you say about the real-world running time of program A versus program B?)
plz im waiting 4 u r answers.