1)UNIX operating system uses Round robin Time slicing with multilevel feed back.

Assume that there are 10 processes waiting in a queue which is implemented as a linked list of PCB’s ( process control blocks). Assume the PCB have information about Process ID , CPU burst time required , amount of memory being used .

Assume the time slice is 2 units. Simulate Round robin Time slicing until all the jobs complete and find average waiting time. Modify your program to include random arrival of jobs with a fixed burst time required and find the average waiting time of the jobs completed over a simulation time of 100 units.

I really need your help to do this in C or C++ for my assignment.Due date is on Monday. Any example for me to refer?I desperately need the solution. Anyone just help me please..I don't want to fail the subject.


It seems like your highlighted text would make some good Googling. STFW first. Follow up here to try to get questions to specific problems related to whichever language you choose.