At that point I might be complaining to your cloud provider? (Although how Amazon is with that stuff I have no idea)...
Amazon is terrible for things like this. I believe you can pay for the privilege of speaking with a tech, but I have never in a decade had any real live help from an amazon employee. They sometimes chime in on the forums, but it's rare.
And yes it's my understanding that these EC2 instances are timeslices running on a hypervisor on some actual "box" -- although that box might be some kind of awesome hardware like a mainframe or something I do not know. AWS does offer specs on their EC2 instances, but I don't really know what tools or techniques might be used to test these promised specs -- especially not on a production system.
Given that I don't lack memory and that the CPU usage seems fine, and that the apache log doesn't complain much about MaxRequestWorkers (or anything else really) is there some other resource or latency problem? Is there some utility/command/file that i might check to definitely identify and document the problem? Possible candidates that occur to me are database latencies or network congestion, disks being maxed out, etc.
As for "fullstatus", I have this in my shell resource file for some of our servers
How do you use this information? If you're capturing it via cron, do you somehow rotate it? One of the problems I have is that these periods of slowness often happen while I'm busy with something else. Seems to me that you'd need some kind of file rotation scheme. Looking at the output of mine, the first bit of output seems helpful for determining network usage:
Current Time: Friday, 28-Sep-2018 21:59:49 UTC
Restart Time: Tuesday, 25-Sep-2018 18:53:19 UTC
Parent Server Config. Generation: 1
Parent Server MPM Generation: 0
Server uptime: 3 days 3 hours 6 minutes 29 seconds
Server load: 0.04 0.05 0.07
Total accesses: 4939977 - Total Traffic: 12.9 GB
CPU Usage: u9.08 s4.58 cu0 cs.03 - .00506% CPU load
18.3 requests/sec - 50.0 kB/second - 2800 B/request
18 requests currently being processed, 16 idle workers
This just jogged a thought loose: my network input exceeding my network output might be related to the fact that db traffic is also on the network. That said, 12.9GB over 3 days hardly seems to approach any network limits.