Q: 5
A financial company receives a high volume of real-time market data streams from an external
provider. The streams consist of thousands of JSON records every second.
The company needs to implement a scalable solution on AWS to identify anomalous data points.
Which solution will meet these requirements with the LEAST operational overhead?
Options
Discussion
Its D if you batch, but since you need real-time and as little ops work as possible, option A is built for this. The built-in RANDOM_CUT_FOREST in Flink means no custom ML or extra infra. Pretty sure that's what AWS wants here.
Option A seems right, but if "least operational overhead" means you can't do model customization at all, would B ever make sense?
Be respectful. No spam.