Some visualization fields are referenced that don't exist in the actual data in AWS integration. For example in the "Cloudwatch Lambda Errors Top5" visualization, it is referencing the field aws.lambda.metrics.Errors, but this field doesn't exist. We should be using aws.lambda.metrics.Errors.avg instead. Similarly, we are using aws.lambda.metrics.Throttles in the dashboard as well.
Similar problem: for elb,aws.elb.metrics.RequestCount in the vis vs aws.elb.metrics.RequestCount.sum in the documentation and in the index template. Cloudwatch Memory Available [Metrics AWS] is using multiple fields such as aws.ecs.metrics.MemoryUlitization which does not exist. Also same problem with the fields starting with aws.metworkfirewall.metrics.

Some visualization fields are referenced that don't exist in the actual data in AWS integration. For example in the "Cloudwatch Lambda Errors Top5" visualization, it is referencing the field
aws.lambda.metrics.Errors, but this field doesn't exist. We should be usingaws.lambda.metrics.Errors.avginstead. Similarly, we are usingaws.lambda.metrics.Throttlesin the dashboard as well.Similar problem: for elb,
aws.elb.metrics.RequestCountin the vis vsaws.elb.metrics.RequestCount.sumin the documentation and in the index template.Cloudwatch Memory Available [Metrics AWS]is using multiple fields such asaws.ecs.metrics.MemoryUlitizationwhich does not exist. Also same problem with the fields starting withaws.metworkfirewall.metrics.