AWS Advanced Networking Specialty Exam Tips and Tricks

21 Dec 2018

If you’re planning on taking the AWS Advanced Networking Specialty exam, I’ve compiled a quick list of tips that you may want to remember headed into the exam.

Note: This exam is heavily oriented towards long, detailed, and exhaustive networking questions. You cannot brute-force study your way through this exam. You need a thorough working understanding of Direct Connect and hybrid architecture scenarios and how to solve them.

I've created a bubble graph showing my experiences with all of the AWS certifications.

In my opinion, this exam is by far the hardest out of all AWS certifications, easily surpassing the Big Data Specialty or Solutions Architect Professional certifications for difficulty.

To give you an idea of how hard this exam really is, it’s worth noting that I currently hold all 9 AWS certifications. I had finished 8 certifications before focusing on Advanced Networking, and comfortably passed all 8 attempts. It usually takes me 30-45 minutes to sit for an AWS exam, and I generally feel extremely well-prepared when I sit down for an exam.

The first time I took Advanced Networking - Specialty, I failed with a 60% score. Unlike all of my other certification exams, I knew within the first 10 questions that my chances of passing were slim. I chewed my fingernails to bits as I tried to troubleshoot scenarios that I didn’t even know existed. I walked out of the testing center miserable - I haven’t failed a certification exam since high school (MCSE). I had wanted to get all of the AWS certifications by the end of 2018 - but AWS mandates a two-week pause before you can re-take the exam, so I would be forced to finish my goal in 2019. This stung.

But, so it goes, right? I already had eight AWS exams, and I just needed this last one. I went home and immediately scheduled the exam for two weeks later.

I wrote down all of the things that I didn’t know, and I focused on where I felt the gaps in my knowledge were. I mostly realized that my Direct Connect troubleshooting skills were terrible, and I was not very good at VPC endpoint/interface stuff.

I crammed as hard as I could. Harder than any test I have ever taken. I studied more for this test than I had ever studied in my life. I drew diagrams, flashcards, and mental maps. I had my girlfriend quiz me over and over again on CIDR block properties and subnetting math. It was not a very fun two weeks. I gained weight, avoided social contact, and thought of nothing but passing. I am not used to failing exams, and I was frustrated with myself.

I took the exam for the second time on January 5th, 2019. I passed with a 66% score.

Two weeks of effort and studying had netted me… a 6% boost in score. Luckily, AWS scores on an unpublished curve, so this was sufficient to pass.

I think the exam is so difficult for a few reasons:

  1. It is rare to have practical experience with Direct Connect and all of the networking concerns that accompany it.
  2. The core material is simply very, very dense, and requires great knowledge of a great deal of AWS services and how they interact with each other.
  3. You need strong networking skills in order to troubleshoot tough scenarios.
  4. The breadth of the exam is extreme. If you haven’t taken the Big Data certification, I doubt you will have any knowledge of some of the products that are used in the exam’s scenario questions. You need very, very broad knowledge of AWS services - on top of that, you need to be familiar with technical implementation details with them.

I found it very useful to use pen and paper to re-draw all of the various routing scenarios that are presented throughout AWS Direct Connect/VPN/VIF documentation. You need to become intimately familiar with the flow of DNS requests for different scenarios.

All that being said, I’ve laid out some of my study notes below. Please enjoy!

Scenarios

  • If you need to set Active/Passive mode, use AS_PATH prepending to make certain routes less attractive than others.
    • For example, if you had two DX connections and wanted to configure in an Active/Passive configuration, you should prepend the AS_PATH of whichever DX connection you want to be the Passive.
  • The only way AWS will prefer a VPN over Direct Connect is if the VPN has a more specific route table entry.
Typical layout for on-premise resources that need to resolve DNS requests using your VPC resolver (or Route53 if you forward the requests along).
Credit: ToTheNew.com for the diagram.
Although this diagram may be intimidating, it is crucial that you understand the route tables necessary for this architecture.

DNS

  • Configure your custom DNS servers on EC2 instances in your VPC to route your private DNS queries to the IP address of the Amazon-provided DNS servers for your VPC. This IP address is the IP address at the base of the VPC network range “plus two.” If your VPC CIDR block is 10.0.0.0/16, the IP address of the DNS server is 10.0.0.2. (Source)
  • If you’re using custom DNS servers that are outside of your VPC and you want to use private DNS, you must use custom DNS servers on EC2 instances within your VPC.
  • If you have integrated your on-premises network with one or more VPC virtual networks and you want your on-premises network to resolve domain names in private hosted zones, you can create a Simple AD directory. Simple AD provides IP addresses that you can use to submit DNS queries from your on-premises network to your private hosted zone.
  • You cannot further subnet a /28 CIDR block.
  • DNS servers require TCP/UDP port 53 to be open.
  • For each public hosted zone that you create, Amazon Route 53 automatically creates a name server (NS) record and a start of authority (SOA) record. Don’t change these records.

BGP

  • BGP is typically classified as a path vector protocol.
    • A path vector protocol is a network routing protocol which maintains the path information that gets updated dynamically. (Source)

Routing

  • AS_PATH prepending assists in setting Active/Passive mode.
  • Keep these route preference rules in mind. They are evaluated from the top down.
Remember that low MED values and the shortest AS_PATH are preferred!
Credit: ACloudGuru.com for the slide.
  • Bidirectional Forwarding Detection (BFD) is a network fault detection protocol that provides fast failure detection times, which facilitates faster re-convergence time for dynamic routing protocols. Enable BFD when configuring multiple DX connections or a single DX + VPN backup. You can configure BFD to detect failures and update dynamic routing as Direct Connect quickly terminates BGP peering so that backup routes can kick in. (Source)
  • A forwarding equivalence class (FEC) is a term used in Multi-protocol Label Switching (MPLS) to describe a set of packets with similar or identical characteristics which may be forwarded the same way. An FEC label is used by the router to know where to forward packets in an MPLS network. (Source)
  • DSCP can be used as a means of classifying and managing network traffic and of providing QoS in modern Layer 3 IP networks. (Source)
  • Make sure to allow ephemeral ports in your outbound NACL rules.

VPCs

  • The default VPC can be deleted and recreated.
  • Broadcasting is not allowed by AWS.

Direct Connect

  • AWS will always prefer a DX connection over VPN, unless the VPN has a more specific route.
  • A sub-1Gbps connection is indicative of a hosted connection.
  • A VIF is required for each VPC connected to the DX connection.
  • You cannot change the port speed of a connection after creation.
  • There is a limit of 100 advertised BGP routes. Summarize routes if you have more than this.
  • Study the Direct Connect troubleshooting tips.
Sample Direct Connect configuration
  • A Direct Connect Gateway is a grouping of VGW’s and private VIFs that belong to the same account.
    • Read through this great resource on Direct Connect Gateways.
    • Remember: The VPCs to which you connect through a Direct Connect gateway cannot have overlapping CIDR blocks.
    • DX Gateway is a global service.
    • Cannot be used cross account. Cannot be used with public VIFs.
    • Cannot be used to send traffic to other VPCs that are connected to the same Gateway.
Sample DX Gateway
  • You may be asked how many LOA’s will be generated for a customer request. Remember the formula: 1 LOA per Connection per Datacenter.
    • A Link Aggregation Group (LAG) counts as one LOA.

Direct Connect Requirements

I noticed that the exam had an unusual emphasis on the Direct Connect requirements. Make sure you know these requirements by heart. It’ll be worth a few points on the exam.

  • BGP
  • BGP MD5 authentication
  • 802.1Q VLAN encapsulation
  • Single-mode fiber with a 1000BASE-LX (1310nm) transceiver for 1gb Ethernet or a 10GBASE-LR (1310nm) transceiver for 10gb Ethernet.
  • Auto-negotiation for the port must be disabled.
  • Port speed and full-duplex mode must be configured manually.

VPNs

  • AWS will always prefer a DX connection over VPN, unless the VPN has a more specific route.
  • Use an IPSec VPN over a public VIF to ensure protection of your data en-route from your on-premise datacenter.
  • AWS VPN does not support 128 bit AES encryption.
  • You need to use a public VIF over a DX connection in order to use an IPSec VPN.
  • You can monitor your VPN tunnels using Cloudwatch. (Source)
    • There are three VPN metrics: TunnelState, TunnelDataOut, and TunnelDataIn.
    • Use aws cloudwatch list-metrics --namespace "AWS/DX" to view metrics using the CLI.
  • If the VPN connection is idle for 10 seconds, the connection may close. Use a keep-alive tool to send continuous traffic.
  • Static routes take priority over propagated routes - it’s assumed that you’re using static routes for a good reason.
  • The VPN tunnel comes up when traffic is generated from your side of the Site-to-Site VPN connection. The virtual private gateway is not the initiator; your customer gateway must initiate the tunnels. (Source)
  • Allow UDP 500 (and UDP 4500 if using NAT Traversal).

OSI Layers Cheatsheet

  • Application - DNS, DHCP, BGP, FTP, SMTP
  • Network/Internet - IPSec
  • Transport - TCP/UDP

Protocol/Port Cheatsheet

  • IPSec: IP 50, UDP 500
  • VPN: UDP 500, UDP 4500 for NAT-Traversal
  • DNS: TCP/UDP 53
  • Ephemeral Port Range: 1024-65535

Elastic Load Balancers

  • ELB needs a minimum subnet size of /27 (not /28).
  • ELB SSL Negotiation: A security policy is a combination of:
    • SSL protocols
    • SSL ciphers
    • Server Order Preference option.
  • ELB SSL Negotiation: SSL 2.0 has been deprecated. Use SSL 3.0, TLS 1.0, TLS 1.1, or TLS 1.2.

CIDR Blocks and Subnetting

  • You need to understand CIDR blocks, addresses ranges, and be able to do subnet math in your head on the fly. Use this calculator to get better at subnetting.
  • The minimum size of a VPC subnet is a /28 (16 addresses) for IPv4.
  • The maximum size is a /16 (65,535 addresses).
  • For IPv6, the VPC is a fixed size of /56.
  • Use the 10.0.0.0/8, 172.16.0.0/12 and 192.168.0.0/16 ranges. (Source)
    • For IPv6, the subnet size is fixed to be a /64. Only one IPv6 CIDR block can be allocated to a subnet.
  • The default VPC range is 172.31.0.0/16
  • Remember that 5 addresses are reserved per subnet:
    • 0 - Network address
    • 1 - VPC router
    • 2 - DNS
    • 3 - Reserved by AWS for future use
    • 255 - Broadcast
  • Subnets cannot be larger than the VPC in which they are created.

Enhanced Networking and Performance

  • You can only use Jumbo frames/high MTU values within a VPC - once the traffic leaves the VPC, use 1500 MTU. The only exception to this is VPC peering traffic - it can take advantage of 9001 MTU. (Source)
  • To maximize instance speed for an external/internal facing instance, attach two ENIs:
    • An externally-facing ENI with an MTU of 1,500 bytes
    • An internally-facing ENI with an MTU of 9,001 bytes
  • DPDK is the Data Plane Development Kit. It consists of libraries to accelerate packet processing workloads running on a wide variety of CPU architectures.
  • General Rules of Placement Groups: (Source)
    • The name you specify for a placement group must be unique within your AWS account for the Region.
    • You can’t merge placement groups.
    • An instance can be launched in one placement group at a time; it cannot span multiple placement groups.
    • Instances with a tenancy of host cannot be launched in placement groups.
  • Use spread placement groups when you are aiming for high availability.
  • Use cluster placement groups when high performance is your only concern.

Public and Private VIFs

  • General VIF Creation Requirements:
    • Name
    • Owner
    • Connection
    • VLAN ID
    • BGP Information
      • ASN
      • BGP MD5 key
    • Peer IP’s
    • Public VIFs: Choose which prefixes to advertise
    • Private VIFs: Jumbo Frame setting
  • You really need to understand public vs. private VIF’s, hosted VIF’s, Direct Connect Gateways, IPSec VPN’s, and how all of them interact.
    • What are the differences between public virtual interfaces and private virtual interfaces for Direct Connect?
      • Public Virtual Interface
        • To connect to AWS public endpoints, such as DynamoDB or S3, with dedicated network performance, use a public VIF.
        • Use an IPSec VPN over a public VIF to ensure protection of your data en-route from your on-premise datacenter.
      • Private Virtual Interface
        • To connect to private services, such as a VPC, with dedicated network performance, use a private VIF.
        • A private VIF allows you to connect to your VPC resources on your private IP address or endpoint. A private VIF can connect to a DX gateway, which can be associated with one or more VGWs.
        • A virtual private gateway is associated with a single VPC, so you can connect to multiple VPCs using a private VIF. For a private VIF, AWS only advertises the entire VPC CIDR over the BGP neighbor.
  • Do not advertise more than 100 prefixes for private virtual interfaces or 1,000 prefixes for public virtual interfaces. These are hard limits and cannot be exceeded. (Source)

VPC Endpoints

  • You cannot extend endpoint functionality outside of the VPC - this means you cannot access a VPC endpoint from another location (like a VPC peer, or over a VPN).
  • DNS resolution is required within the VPC.
  • Default VPCE policy is unrestricted. You can lock it down after creation.
  • You can’t reach VPC endpoints over a private VIF.
  • All connections in a LAG run in an Active/Active configuration. LAG is a Layer 2 connection. (Source)
  • Remember the three rules of creating Link Aggregation Groups:
    • All connections in the LAG must use the same bandwidth.
    • You can have a maximum of four connections in a LAG. Each connection in the LAG counts towards your overall connection limit for the Region.
    • All connections in the LAG must terminate at the same AWS Direct Connect endpoint.

Workspaces

  • Remember that AWS Workspaces requires a minimum of 1200 MTU. (Source)
AWS Workspaces also recommends two private subnets and one public subnet for Microsoft AD or Simple AD

Route53

  • Use Route53 to configure split-view DNS, also known as “split-horizon DNS”. This feature can be used when you want to maintain internal and external versions of the same website or application. (Source)
  • Create a Reusable Delegation Set using the Route 53 API command CreateReusableDelegationSet (you can’t do this through the AWS console) if you want to re-use nameservers across different hosted zones.

Cloudhub

  • The hub and spoke model involves creating multiple Customer Gateways, each with a public IP address.
  • The remote network prefixes for each spoke must have unique ASNs, and the sites must not have overlapping IP ranges.

Cloudwatch

  • Use Lambda@Edge in conjunction with Cloudwatch to perform actions based on incoming requests. For example:
    • A Lambda function can inspect cookies and rewrite URLs so that users see different versions of a site for A/B testing.
    • CloudFront can return different objects to viewers based on the device they’re using by checking the User-Agent header, which includes information about the devices.
    • On a retail website that sells clothing, if you use cookies to indicate which color a user chose for a jacket, a Lambda function can change the request so that CloudFront returns the image of a jacket in the selected color.
    • Generate HTTP responses when CloudFront viewer request or origin request events occur.
    • inspect headers or authorization tokens, and insert a header to control access to your content before CloudFront forwards the request to your origin.
    • Make network calls to external resources to confirm user credentials, or fetch additional content to customize a response.

Security Groups and Network Access Control Lists

  • In case of a DDOS, the fastest way to stop incoming requests is to delete the default NACL in a VPC.
  • You cannot apply Security Groups to NAT Gateways or Egress-Only Gateways (which makes sense, because you can’t set outgoing rules with Security Groups anyways!)

Misc

  • Know your basic service limits:
    • Max. 5 Internet Gateways per region
    • Direct Connect: 100 BGP advertised routes per route table maximum
  • You need to assign an Elastic IP to NAT Gateways.
  • Use tracepath to check MTU between two hosts. tracepath needs UDP traffic allowed in order to function.
  • A hosted VIF is a VIF created on your Direct Connect connection, and shared with another AWS account.
  • By contrast, a hosted connection is a sub-1gbps connection allocated by an AWS partner for your use.

Training Materials I Used

Whitepapers I Read

Videos I Watched


AWS Big Data Specialty Exam Tips and Tricks

05 Dec 2018

If you’re planning on taking the AWS Big Data Specialty exam, I’ve compiled a quick list of tips that you may want to remember headed into the exam.

I passed the exam on December 6, 2018 with a score of 76%. In my opinion, this exam is more difficult than the AWS Solutions Architect Pro!

  • You really, really need to understand Redshift distribution strategies. Here are some things to remember:
    • Automatic Distribution: The default option, Redshift automatically manages your distribution strategy for you, shifting from an initial ALL strategy (for smaller tables) to EVEN distribution (for larger tables). Note: Redshift will not automatically switch back from EVEN to ALL.
    • Even Distribution: With the EVEN distribution, the leader node distributes rows equally across all slices. This is appropriate for tables that do not participate in joining.
    • Key Distribution: With the KEY distribution, rows are distributed according to a selected column. Tables that share common join keys are physically co-located for performance.
    • All Distribution: A copy of the entire data set is stored on each node. This slows down inserting, updating, and querying. This distribution method is only appropriate for small or rarely-updated data sets.
  • You need to know the DynamoDB partition sizing formula by heart: (Desired RCU/3000 RCU) + (Desired WCU/1000 RCU) = # of partitions needed
  • AWS Machine Learning does not support unsupervised learning - you will need Apache Spark or Spark MLLib for real-time anomaly detection.
  • AWS IoT accepts four forms of identity verification: X.509 certificates, IAM users/roles, Cognito identities, and Federated identities. “Typically, AWS IoT devices use X.509 certificates, while mobile applications use Amazon Cognito identities. Web and desktop applications use IAM or federated identities. CLI commands use IAM.”
  • In the context of evaluating a Redshift query plan, DS_DIST_NONE and DS_DIST_ALL_NONE are good. They indicate that no distribution was required for that step because all of the joins are co-located.
  • DS_DIST_INNER means that the step will probably have a relatively high cost because the inner table is being redistributed to the nodes. DS_DIST_ALL_INNER, DS_BCAST_INNER and DS_DIST_BOTH are not good. (Source)
  • You must disable cross-region snapshots for Redshift before changing the encryption type of the cluster.
  • Amazon recommends allocating three dedicated master nodes for each production ElasticSearch domain.
  • Read up on DAX and DynamoDB.
In most cases, the DynamoDB response times can be measured in single-digit milliseconds. However, for use cases that require response times in microseconds, DynamoDB Accelerator (DAX) delivers fast response times for accessing eventually consistent data.
  • After a Redshift load operation is complete, query the STL_LOAD_COMMITS table to verify that the expected files were loaded.
  • Mahout is a machine learning library with tools for clustering, classification, and several types of recommenders, including tools to calculate most-similar items or build item recommendations for users. Use it to carry out Machine Learning work on top of Hadoop.
  • When preparing to use a Lambda/Kinesis combination, make sure to optimize your Lambda memory and batch size, and adjust the number of shards used by the Kinesis streams.
  • Triggers do not exist in Redshift.
  • Amazon Kinesis Aggregators is a Java framework that enables the automatic creation of real-time aggregated time series data from Kinesis streams. (Source)
  • You can’t encrypt an existing DynamoDB table. You need to create a new, encrypted table and transfer your data over. (Source)
  • Presto is a fast SQL query engine designed for interactive analytic queries over large datasets from multiple sources. (Source)
  • Redshift creates the following log types:
    • Connection log — logs authentication attempts, and connections and disconnections.
    • User log — logs information about changes to database user definitions.
    • User activity log — logs each query before it is run on the database.
  • Kinesis Data Firehose can send records to S3, Redshift, or Elasticsearch. It cannot send records to DynamoDB. (Source)

A serverless solution to ingest, query, and visualize Big Data using Kinesis, Glue, Athena, and QuickSight.
  • Use Spark for general purpose Amazon EMR operations, use Presto for interactive queries, and use Hive for batch operations.
  • Use Athena generally to query existing data in S3. You should be aware that Redshift Spectrum exists, and that it can query data in S3.
  • If a question is asking how to handle joins or manipulations on millions of rows in DynamoDB, there’s a good chance that EMR with Hive is the answer.
  • When using Spark, you should aim for a memory-optimized instance type.
  • To improve query performance and reduce cost, AWS recommends partitioning data used for Athena, and storing your data in Apache Parquet or ORC form - not .csv!
  • Use the COPY command to transfer data from DynamoDB to Redshift. Use UNLOAD to transfer the results of a query from Redshift to S3.
  • Redshift clusters have two types of nodes: Leader nodes and Compute nodes.
Redshift Architecture: If a Leader node goes down, the cluster health will suffer.
  • Not to be confused with EMR, which uses Master, Core, and Task nodes.
EMR Node Structure
EMR Architecture: EMR stores log files on the Master node by default.
  • Use ElasticSearch to analyze data stream updates from other services, such as Kinesis Streams and DynamoDB.
  • Amazon Schema Conversion Tool is sometimes referred to as SCT.
Review EMR security and encryption

Training Materials I Used

Whitepapers I Read


AWS DevOps Engineer Professional Exam Tips and Tricks

30 Nov 2018

If you’re planning on taking the AWS DevOps Engineer Professional exam, I’ve compiled a quick list of tips that you may want to remember headed into the exam.

I passed the exam on November 30th, 2018. Before taking this exam, I had already achieved my SA Pro. There is a fair amount of overlap between the two tests - I would highly recommend taking the SA Pro before this exam. I scored an 85% on the exam.

This exam is fairly difficult, although I would argue it’s slightly easier than the Solutions Architect Professional exam. There is a greater focus on implementation and debugging in the DevOps exam - the SA Pro exam is more about breadth of knowledge across AWS services.

  1. Remember to solve exactly what the question is asking. If they want performance, give them read replicas, sharding, PIOPS, etc. If they want elasticity and high availability, start using Multi-AZ, cross-region replication, Auto Scaling Groups, ELB’s, etc. If they are worried about cost, remember the S3 storage classes, and use AutoScaling groups to scale down unnecessary instances.
  2. The basic stages of a Continuous Integration pipeline are Source Control -> Build -> Staging -> Production.
  3. ELB can be configured to publish logs at a 5 or 60 minute interval. Access logs are disabled by default.
  4. When an instance is launched by Autoscaling, the instance can be in the pending state for up to 60 minutes by default.
  5. Cloudwatch logs can be streamed to Kinesis, Lambda, or ElasticSearch for further processing.
  6. The fastest deployment times in Elastic Beanstalk are achieved using the All At Once deployment method.
  7. Remember this Opswork trivia: “For Linux stacks, if you want to associate an Amazon RDS service layer with your app, you must add the appropriate driver package to the associated app server layer”
  8. Generally, if you see Docker, think Elastic Beanstalk!
  9. AWS OpsWorks Stacks agents communicate regularly. If an agent does not communicate with the service for more than approximately five minutes, AWS OpsWorks Stacks considers the instance to have failed.
  10. You can configure an Opswork stack to accept custom cookbooks after creation. Just enable the option.
  11. Speaking of custom cookbooks, remember that the custom cookbook option is enabled at the stack level, not the layer level.
  12. The CodeDeploy service relies on a appspec.yml file included with your source code binaries on an EC2/On-Premises Compute Platform.
  13. A nested stack is a stack that you create within another stack by using the AWS::CloudFormation::Stack resource.
  14. When attempting to secure data in transit through an ELB, remember you can use either an HTTPS or SSL listener.
  15. AWS CodeDeploy can deploy application content stored in Amazon S3 buckets, GitHub repositories, or Bitbucket repositories. Note: Subversion is not supported.
  16. CodeDeploy offers two deployment types: In-place deployment and Blue-Green deployment.
  17. A Dockerrun.aws.json file is an Elastic Beanstalk–specific JSON file that describes how to deploy a set of Docker containers as an Elastic Beanstalk application.
  18. This exam is extremely tough to parse at first glance. There are lots of grammatical twists and turns that are not present in typical exams. You may need to check your proposed architecture against the question 4-5 times to ensure that you are solving the right problem.
  19. Remember to use the NoEcho property in CloudFormation templates to prevent a describe command from listing sensitive values, such as passwords.
  20. You really need to understand the lifecycle of an EC2 instance (pictured below) being brought into service.
EC2 Auto Scaling Lifecycle

Training Materials I Used

Whitepapers I Read


AWS Solutions Architect Professional Exam Tips and Tricks

26 Nov 2018

If you’re planning on taking the AWS Solutions Architect Professional exam, I’ve compiled a quick list of tips that you may want to remember headed into the exam.

I passed the exam on November 24th, 2018. Before taking this exam, I held all three Associate certifications and the Security Specialty certification. I passed with an 80% score, and it took 69 minutes.

This exam is very difficult - on par with the Security Specialty exam! You will not accidentally pass this exam. :)

STS and Identity Broker Diagram
  1. You need to know your STS use cases inside and out. Remember that you generally need to develop an identity broker, your application will authenticate against LDAP or your broker, and then against STS. Your application will not directly authenticate against STS first!
  2. Understand how cross-account access, and granting that access, works.
  3. Remember to turn on the Cloudtrail “global services” in order to track IAM usage.
  4. Prefer Cloudformation for version controlled infrastructure configuration.
  5. You need to understand what BGP does (and what it doesn’t do) in relation to Direct Connect. The ACloudGuru course did not cover eBGP or weighting policies very heavily, so you will need to do additional research. I recommend watching this video before taking the exam.
  6. Use SQS queues placed in front of RDS/DynamoDB to reduce the load on your databases!
  7. Use Kinesis for large amounts of incoming data, especially when it’s coming from multiple sources.
  8. Any time you see the words “mobile app”, “social networking login”, “Login with Amazon”, or “Facebook” on the exam, you should immediately be able to narrow down the answers - hint: the correct answer will probably involve Web Identity Federation.
  9. Play around with this CIDR calculator if you have trouble understanding VPC/subnet CIDR ranges and possibilities. Always remember that AWS reserves five IP addresses for their use (first four IP’s and the last IP).
  10. There are usually two blatantly incorrect answers, and two answers that could be right. Narrow down your choices.
  11. Understand the various best practices for encrypting data at rest and in transit.
  12. You need to understand public VIF’s vs. private VIF’s, and which services use which type of VIF.
  13. You need to understand ELBs, Auto Scaling, and how to architect a scalable architecture. By this point in your AWS studies (assuming you already have an Associate certificate), you should have this pretty down pat. I recommend watching this video for a cool overview of setting up low latency multiplayer servers globally.
  14. Even though Data Pipeline has been pretty much usurped by Lambda, it features fairly heavily on the exam. Definitely watch the ACloudGuru videos on Data Pipeline (and follow the labs) before taking the exam.
  15. AWS Connector for vCenter shows up in one or two questions, and is worth knowing about.
  16. You need to know the different instance types and what usage scenarios are appropriate for them. Remember that if you receive a capacity error when resizing a placement group, you just need to stop and restart the placement group - this will allocate the group to a new physical cluster with a proper amount of instances.

Training Materials I Used

Videos I Watched (in order of importance)

Whitepapers I Read