Friday, December 30, 2016

AWS re:Invent 2016 Video & Slide Presentation Links with Easy Index

As with last year, here is my quick index of all re:Invent sessions. I'll keep running the tool to fill in the index.  It usually takes Amazon a few weeks to fully upload all the videos and presentations. This year it looks like Amazon got the majority of content on Youtube and Slideshare very quick with a few Slideshares still trickling in.

See below for how I created the index (with code):


ALX201 - How Capital One Built a Voice-Based Banking Skill for Amazon Echo
As we add thousands of skills to Alexa, our developers have uncovered some basic and more complex tips for building better skills. Whether you are new to Alexa skill development or if you have created skills that are live today, this session helps you understand how to create better voice experiences. Last year, Capital One joined Alexa on stage at re:Invent to talk about their experience building an Alexa skill. Hear from them one year later to learn from the challenges that they had to overcome and the results they are seeing from their skill. In this session, you will learn the importance of flexible invocations, better VUI design, how OAuth and account linking can add value to your skill, and about Capital One's experience building an Alexa skill.
ALX202 - How Amazon is enabling the future of Automotive
The experience in the auto industry is changing. For both the driver and the car manufacturer, a whole new frontier is on the near horizon. What do you do with your time while the car is driving itself? How do I have a consistent experience while driving shared or borrowed cars? How do I stay safer and more aware in the ever increasing complexity of traffic, schedules, calls, messages and tweets? In this session we will discuss how the auto industry is facing new challenges and how the use of Amazon Alexa, IoT, Logistics services and the AWS Cloud is transforming the Mobility experience of the (very near) future.
ALX203 - Workshop: Creating Voice Experiences with Alexa Skills: From Idea to Testing in Two Hours
This workshop teaches you how to build your first voice skill with Alexa. You bring a skill idea and well show you how to bring it to life. This workshop will walk you through how to build an Alexa skill, including Node.js setup, how to implement an intent, deploying to AWS Lambda, and how to register and test a skill. Youll walk out of the workshop with a working prototype of your skill idea. Prerequisites: Participants should have an AWS account established and available for use during the workshop. Please bring your own laptop.
ALX204 - Workshop: Build an Alexa-Enabled Product with Raspberry Pi
Fascinated by Alexa, and want to build your own device with Alexa built in? This workshop will walk you through to how to build your first Alexa-powered device step by step, using a Raspberry Pi. No experience with Raspberry Pi or Alexa Voice Service is required. We will provide you with the hardware and the software required to build this project, and at the end of the workshop, you will be able to walk out with a working prototype of Alexa on a Pi. Please bring a WiFi capable laptop.
ALX301 - Alexa in the Enterprise: How JPL Leverages Alexa to Further Space Exploration with Internet of Things
The Jet Propulsion Laboratory designs and creates some of the most advanced space robotics ever imagined. JPL IT is now innovating to help streamline how JPLers will work in the future in order to design, build, operate, and support these spacecraft. They hope to dramatically improve JPLers' workflows and make their work easier for them by enabling simple voice conversations with the room and the equipment across the entire enterprise. What could this look like? Imagine just talking with the conference room to configure it. What if you could kick off advanced queries across AWS services and kick off AWS Kinesis tasks by simply speaking the commands? What if the laboratory could speak to you and warn you about anomalies or notify you of trends across your AWS infrastructure? What if you could control rovers by having a conversation with them and ask them questions? In this session, JPL will demonstrate how they leveraged AWS Lambda, DynamoDB and CloudWatch in their prototypes of these use cases and more. They will also discuss some of the technical challenges they are overcoming, including how to deploy and manage consumer devices such as the Amazon Echo across the enterprise, and give lessons learned. Join them as they use Alexa to query JPL databases, control conference room equipment and lights, and even drive a rover on stage, all with nothing but the power of voice!
ALX302 - Build a Serverless Back End for Your Alexa-Based Voice Interactions
Learn how to develop voice-based serverless back ends for Alexa Voice Service (AVS) and Alexa devices using the Alexa Skills Kit (ASK), which allows you to add new voice-based interactions to Alexa. Well code a new skill, implemented by a serverless backend leveraging AWS services such as Amazon Cognito, AWS Lambda, and Amazon DynamoDB. Often, your skill needs to authenticate your users and link them back to your backend systems and to persist state between user invocations. User authentication is performed by leveraging OAuth compatible identity systems. Running such a system on your back end requires undifferentiated heavy lifting or boilerplate code. Well leverage Login with Amazon as the identity provider instead, allowing you to focus on your application implementation and not on the low-level user management parts. At the end of this session, youll be able to develop your own Alexa skills and use Amazon and AWS services to minimize the required backend infrastructure. This session shows you how to deploy your Alexa skill code on a serverless infrastructure, leverage AWS Lambda, use Amazon Cognito and Login with Amazon to authenticate users, and leverage AWS DynamoDB as a fully managed NoSQL data store.
ALX303 - Building a Smarter Home with Alexa
Natural user interfaces, such as those based on speech, enable customers to interact with their home in a more intuitive way. With the VUI (Voice User Interface) smart home, now customers don't need to use their hands or eyes to do things around the home they only have to ask and it's at their command. This session will address the vision for the VUI smart home and how innovations with Amazon Alexa make it possible.
ALX304 - Tips and Tricks on Bringing Alexa to Your Products
Ever wonder what it takes to add the power of Alexa to your own products? Are you curious about what Alexa partners have learned on their way to a successful product launch? In this session you will learn about the top tips and tricks on how to go from VUI newbie to an Alexa-enabled product launch. Key concepts around hardware selection, enabling far field voice interaction, building a robust Alexa Voice Service (AVS) client and more will be discussed along with customer and partner examples on how to plan for and avoid common challenges in product design, development and delivery.
ALX305 - From VUI to QA: Building a Voice-Based Adventure Game for Alexa
Hitting the submit button to publish your skill is similar to sending your child to their first day of school. You want it to be set up for a successful launch day and for many days thereafter. Learn how to set your skill up for success from Andy Huntwork, Alexa Principal Engineer and one of the creators of the popular Alexa skill The Magic Door. You will learn the most common reasons why skills fail and also some of the more unique use cases. The purpose of this session is to help you build better skills by knowing what to look out for and what you can test for before submitting. In this session, you will learn what most developers do wrong, how to successfully test and QA your skill, how to set your skill up for successful certification, and the process of how a skill gets certified.
ALX306 - State of the Union: Amazon Alexa and Recent Advances in Conversational AI
The way humans interact with machines is at a turning point, and conversational artificial intelligence (AI) is at the center of the transformation. Learn how Amazon is using machine learning and cloud computing to fuel innovation in AI, making Amazon Alexa smarter every day. Alexa VP and Head Scientist Rohit Prasad presents the state of the union Alexa and Recent Advances in Conversational AIn for Alexa. He addresses Alexa's advances in spoken language understanding and machine learning, and shares Amazon's thoughts about building the next generation of user experiences.
ALX307 - Voice-enabling Your Home and Devices with Amazon Alexa and AWS IoT
Want to learn how to Alexa-power your home? Join Brookfield Residential CIO and EVP Tom Wynnyk and Senior Solutions Architect Nathan Grice, for Alexa Smart Homefor an overview of building the next generation of integrated smart homes using Alexa to create voice-first experiences. Understand the technologies used and how to best expose voice experiences to users through Alexa. Paul and Nathan cover the difference between custom Alexa skills and Smart Home Skill API skills, and build a home automation control from the ground up using Alexa and AWS IoT.
ARC201 - Scaling Up to Your First 10 Million Users
Cloud computing gives you a number of advantages, such as the ability to scale your web application or website on demand. If you have a new web application and want to use cloud computing, you might be asking yourself, "Where do I start?" Join us in this session to understand best practices for scaling your resources from zero to millions of users. We show you how to best combine different AWS services, how to make smarter decisions for architecting your application, and how to scale your infrastructure in the cloud.
ARC202 - Accenture Cloud Platform Serverless Journey
Accenture Cloud Platform helps customers manage public and private enterprise cloud resources effectively and securely. In this session, learn how we designed and built new core platform capabilities using a serverless, microservices-based architecture that is based on AWS services such as AWS Lambda and Amazon API Gateway. During our journey, we discovered a number of key benefits, including a dramatic increase in developer velocity, a reduction (to almost zero) of reliance on other teams, reduced costs, greater resilience, and scalability. We describe the (wild) successes weve had and the challenges weve overcome to create an AWS serverless architecture at scale. Session sponsored by Accenture. AWS Competency Partner
ARC203 - Achieving Agility by Following Well-Architected Framework Principles on AWS
The AWS Well-Architected Framework enables customers to understand best practices around security, reliability, performance, and cost optimization when building systems on AWS. This approach helps customers make informed decisions and weigh the pros and cons of application design patterns for the cloud. In this session, you'll learn how National Instruments used the Well-Architected Framework to follow AWS guidelines and best practices. By developing a strategy based on the AWS Well-Architected Framework, National Instruments was able to triple the number of applications running in the cloud without additional head count, significantly increase the frequency of code deployments, and reduce deployment times from two weeks to a single day. As a result, National Instruments was able to deliver a more scalable, dynamic, and resilient LabVIEW platform with agility.
ARC204 - From Resilience to Ubiquity - #NetflixEverywhere Global Architecture
Building and evolving a pervasive, global service requires a multi-disciplined approach that balances requirements with service availability, latency, data replication, compute capacity, and efficiency. In this session, well follow the Netflix journey of failure, innovation, and ubiquity. We'll review the many facets of globalization and then delve deep into the architectural patterns that enable seamless, multi-region traffic management; reliable, fast data propagation; and efficient service infrastructure. The patterns presented will be broadly applicable to internet services with global aspirations.
ARC205 - Born in the Cloud; Built Like a Startup
This presentation provides a comparison of three modern architecture patterns that startups are building their business around. It includes a realistic analysis of cost, team management, and security implications of each approach. It covers Elastic Beanstalk, Amazon ECS, Docker, Amazon API Gateway, AWS Lambda, Amazon DynamoDB, and Amazon CloudFront, as well as Docker.
ARC207 - NEW LAUNCH! Additional transparency and control for your AWS environment through AWS Personal Health Dashboard
When your business is counting on the performance of your cloud solutions, having relevant and timely insights into events impacting your AWS resources is essential. AWS Personal Health Dashboard serves as the primary destination for you to receive personalized information related to your AWS infrastructure, guiding your through scheduled changes, and accelerating the troubleshooting of issues impacting your AWS resources. The service, powered by AWS Health APIs, integrates with your in-house event management systems, and can be programmatically configured to proactively get the right information into the right hands at the right time. The service is integrated with Splunk App for AWS to enhance Splunks dashboards, reports and alerts to deliver real-time visibility into your environment.
ARC208 - Hybrid Architectures: Bridging the Gap to the Cloud
AWS provides many services to assist customers with their journey to the cloud. Hybrid solutions offer customers a way to continue leveraging existing investments on-premises, while expanding their footprint into the public cloud. This session covers the different technologies available to support hybrid architectures on AWS. We discuss common patterns and anti-patterns for solving enterprise workloads across a hybrid environment.
ARC209 - Attitude of Iteration
In todays world, technology changes at a breakneck speed. What was new this morning is outdated at lunch. Working in the AWS Cloud is no different. Every week, AWS announces new features or improvements to current products. As AWS technologists, we must assimilate these new technologies and make decisions to adopt, reject, or defer. These decisions can be overwhelming: we tend to either reject everything and become stagnant, or adopt everything and never get our project out the door. In this session we will discuss the attitude of iteration. The attitude of iteration allows us to face the challenges of change without overwhelming our technical teams with a constant tug-o-war between implementation and improvement. Whether youre an architect, engineer, developer, or AWS newbie, prepare to laugh, cry, and commiserate as we talk about overcoming these challenges. Session sponsored by Rackspace.
ARC210 - Workshop: Addressing Your Business Needs with AWS
Come and participate with other AWS customers as we focus on the overall experience of using AWS to solve business problems. This is a great opportunity to collaborate with existing and prospective AWS users to validate your thinking and direction with AWS peers, discuss the resources that aid AWS solution design, and give direct feedback on your experience building solutions on AWS.
ARC211 - Solve common problems with ready to use solutions in 5 minutes or less
Regularly, customers at AWS assign resources to create solutions that address common problems shared between businesses of all sizes. Often, this results in taking resources away from products or services that truly differentiate the business in the marketplace. The Solutions Builder team at AWS focuses on developing and publishing a catalog of repeatable, standardized solutions that can be rapidly deployed by customers to overcome common business challenges. In this session, the Solutions Builder team will share ready to use solutions that make it easy for anyone to create a transit VPC, centralized logging, a data lake, scheduling for Amazon EC2, and VPN monitoring. Along the way, the team reveals the architectural tenets and best practices they follow for the development of these solutions. In the end, customers are introduced to a catalog of freely available solutions with a peek into the architectural approaches used by an internal team at AWS.
ARC212 - Salesforce: Helping Developers Deliver Innovations Faster
Salesforce is one of the most innovative enterprise software companies in the world, delivering 3 major releases a year with hundreds of features in each release. In this session, come learn how we enable thousands of engineers within Salesforce to utilize a flexible development environment to deliver these innovations to our customers faster. We show you how we enable engineers at Salesforce to test not only individual services they are developing but also large scale service integrations. Also learn how we can achieve setup of a representative production environment in minutes and teardown in seconds, using AWS.
ARC213 - Open Source at AWS—Contributions, Support, and Engagement
Over the last few years, we have seen a dramatic increase in the use of open source projects as the mainstay of architectures in both startups and enterprises. Many of our customers and partners also run their own open source programs and contribute key technologies to the industry as a whole (see DCS201). At AWS weengage with open source projects in a number of ways. Wecontribute bug fixesand enhancementstopopular projectsincluding ourwork with the Hadoop ecosystem (see BDM401), Chromium(see BAP305) and (obviously) Boto.We have our own standalone projectsincludingthe security library s2n (see NET405)and machine learning project MXnet (see MAC401).Wealsohave services that make open source easier to use like ECS for Docker (see CON316), and RDS for MySQL and PostgreSQL (see DAT305).In this session you will learn about our existing open source work across AWS, and our next steps.
ARC301 - Architecting Next Generation SaaS Applications on AWS
AWS provides a broad array of services, tools, and constructs that can be used to design, operate, and deliver SaaS applications. In this session, Tod Golding, the AWS Partner Solutions Architect, shares the wisdom and lessons learned from working with dozens of customers and partners building SaaS solutions on AWS. We discuss key architectural strategies and patterns that are used to deliver multi-tenant SaaS models on AWS and dive into the full spectrum of SaaS design and architecture considerations, including tenant isolation models, tenant identity management, serverless SaaS, and multi-tenant storage strategies. This session connects the dots between general SaaS best practices and what it means to realize these patterns on AWS, weighing the architectural tradeoffs of each model and assessing its influence on the agility, manageability, and cost profile of your SaaS solution.
ARC302 - From One to Many: Evolving VPC Design
As more customers adopt Amazon VPC architectures, the features and flexibility of the service are squaring off against evolving design requirements. This session follows this evolution of a single regional VPC into a multi-VPC, multi-region design with diverse connectivity into on-premises systems and infrastructure. Along the way, we investigate creative customer solutions for scaling and securing outbound VPC traffic, securing private access to Amazon S3, managing multi-tenant VPCs, integrating existing customer networks through AWS Direct Connect, and building a full VPC mesh network across global regions.
ARC303 - Cloud Monitoring - Understanding, Preparing, and Troubleshooting Dynamic Apps on AWS
Applications running in a typical data center are static entities. Dynamic scaling and resource allocation are the norm in AWS. Technologies such as Amazon EC2, Docker, AWS Lambda, and Auto Scaling make tracking resources and resource utilization a challenge. The days of static server monitoring are over. In this session, we examine trends weve observed across thousands of customers using dynamic resource allocation and discuss why dynamic infrastructure fundamentally changes your monitoring strategy. We discuss some of the best practices weve learned by working with New Relic customers to build, manage, and troubleshoot applications and dynamic cloud services. Session sponsored by New Relic. AWS Competency Partner
ARC304 - Effective Application Data Analytics for Modern Applications
IT is evolving from a cost center to a source of continuous innovation for business. At the heart of this transition are modern, revenue-generating applications, based on dynamic architectures that constantly evolve to keep pace with end-customer demands. This dynamic application environment requires a new, comprehensive approach to traditional monitoring one based on real-time, end-to-end visibility and analytics across the entire application lifecycle and stack, instead of monitoring by piecemeal. This presentation highlights practical advice on how developers and operators can leverage data and analytics to glean critical information about their modern applications. In this session, we will cover the types of data important for todays modern applications. Well discuss visibility and analytics into data sources such as AWS services (e.g., Amazon CloudWatch, AWS Lambda, VPC Flow Logs, Amazon EC2, Amazon S3, etc.), development tool chain, and custom metrics, and describe how to use analytics to understand business performance and behaviors. We discuss a comprehensive approach to monitoring, troubleshooting, and customer usage insights, provide examples of effective data analytics to improve software quality, and describe an end-to-end customer use case that highlights how analytics applies to the modern app lifecycle and stack. Session sponsored by Sumo Logic. AWS Competency Partner
ARC305 - From Monolithic to Microservices: Evolving Architecture Patterns in the Cloud
Gilt, a global e-commerce company, implemented a sophisticated microservices architecture on AWS to handle millions of customers visiting their site at noon every day. The microservices architecture pattern enables independent service scaling, faster deployments, better fault isolation, and graceful degradation. In this session, Emerson Loureiro, Sr. Software Engineer at Gilt, will share Gilt's experiences and lessons learned during their evolution from a single monolithic Rails application in a traditional data center to more than 300 Scala/Java microservices deployed in the cloud.Derek Chiles, AWS Solutions Architect, will review best practices and recommended architectures for deploying microservices on AWS.
ARC306 - Event Handling at Scale: Designing an Auditable Ingestion and Persistence Architecture for 10K+ events/second
How does McGraw-Hill Education use the AWS platform to scale and reliably receive 10,000 learning events per second? How do we provide near-real-time reporting and event-driven analytics for hundreds of thousands of concurrent learners in a reliable, secure, and auditable manner that is cost effective? MHE designed and implemented a robust solution that integrates AWS API Gateway, AWS Lambda, Amazon Kinesis, Amazon S3, Amazon Elasticsearch Service, Amazon DynamoDB, HDFS, Amazon EMR, Amazopn EC2, and other technologies to deliver this cloud-native platform across the US and soon the world. This session describes the challenges we faced, architecture considerations, how we gained confidence for a successful production roll-out, and the behind-the-scenes lessons we learned.
ARC307 - Accelerating Next Generation Healthcare Business on the AWS Cloud
Hear Geneia's design principles for using multiple technologies like Elastic Load Balancing and Auto Scaling in end-to-end solutions to meet regulatory requirements. Explore how to meet HIPAA regulations by using native cloud services like Amazon EC2, Amazon EBS volumes, encryption services, and monitoring features in addition to third-party tools to ensure end-to-end data protection, privacy, and security for protected health information (PHI) data hosted in the AWS Cloud. Learn how Geneia leveraged multiregion and multizone backup and disaster recovery solutions to address the recovery time objective (RTO) and recovery point objective (RPO) requirements. Discover how automated build, deployment, provisioning, and virtual workstations in the cloud enabled Geneia's developers and data scientists to quickly provision resources and work from any location, expediting the onboarding of customers, getting to market faster, and capturing bigger market share in healthcare analytics while minimizing costs. Session sponsored by Cognizant. AWS Competency Partner
ARC308 - Metering Big Data at AWS: From 0 to 100 Million Records in 1 Second
Learn how AWS processes millions of records per second to support accurate metering across AWS and our customers. This session shows how we migrated from traditional frameworks to AWS managed services to support a large processing pipeline. You will gain insights on how we used AWS services to build a reliable, scalable, and fast processing system using Amazon Kinesis, Amazon S3, and Amazon EMR. Along the way we dive deep into use cases that deal with scaling and accuracy constraints. Attend this session to see AWSs end-to-end solution that supports metering at AWS.
ARC309 - Moving Mission Critical Apps from One Region to Multi-Region active/active
In gaming, low latencies and connectivity are bare minimum expectations users have while playing online on PlayStation Network. Alex and Dustin share key architectural patterns to provide low latency, multi-region services to global users. They discuss the testing methodologies and how to programmatically map out a large dependency multi-region deployment with data-driven techniques. The patterns shared show how to adapt to changing bottlenecks and sudden, several million request spikes. Youll walk away with several key architectural patterns that can service users at global scale while being mindful of costs.
ARC310 - Cost Optimizing Your Architecture: Practical Design Steps For Big Savings
Did you know that AWS enables builders to architect solutions for price? Beyond the typical challenges of function, performance, and scale, you can make your application cost effective. Using different architectural patterns and AWS services in concert can dramatically reduce the cost of systems operation and per-transaction costs. This session uses practical examples aimed at architects and developers. Using code and AWS CloudFormation in concert with services such as Amazon EC2, Amazon ECS, Lambda, Amazon RDS, Amazon SQS, Amazon SNS, Amazon S3, CloudFront, and more, we demonstrate the financial advantages of different architectural decisions. Attendees will walk away with concrete examples, as well as a new perspective on how they can build systems economically and effectively. Attendees at this session will receive a free 30 day trial of AWS Trusted Advisor.
ARC311 - Evolving a Responsive and Resilient Architecture to Analyze Billions of Metrics
Nike+ is at the core of the Nike digital product ecosystem, providing services to enhance your athletic experience through quantified activity tracking and gamification. As one of the first movers at Nike to migrate out of the datacenter to AWS, they share the evolution in building a reactive platform on AWS to handle large, complex data sets. They provide a deep technical view of how they process billions of metrics a day in their quantified-self platform, supporting millions of customers worldwide. Youll leave with ideas and tools to help your organization scale in the cloud. Come learn from experts who have built an elastic platform using Java, Scala, and Akka, leveraging the power of many AWS technologies like Amazon EC2, ElastiCache, Amazon SQS, Amazon SNS, DynamoDB, Amazon ES, Lambda, Amazon S3, and a few others that helped them (and can help you) get there quickly.
ARC312 - Compliance Architecture: How Capital One Automates the Guard Rails for 6,000 Developers
What happens when you give 6,000 developers access to the cloud? Introducing Cloud Custodian, an open source project from Capital One, which provides a DSL for AWS fleet management that operates in real-time using CloudWatch Events and Lambda. Cloud Custodian is used for the gamut of compliance, encryption, and cost optimization. What can it do for you?
ARC313 - Running Lean Architectures: How to Optimize for Cost Efficiency
Whether youre a cash-strapped startup or an enterprise optimizing spend, it pays to run cost-efficient architectures on AWS. This session reviews a wide range of cost planning, monitoring, and optimization strategies, featuring real-world experience from AWS customers. We cover how to effectively combine Amazon EC2 On-Demand, Reserved, and Spot instances to handle different use cases; leveraging Auto Scaling to match capacity to workload; choosing the optimal instance type through load testing; taking advantage of Multi-AZ support; and using Amazon CloudWatch to monitor usage and automatically shut off resources when they are not in use. We discuss taking advantage of tiered storage and caching, offloading content to Amazon CloudFront to reduce back-end load, and getting rid of your back end entirely by leveraging AWS high-level services. We also showcase simple tools to help track and manage costs, including Cost Explorer, billing alerts, and AWS Trusted Advisor. This session is your pocket guide for running cost effectively in the Amazon Cloud. Attendees of this session receive a free 30-day trial of enterprise-level Trusted Advisor.
ARC314 - Enabling Enterprise Migrations: Creating an AWS Landing Zone
With customers migrating workloads to AWS, we are starting to see a need for the creation of a prescribed landing zone, which uses native AWS capabilities and meets or exceeds customers' security and compliance objectives. In this session, we will describe an AWS landing zone and will cover solutions for account structure, user configuration, provisioning, networking and operation automation. This solution is based on AWS native capabilities such as AWS Service Catalog, AWS Identity and Access Management, AWS Config Rules, AWS CloudTrail and Amazon Lambda. We will provide an overview of AWS Service Catalog and how it be used to provide self-service infrastructure to applications users, including various options for automation. After this session you will be able to configure an AWS landing zone for successful large scale application migrations. Additionally, Philips will explain their cloud journey and how they have applied their guiding principles when building their landing zone.
ARC315 - The Enterprise Fast Lane - What Your Competition Doesn't Want You To Know About Enterprise Cloud Transformation
Fed up with stop and go in your data center? Shift into overdrive and pull into the fast lane! Learn how AutoScout24, the largest online car marketplace Europe-wide, are building their Autobahn in the Cloud. The secret ingredient? Culture! Because Cloud is only one half of the digital transformation story: The other half is how your organization deals with cultural change as you transition from the old world of IT into building microservices on AWS with agile DevOps teams in a true you build it you run it fashion. Listen to stories from the trenches, powered by Amazon Kinesis, Amazon DynamoDB, AWS Lambda, Amazon ECS, Amazon API Gateway and much more, backed by AWS Partners, AWS Professional Services, and AWS Enterprise Support. Key takeaways: How to become Cloud native, evolve your architecture step by step, drive cultural change across your teams, and manage your companys transformation for the future.
ARC316 - Hybrid IT: A Stepping Stone to All-In
This session demonstrates how customers can leverage hybrid IT as a transitional step on the path to going all-in on AWS. We provide a step-by-step walk-through focusing on seamless migration to the cloud, with consideration given to existing data centers, equipment, and staff retraining. Learn about the suite of capabilities AWS provides to ease and simplify your journey to the cloud.
ARC318 - Busting the Myth of Vendor Lock-In: How D2L Embraced the Lock and Opened the Cage
When D2L first moved to the cloud, we were concerned about being locked-in to one cloud provider. We were compelled to explore the opportunities of the cloud, so we overcame our perceived risk, and turned it into an opportunity by self-rolling tools and avoiding AWS native services. In this session, you learn how D2L tried to bypass the lock buteventually embraced itand opened the cage. Avoiding AWS native tooling and pure lifts of enterprise architecture caused a drastic inflation of costs. Learn how we shifted away from a self-rolled lift into an efficient and effective shift while prioritizing cost, client safety, AND speed of development. Learn from D2L'ssuccesses and missteps, and convert your own enterprise systems into the cloud both through native cloud births and enterprise conversions. This session discusses D2Ls use of Amazon EC2 (with aguest appearance by Reserved Instances), Elastic Load Balancing, Amazon EBS, Amazon DynamoDB, Amazon S3, AWS CloudFormation, AWS CloudTrail, Amazon CloudFront, AWS Marketplace, Amazon Route 53, AWS Elastic Beanstalk, and Amazon ElastiCache.
ARC319 - Datapipe Open Source: Image Development Pipeline
For an IT organization to be successful in rapid cloud assessment or iterative migration of their infrastructure and applications to AWS, they need to effectively plan and execute on a strategic cloud strategy that focuses not only on cloud, but also big data, DevOps, and security.Session sponsored by Datapipe. AWS Competency Partner
ARC320 - Workshop: AWS Professional Services Effective Architecting Workshop
The AWS Professional Services team will be facilitating an architecture workshop exercise for certified AWS Architects. Class size will be limited to 48. This workshop will be a highly interactive architecture design exercise where the class will be randomly divided into teams and given a business case for which they will need to design an effective AWS solution. Past participants have found the interaction with people from other organizations and the creative brainstorming that occurs across 6 different teams greatly enhances the learning experience. Flipcharts will be provided and students are encouraged to bring their laptops to document their designs. Each team will be expected to present their solution to the class.
ARC402 - Serverless Architectural Patterns and Best Practices
As serverless architectures become more popular, AWS customers need a framework of patterns to help them deploy their workloads without managing servers or operating systems. This session introduces and describes four re-usable serverless patterns for web apps, stream processing, batch processing, and automation. For each, we provide a TCO analysis and comparison with its server-based counterpart. We also discuss the considerations and nuances associated with each pattern and have customers share similar experiences. The target audience is architects, system operators, and anyone looking for a better understanding of how serverless architectures can help them save money and improve their agility.
ARC403 - Building a Microservices Gaming Platform for Turbine Mobile Games
Warner Bros Turbine team shares lessons learned from their enhanced microservices game platform, which uses Docker, Amazon EC2, Elastic Load Balancing, and Amazon ElastiCache to scale up in anticipation of massive game adoption. Learn about their Docker-based microservices architecture, tuned and optimized to support the demands of the massively popular [Batman: Arkham Underworld and other franchises]. Turbine invent and simplify microservices persistence services consolidating their previous NoSQL database solution with highly performant PostgreSQL on Amazon EC2 and Amazon EBS. Turbine also describes other innovative strategies, including integrated analytic techniques to anticipate and predict their scaling operations.
ARC404 - Migrating a Highly Available and Scalable Database from Oracle to Amazon DynamoDB
In this session, we share how an Amazon.com team that owns a document management platform that manages billions of critical customer documents for Amazon.com migrated from a relational to a non-relational database. Initially, the service was built as an Oracle database. As it grew, the team discovered the limits of the relational model and decided to migrate to a non-relational database. They chose Amazon DynamoDB for its built-in resilience, scalability, and predictability. We provide a template that you can use to migrate from a relational data store to DynamoDB. We also provide details about the entire process: design patterns for moving from a SQL schema to a NoSQL schema; mechanisms used to transition from an ACID (Atomicity, Consistency, Isolation, Durability) model to an eventually consistent model; migration alternatives considered; pitfalls in common migration strategies; and how to ensure service availability and consistency during migration.
ARC405 - Running, Configuring, and Securing Windows Workloads
From servers to workstations, AWS provides the best place to run your Windows workloads. In this session, we'll discuss the ease of deploying Windows workloads on AWS, and architecting for performance, scalability, security, and cost savings. We will explore the use of AWS Directory Service, the Amazon EC2 Run command, and Windows PowerShell to bootstrap your instances for seamless Microsoft Active Directory integration, application installation, and management. We will walk through an architecture that includes Amazon RDS, Amazon EC2, and Amazon WorkSpaces, and discuss the secure relationships among these services. You will learn how you can use native AWS services as well as the tools you are already familiar with to manage your Windows environment.
ARC406 - Encoding Artifacts to Emmy Awards: Taking on Terabyte-Scale, 1-Gbps, 4K Video Processing in the Cloud
The advent of 4K video has resulted in a huge uptick in resource requirements, which is difficult to scale in a traditional environment. The cloud is a perfect environment for handling problems of this scale; however, there are many unanswered questions around best practices and suitable architectures for dealing with massive, high-quality assets. In this session, we will define problem cases and discuss practical architectural patterns for dealing with these challenges by using AWS services such as Amazon EC2 (graphical instances), Amazon EMR, Amazon S3, Amazon S3 Transfer Acceleration, Amazon Glacier, AWS Snowball, and the new magnetic EBS volumes. The best practices that we'll discuss will also be helpful to architects and engineers who are dealing with non-video data. Amazon Studios will present how, powered by AWS, they solved many of these problems and are able to create, manage, and distribute Emmy Award-winning content.
ARC408 - Optimizing SaaS Solutions for AWS
OK, so you have your SaaS application up and running on AWS and now the real fun begins. The adoption of SaaS and a shared multi-tenant infrastructure brings with it a unique set of challenges that require developers to find new and creative ways to optimize their SaaS applications. For many engineers, this means tackling a whole new realm of questions. How do you optimize tenant experience by tiers? How do you perform real-time performance tuning or centralized tenant policy management? How do you collect and aggregate tenant analytics? How do you implement scalable policy-driven configuration of tenant experience? These and many other challenging topics will be covered in this technical deep dive. Tod Golding, an AWS Solutions Architect specializing in helping AWS partners build SaaS systems, will explore a collection of strategies that can be leveraged to increase the performance, flexibility, and agility of SaaS environments. You will learn how to identify areas where services, tools, and design strategies can be combined to enrich the agility of your SaaS architecture.
ARC409 - Deploying Your First 100K Windows Users
Deploy, scale, and manage the Microsoft servers on AWS supporting 100K users. This session covers the AWS and Microsoft architectures to stand up Exchange, SharePoint, Lync, SQL Server, and Active Directory for enterprises. We focus on the architecture, automation, and management of Windows resources on AWS, including cloud-first technologies like AWS CloudFormation, Amazon EC2 Simple Systems Manager, and Amazon CloudWatch Logs.
BDA203 - Billions of Rows Transformed in Record Time Using Matillion ETL for Amazon Redshift
Billions of Rows Transformed in Record Time Using Matillion ETL for Amazon Redshift GE Power Water develops advanced technologies to help solve some of the worlds most complex challenges related to water availability and quality. They had amassed billions of rows of data on on-premises databases, but decided to migrate some of their core big data projects to the AWS Cloud. When they decided to transform and store it all in Amazon Redshift, they knew they needed an ETL/ELT tool that could handle this enormous amount of data and safely deliver it to its destination. In this session, Ryan Oates, Enterprise Architect at GE Water, shares his use case, requirements, outcomes and lessons learned. He also shares the details of his solution stack, including Amazon Redshift and Matillion ETL for Amazon Redshift in AWS Marketplace. You learn best practices on Amazon Redshift ETL supporting enterprise analytics and big data requirements, simply and at scale. You learn how to simplify data loading, transformation and orchestration on to Amazon Redshift and how build out a real data pipeline. Get the insights to deliver your big data project in record time.
BDA204 - Leverage the Power of the Crowd To Work with Amazon Mechanical Turk
With Amazon Mechanical Turk (MTurk), you can leverage the power of the crowd for a host of tasks ranging from image moderation and video transcription to data collection and user testing. You simply build a process that submit tasks to the Mechanical Turk marketplace and get results quickly, accurately, and at scale. In this session, Russ, from Rainforest QA, shares best practices and lessons learned from his experience using MTurk. The session covers the key concepts of MTurk, getting started as a Requester, and using MTurk via the API. You learn how to set and manage Worker incentives, achieve great Worker quality, and how to integrate and scale your crowdsourced application. By the end of this session, you will have a comprehensive understanding of MTurk and know how to get started harnessing the power of the crowd.
BDA205 - Delighting Customers Through Device Data with Salesforce IoT Cloud and AWS IoT
The Internet of Things (IoT) produces vast quantities of data that promise a deep, always connected view into customer experiences through their devices. In this connected age, the question is no longer how do you gather customer data, but what do you do with all that data. How do you ingest at massive scale and develop meaningful experiences for your customers? In this session, you learn how Salesforce IoT Cloud works in concert with the AWS IoT engine to ingest and transform all of the data generated by every one of your customers, partners, devices, and sensors into meaningful action. You also see how customers are using Salesforce and AWS together to process massive quantities of data, build business rules with simple, intuitive tools, and engage proactively with customers in real time. Session sponsored by Salesforce.
BDA206 - Building Big Data Applications with the AWS Big Data Platform
Building big data applications often requires integrating a broad set of technologies to store, process, and analyze the increasing variety, velocity, and volume of data being collected by many organizations. In this session, we show how you can build entire big data applications using a core set of managed services including Amazon S3, Amazon Kinesis, Amazon EMR, Amazon Elasticsearch Service, Amazon Redshift, and Amazon QuickSight. We walk you through the steps of building and securing a big data application using the AWS Big Data Platform. We also share best practices and common use cases for AWS big data services, including tips to help you choose the best services for your specific application.
BDA207 - Fanatics: Deploying Scalable, Self-Service Business Intelligence on AWS
Data is growing at a quantum scale and one of challenges you face is to enable your users to analyze all this data, extract timely insights from it, and visualize it. In this session, you learn about business intelligence solutions available on AWS. We discuss best practices for deploying a scalable and self-serve BI platform capable of churning through large datasets. Fanatics, the nations largest online seller of licensed sports apparel, talks about their experience building a globally distributed BI platform on AWS, that delivers massive volumes of reports, dashboards, and charts on a daily basis to an ever growing user base. Fanatics shares the architecture of their data platform, built using Amazon Redshift, Amazon S3, and open source frameworks like Presto and Spark. They talk in detail about their BI platform including Tableau, Microstrategy, and other tools on AWS to make it easy for their analysts to perform ad-hoc analysis and get real-time updates, alerts, and visualizations. You also learn about the experimentation-based approach that Fanatics adopted to fully engage their business intelligence community and make optimal use of their BI platform resources on AWS.
BDA209 - NEW LAUNCH! Introducing AWS Glue: A Fully Managed ETL Service
AWS Glue is a fully managed ETL service that makes it easy to understand your data sources, prepare the data for analytics, and load it reliably to your data stores. In this session, we will introduce AWS Glue, provide an overview of its components, and discuss how you can use the service to simplify and automate your ETL process. We will also talk about when you can try out the service and how to sign up for a preview.
BDA303 - NEW LAUNCH! Intro to Amazon Athena. Easily analyze data in S3, using SQL.
Amazon Athena is a new interactive query service that makes it easy to analyze data in Amazon S3, using standard SQL. Athena is serverless, so there is no infrastructure to setup or manage, and you can start analyzing your data immediately. You dont even need to load your data into Athena, it works directly with data stored in S3. In this session, we will show you how easy is to start querying your data stored in Amazon S3, with Amazon Athena. First we will use Athena to create the schema for data already in S3. Then, we will demonstrate how you can run interactive queries through the built-in query editor. We will provide best practices and use cases for Athena. Then, we will talk about supported queries, data formats, and strategies to save costs when querying data with Athena.
BDA304 - What’s New with Amazon Redshift
In this session, you learn about the latest and hottest features of Amazon Redshift. Join Vidhya Srinivasan, General Manager of Amazon Redshift, to take a deep dive into the architecture and inner workings of Amazon Redshift. You discover how the recent availability, performance, and manageability improvements weve made can significantly enhance your end user experience. You also get a glimpse of what we are working on and our plans for the future.
BDM201 - Big Data Architectural Patterns and Best Practices on AWS
The world is producing an ever increasing volume, velocity, and variety of big data. Consumers and businesses are demanding up-to-the-second (or even millisecond) analytics on their fast-moving data, in addition to classic batch processing. AWS delivers many technologies for solving big data problems. But what services should you use, why, when, and how? In this session, we simplify big data processing as a data bus comprising various stages: ingest, store, process, and visualize. Next, we discuss how to choose the right technology in each stage based on criteria such as data structure, query latency, cost, request rate, item size, data volume, durability, and so on. Finally, we provide reference architecture, design patterns, and best practices for assembling these technologies to solve your big data problems at the right cost.
BDM202 - Workshop: Building Your First Big Data Application with AWS
Want to get ramped up on how to use Amazon's big data web services and launch your first big data application on AWS? Join us in this workshop as we build a big data application in real time using Amazon EMR, Amazon Redshift, Amazon Kinesis, Amazon DynamoDB, and Amazon S3. We review architecture design patterns for big data solutions on AWS, and give you access to a take-home lab so that you can rebuild and customize the application yourself.
BDM203 - FINRA: Building a Secure Data Science Platform on AWS
Data science is a key discipline in a data-driven organization. Through analytics, data scientists can uncover previously unknown relationships in data to help an organization make better decisions.However, data science is often performed from local machines with limited resources and multiple datasets on a variety of databases. Moving to the cloud can help organizations provide scalable compute and storage resources to data scientists, while freeing them from the burden of setting up and managing infrastructure. In this session, FINRA, the Financial Industry Regulatory Authority, shares best practices and lessons learned when building a self-service, curated data science platform on AWS. A project that allowed us to remove the technology middleman and empower users to choose the best compute environment for their workloads. Understand the architecture and underlying data infrastructure services to provide a secure, self-service portal to data scientists, learn how we built consensus for tooling from of our data science community, hear about the benefits of increased collaboration among the scientists due to the standardized tools, and learn how you can retain the freedom to experiment with the latest technologies while retaining information security boundaries within a virtual private cloud (VPC).
BDM204 - Visualizing Big Data Insights with Amazon QuickSight
Amazon QuickSight is a fast BI service that makes it easy for you to build visualizations, perform ad-hoc analysis, and quickly get business insights from your data. QuickSight is built to harness the power and scalability of the cloud, so you can easily run analysis on large datasets, and support hundreds of thousands of users. In this session, well demonstrate how you can easily get started with Amazon QuickSight, uploading files, connecting to S3 and Redshift and creating analyses from visualizations that are optimized based on the underlying data. Once weve built our analysis and dashboard, well show you easy it is to share it with colleagues and stakeholders in just a few seconds. And with SPICE QuckSights in-memory calculation engine you can go from data to insights, faster than ever.
BDM205 - Big Data Mini Con State of the Union
Join us for this general session where AWS big data experts present an in-depth look at the current state of big data. Learn about the latest big data trends and industry use cases. Hear how other organizations are using the AWS big data platform to innovate and remain competitive. Take a look at some of the most recent AWS big data announcements, as we kick off the Big Data re:Source Mini Con.
BDM206 - Understanding IoT Data: How to Leverage Amazon Kinesis in Building an IoT Analytics Platform on AWS
The growing popularity and breadth of use cases for IoT are challenging the traditional thinking of how data is acquired, processed, and analyzed to quickly gain insights and act promptly. Today, the potential of this data remains largely untapped. In this session, we explore architecture patterns for building comprehensive IoT analytics solutions using AWS big data services. We walk through two production-ready implementations. First, we present an end-to-end solution using AWS IoT, Amazon Kinesis, and AWS Lambda. Next, Hello discusses their consumer IoT solution built on top of Amazon Kinesis,Amazon DynamoDB, and Amazon Redshift.
BDM301 - Best Practices for Apache Spark on Amazon EMR
Organizations need to perform increasingly complex analysis on data streaming analytics, ad-hoc querying, and predictive analytics in order to get better customer insights and actionable business intelligence. Apache Spark has recently emerged as the framework of choice to address many of these challenges. In this session, we show you how to use Apache Spark on AWS to implement and scale common big data use cases such as real-time data processing, interactive data science, predictive analytics, and more. We talk about common architectures, best practices to quickly create Spark clusters using Amazon EMR, and ways to integrate Spark with other big data services in AWS. This session will feature DataXu, a provider of programmatic marketing and analytics software. DataXu will share how they architected their petabyte-scale ETL processing pipeline and data science workflows using Spark.
BDM302 - Real-Time Data Exploration and Analytics with Amazon Elasticsearch Service and Kibana
Elasticsearch is a fully featuredsearch engine used for real-time analytics, and Amazon Elasticsearch Service makes it easy to deploy Elasticsearch clusters on AWS. With Amazon ES, you can ingest and process billions of events per day, and explore the data using Kibana to discover patterns. In this session, we use Apache web logs as example and show you how to build an end-to-end analytics solution. First, we cover how to configure an Amazon ES cluster and ingest data into it using Amazon Kinesis Firehose. We look at best practices for choosing instance types, storage options, shard counts, and index rotations based on the throughput of incoming data. Then we demonstrate how to set up a Kibana dashboard and build custom dashboard widgets. Finally, we dive deep into the Elasticsearch query DSL and review approaches for generating custom, ad-hoc reports.
BDM303 - JustGiving: Serverless Data Pipelines, Event-Driven ETL, and Stream Processing
Organizations need to gain insight and knowledge from a growing number of Internet of Things (IoT), application programming interfaces (API), clickstreams, unstructured and log data sources. However, organizations are also often limited by legacy data warehouses and ETL processes that were designed for transactional data. Building scalable big data pipelines with automated extract-transform-load (ETL) and machine learning processes can address these limitations. JustGiving is the worlds largest social platform for online giving. In this session, we describe how we created several scalable and loosely coupled event-driven ETL and ML pipelines as part of our in-house data science platform called RAVEN. You learn how to leverage AWS Lambda, Amazon S3, Amazon EMR, Amazon Kinesis, and other services to build serverless, event-driven, data and stream processing pipelines in your organization. We review common design patterns, lessons learned, and best practices, with a focus on serverless big data architectures with AWS Lambda.
BDM304 - Analyzing Streaming Data in Real-time with Amazon Kinesis Analytics
As more and more organizations strive to gain real-time insights into their business, streaming data has become ubiquitous. Typical streaming data analytics solutions require specific skills and complex infrastructure. However, with Amazon KinesisAnalytics, you can analyze streaming data in real-time with standard SQLthere is no need to learn new programming languages or processing frameworks. In this session, we dive deep into the capabilities of Amazon Kinesis Analytics using real-world examples. Well present an end-to-end streaming data solution using Amazon KinesisStreams for data ingestion, Amazon KinesisAnalytics for real-time processing, and Amazon KinesisFirehose for persistence. We review in detail how to write SQL queries using streaming data and discuss best practices to optimize and monitor your Amazon KinesisAnalytics applications. Lastly, we discuss how to estimate the cost of the entire system.
BDM306 - Netflix: Using Amazon S3 as the fabric of our big data ecosystem
Amazon S3 is the central data hub for Netflix's big data ecosystem. We currently have over 1.5 billion objects and 60+ PB of data stored in S3. As we ingest, transform, transport, and visualize data, we find this data naturally weaving in and out of S3. Amazon S3 provides us the flexibility to use an interoperable set of big data processing tools like Spark, Presto, Hive, and Pig. It serves as the hub for transporting data to additional data stores / engines like Teradata, Redshift, and Druid, as well as exporting data to reporting tools like Microstrategy and Tableau. Over time, we have built an ecosystem of services and tools to manage our data on S3. We have a federated metadata catalog service that keeps track of all our data. We have a set of data lifecycle management tools that expire data based on business rules and compliance. We also have a portal that allows users to see the cost and size of their data footprint. In this talk, well dive into these major uses of S3, as well as many smaller cases, where S3 smoothly addresses an important data infrastructure need. We will also provide solutions and methodologies on how you can build your own S3 big data hub.
BDM401 - Deep Dive: Amazon EMR Best Practices Design Patterns
Amazon EMR is one of the largest Hadoop operators in the world. In this session, we introduce you to Amazon EMR design patterns such as using Amazon S3 instead of HDFS, taking advantage of both long and short-lived clusters, and other Amazon EMR architectural best practices. We talk about how to scale your cluster up or down dynamically and introduce you to ways you can fine-tune your cluster. We also share best practices to keep your Amazon EMR cluster cost-efficient. Finally, we dive into some of our recent launches to keep you current on our latest features. This session will feature Asurion, a provider of device protection and support services for over 280 million smartphones and other consumer electronics devices.Asurion will share how they architected their petabyte-scale data platform using Apache Hive, Apache Spark, and Presto on Amazon EMR.
BDM402 - Best Practices for Data Warehousing with Amazon Redshift
Analyzing big data quickly and efficiently requires a data warehouse optimized to handle and scale for large datasets. Amazon Redshift is a fast, petabyte-scale data warehouse that makes it simple and cost-effective to analyze all of your data for a fraction of the cost of traditional data warehouses. In this session, we take an in-depth look at data warehousing with Amazon Redshift for big data analytics. We cover best practices totake advantage of Amazon Redshift's columnar technology and parallel processing capabilities to deliver high throughput and query performance. We also discuss how to design optimal schemas,loaddata efficiently, and usework load management.
BDM403 - Beeswax: Building a Real-Time Streaming Data Platform on AWS
Amazon Kinesis is a platform of services for building real-time, streaming data applications in the cloud. Customers can use Amazon Kinesis to collect, stream, and process real-time data such as website clickstreams, financial transactions, social media feeds, application logs, location-tracking events, and more. In this session, we first cover best practices for building an end-to-end streaming data applications using Amazon Kinesis. Next, Beeswax, which provides real-time Bidder as a Service for programmatic digital advertising, will talk about how they built a feature-rich, real-time streaming data solution on AWS using Amazon Kinesis, Amazon Redshift, Amazon S3, Amazon EMR, and Apache Spark. Beeswax will discuss key components of their solution including scalable data capture, messaging hub for archival, data warehousing, near real-time analytics, and real-time alerting.
BAP201 - Move your desktops to the cloud with Amazon WorkSpaces
Organizations today are striving to provide a more flexible environment for their end users, allowing them access to corporate resources from any device, anytime. At the same time, securing corporate information remains a top priority for CIOs. Traditional solutions such as Virtual Desktop Infrastructure (VDI) offer worker flexibility and security benefits, but they are expensive to purchase, complex to deploy, and they dont scale well. In this session, youll learn how Amazon WorkSpaces combines the benefits of VDI with the economics of the cloud to deliver better value for organizations. Well show you just how easy it is to get started, and talk through how customers are using Amazon WorkSpaces today.
BAP202 - Extra, Extra! News UK Saves Money with Hourly Amazon WorkSpaces, Read All About It!
Amazon WorkSpaces is a desktop computing service that runs in the cloud, and now supports both monthly and hourly billing. In this session, we help you determine the right billing method for your use cases, show you how to provision Amazon WorkSpaces for monthly or hourly billing, and work through a real-world example. News UK, a media organization that owns a stable of news and media brands, describes how they use Amazon WorkSpaces to solve a unique business need.
BAP203 - How Higher Ed Graduated from Traditional Desktops to DaaS Using Amazon WorkSpaces
Join leaders in higher education in this panel discussion as they share how they use Amazon WorkSpaces to move to desktop-as-a-service (DaaS) to provide faculty, staff, and students, access to applications and information they need in classrooms, research labs, and across campus. Amazon WorkSpaces enables major educational institutions be more agile, improve their security posture, and offer end users a more flexible experience, while meeting stringent compliance requirements and remaining cost effective. Hear directly from the world of higher education about their projects to deliver the next generation of end-user computing in their organizations.
BAP204 - Deliver Desktop Applications to Any Device, Anywhere with Amazon AppStream 2.0
Amazon AppStream 2.0 is an application streaming service that provides users instant access to their applications from anywhere, on any connected device. Enterprises can use Amazon AppStream 2.0 to securely deliver the latest versions of their desktop apps to their users. ISVs can reach more users across more devices, without rewriting applications, by streaming their desktop apps from the cloud to their users browsers. In this session, we show you how easy it is to upload your apps to Amazon AppStream and start delivering them to thousands of users. Well also learn from Siemens PLM how they are moving their desktop apps to the cloud with Amazon AppStream 2.0, and enabling their enterprise customers greater flexibility and productivity through streamed applications.
BAP205 - Stop Managing Email Infrastructure: Move to Amazon WorkMail
Amazon WorkMail is a secure, managed business email and calendar service with support for existing desktop and mobile email clients. Amazon WorkMail enables users to access email, calendars, and contacts using the Amazon WorkMail web app, Microsoft Outlook, or native iOS or Android email applications. In this session, well show you how to quickly create an Amazon WorkMail organization, take a look at the encryption model employed by Amazon WorkMail for data at rest, set up interoperability between existing on-premises Microsoft Exchange servers and Amazon WorkMail and dive into a host of new features that are now available on Amazon WorkMail.
BAP206 - Store and collaborate on content securely with Amazon WorkDocs and use the SDKs to integrate with your existing IT tools
File sync and share applications have become a cornerstone of collaboration in modern organizations; securing and managing content is an ongoing challenge. Amazon WorKDocs is a fully managed, secure file storage and sharing service with strong administrative controls and feedback capabilities. In this session, well demonstrate how you can use Amazon WorkDocs to manage files and secure content, and show you how were shaping Amazon WorkDocs to be a full-fledged productivity tool for users. We'll also present a technical overview of the WorkDocs SDKs, and how they allow you to use existing IT tools for antivirus protection, data loss prevention, analystics, and more.
BAP210 - Hardware-Accelerating Graphics Desktop Workloads with Amazon WorkSpaces
Amazon WorkSpaces is a desktop computing service that runs in the cloud, and now offers GPU configurations to support design and engineering applications and three-dimensional modeling. We show you how running these applications on Amazon WorkSpaces graphics bundles, in close proximity to data you already store on AWS, can help you process and visualize the results you need. We discuss the economics of running Amazon WorkSpaces graphics bundles, and demonstrate the experience of running a graphics-intensive application on a GPU-enabled Amazon WorkSpace. We also invite Autodesk (or TRC or ESRi) to discuss how they are using Amazon WorkSpaces graphics bundles in their business.
BAP212 - Deliver Engaging Experiences with Custom Apps Built on Salesforce App Cloud and AWS
Your developers are the most important part of transforming your customer interactions into engaging experiences. Salesforce App Cloud, which brings together Heroku, Force.com and Lightning, abstracts away infrastructure and devops complexity, so you can focus on what matters most: building differentiated experiences through apps. Reducing time to market and letting you iterate fast helps you rise above the competition and build lasting customer relationships. In this session, you hear from Zayo, aleading global communications infrastructure services provider, and how they are leveraging the power of integrating the Salesforce and AWS platforms to deliver highly engaging customer experiences, enhancing developer productivity and driving faster innovation cycles. We spotlight Heroku Connect, which makes it easy to extend and synchronize your customer data between Salesforce and AWS and enhance it in ways that empower your developers to do what they do best: innovate. Session sponsored by Salesforce.
BAP305 - Zero to Google Chrome in 60 Minutes: Lightweight and Inexpensive Client Devices for Amazon WorkSpaces
Youve bet big with Amazon WorkSpaces to remove challenges managing your physical fleet of Macs and PCs. Now what? In this session, well demonstrate how you can deploy a rich cloud-based Windows experience on lightweight hardware to reign in management issues, improve TCO, and be at parity with your traditional environment. Well take you through the client device ecosystem from Zero to thin to Google Chrome and Chromium OS clients and strengthen your ability to determine the right client device strategy moving forward. Live product demonstrations will be provided as we journal how customers are moving to lightweight devices, and what best practices weve learned along the way.
BAP402 - Best Practices from the Trenches: Deploy Amazon WorkSpaces Like a Pro
Amazon WorkSpaces is an enterprise desktop computing service that runs in the cloud. In this deep-dive session, we discuss advanced topics and best practices for deploying Amazon WorkSpaces in your organization. We discuss Amazon VPC design and public endpoints, AWS Directory Service, integrating with your on-premises Microsoft Active Directory, using multi-factor authentication, and monitoring and logging with Amazon CloudWatch metrics. We walk through how to do all this using a combination of the AWS Command Line Interface, the AWS Management Console, and AWS CloudFormation templates.
CMP201 - Auto Scaling – the Fleet Management Solution for Planet Earth
Scaling allows cloud resources to scale automatically in reaction to the dynamic needs of customers. This session will show how Auto Scaling offers an advantage to everyone whether its basic fleet management to keep instances healthy as an EC2 best practice, or dynamic scaling to manage extremes. Well share examples of how Auto Scaling is helping customers of all sizes and industries unlock use cases and value. Well also discuss how Auto Scaling is evolving to scaling different types of elastic AWS resources beyond EC2 instances. NASA Jet Propulsion Laboratory (JPL) / California Institute of Technology will share how Auto Scaling is used to scale science data processing of Interferometric Synthetic Aperture Radar (InSAR) data from earth-observing satellite missions, and reduce response times during hazard response events such as those from earthquakes, floods, and volcanoes. JPL will also discuss how they are integrating their science data systems with the AWS ecosystem to expand into NASAs next two large-scale missions with remote-sensing radar-based observations. Learn how Auto Scaling is being used at a global scale and beyond!
CMP202 - Getting the most Bang for your buck with #EC2 #Winning
Amazon EC2 provides you with the flexibility to cost optimize your computing portfolio through purchasing models that fit your business needs. With the flexibility of mix-and-match purchasing models, you can grow your compute capacity and throughput and enable new types of cloud computing applications with the lowest TCO. In this session, we will explore combining pay-as-you-go (On-Demand), reserve ahead of time for discounts (Reserved), and high-discount spare capacity (Spot) purchasing models to optimize costs while maintaining high performance and availability for your applications. Common application examples will be used to demonstrate how to best combine EC2s purchasing models. You will leave the session with best practices you can immediately apply to your application portfolio.
CMP203 - Amazon EC2 Foundations
Amazon EC2 changes the economics of computing and provides you with complete control of your computing resources. It is designed to make web-scale cloud computing easier for developers. In this session, we will take you on a journey, starting with the basics of key management and security groups and ending with an explanation of Auto Scaling and how you can use it to match capacity and costs to demand using dynamic policies. We will also discuss tools and best practices that will help you build failure resilient applications that take advantage of the scale and robustness of AWS regions
CMP205 - Lessons Learned from a Year of Using Spot Fleet
Over the last year, Yelp has transitioned its scalable and reliable parallel task execution system, Seagull, from On-Demand and Reserved Instances entirely to Spot Fleet. Seagull runs over 28 million tests per day, launches more than 2.5 million Docker containers per day, and uses over 10,000 vCPUs in Spot Fleet at peak capacity. To deal with rising infrastructure costs for Seagull, we have extended our in-house Auto Scaling Engine called FleetMiser to scale the Spot Fleet in response to demand. FleetMiser has reduced Seagulls cluster costs by 60% in the past year and saved Yelp thousands of dollars every month. In this session, we describe how Yelp uses Spot Fleet for Seagull and lessons weve learned over the past year, along with our recommendations on how to use it reliably (pro tip: dont get outbid for your whole Spot Fleet). We conclude by looking at our future plans for extending Spot Fleet usage at Yelp.
CMP206 - NEW LAUNCH! Introducing Amazon Lightsail: Jumpstart Your Cloud Project for a Low, Predictable Price
Amazon Lightsail is the latest addition to the AWS family of compute services and the fastest way to get your next cloud server up and running. For a low price that starts at $5/month, Lightsail offers a bundle of resources and services that let you jumpstart your cloud project in a few clicks. The new, intuitive Lightsail console makes it simple to manage your virtual resources, letting you focus on code, not system administration. Come to this session and learn how Lightsail can get you started on AWS quickly and efficiently.
CMP207 - High Performance Computing on AWS
High performance computing in the cloud is enabling high scale compute- and graphics-intensive workloads across industries, ranging from aerospace, automotive, and manufacturing to life sciences, financial services, and energy. AWS provides application developers and end users with unprecedented computational power for massively parallel applications, in areas such as large-scale fluid and materials simulations, 3D content rendering, financial computing, and deep learning. This session provides an overview of HPC capabilities on AWS, describes the newest generations of accelerated computing instances (including P2), as well as highlighting customer and partner use-cases across industries. Attendees learn about best practices for running HPC workflows in the cloud, including graphical pre- and post-processing, workflow automation, and optimization. Attendees also learn about new and emerging HPC use cases: in particular, deep learning training and inference, large-scale simulations, and high performance data analytics.
CMP209 - Getting Started with Docker on AWS
AWS is an elastic, secure, flexible, and developer-centric ecosystem that serves as an ideal platform for Docker deployments. AWS offers the scalable infrastructure, APIs, and SDKs that integrate tightly into a development lifecycle and accentuate the benefits of the lightweight and portable containers that Docker offers to its users. This session familiarizes you with the benefits of containers, introduce Amazon EC2 Container Service, and demonstrates how to use Amazon ECS to run containerized applications at scale in production.
CMP211 - Getting Started with Serverless Architectures
Serverless architectures let you build and deploy applications and services with infrastructure resources that require zero administration. In the past, you had to provision and scale servers to run your application code, install and operate distributed databases, and build and run custom software to handle API requests. Now, AWS provides a stack of scalable, fully-managed services that eliminates these operational complexities. In this session, you learn about the concepts and benefits of serverless architectures and the basics of the serverless stack AWS provides (e.g., AWS Lambda and Amazon API Gateway). We discuss use cases such as data processing, website backends, serverless applications and operational glue. After that, you get practical tips and tricks, best practices, and architecture patterns that you can take back and implement immediately.
CMP212 - Host a massively scalable website around the world for a fraction of the cost in a DevOps model
A worldwide dynamic website infrastructure at the cost and simplicity of a static deployment was once an unachievable dream. Thats no longer the case with the tools and services that AWS offers like Amazon CloudFront, Amazon S3, Amazon API Gateway and AWS Lambda. During this session we will discuss a DevOps approach to managing your code repositories, code quality, continuous integration and deployments. We discuss some static website generators such as mkdocs, Jekyll, and middleman for building knowledge bases, blogs and websites in a serverless architecture. Featuring an ideal use case for implementing a serverless architecture combining Amazon CloudFront, Amazon S3 and Amazon API Gateway; we explore how we deployed a massively scaled web application for a fraction of the cost of traditional dynamic websites leveraging a world-wide CDN. Going serverless significantly reduced operational costs and complexity because the infrastructure scales on demand without having to manually provision new servers on the ready. Finally, we share the lessons learned and the advantages of this move, leveraging Cloudnexas vNOC product which allows for a faster, more affordable user experience. Session sponsored byCloudnexa. AWS Competency Partner
CMP213 - Case Study: Librato's Experience Running Cassandra Using Amazon EBS, ENIs, and Amazon VPC
At Librato, a Solarwinds company, we run hundreds of Cassandra instances across multiple rings and use it as our primary data store. In the past year, we embarked on a process to upgrade our fleet of Cassandra Amazon EC2 instances from instance store to instances using Amazon EBS and attached elastic network interfaces (ENIs). We find running Cassandra on EBS gives us the flexibility to choose the best instances for the best performance of our workload while saving us significant costs on infrastructure. In this session, we discuss how Librato operates Cassandra on EBS. Topics include how we chose the right instance for our workload, use detached EBS volumes and ENI mobility to reduce MTTR, use mixed EBS storage types for the best cost/performance tradeoff, debug performance issues, and continuously monitor Cassandra to get the most from AWS. We also look at performance tradeoffs made in the implementation of storage engines of large data systems like Cassandra.
CMP301 - Deep Dive on Amazon EC2 Instances, Featuring Performance Optimization Best Practices
Amazon EC2 provides a broad selection of instance types to accommodate a diverse mix of workloads. In this session, we provide an overview of the Amazon EC2 instance platform, key platform features, and the concept of instance generations. We dive into the current generation design choices of the different instance families, including the General Purpose, Compute Optimized, Storage Optimized, Memory Optimized, and GPU instance families. We also detail best practices and share performance tips for getting the most out of your Amazon EC2 instances.
CMP302 - Disrupting Big Data with Cost-effective Compute
Amazon EC2 Spot instances provide acceleration, scale, and deep cost savings to run time-critical, hyper-scale workloads for rapid data analysis. In this session, AOL and Metamarkets will present lessons learned and best practices from scaling their big data workloadsusing popular platforms like Presto, Spark and Druid. AOL will present how they process, store, and analyze big data securely and cost effectively using Presto. AOL achieved 70% savings by separating compute and storage, dynamically resizing clusters based on volume and complexity, and usingAWSLambda to orchestrate processing pipelines. Metamarkets, an industry leader in interactive analytics, will present how they leverage Amazon EBS to persist 185 TiB of (compressed) state to run Druid historical nodes on EC2 Spot instances. They will also cover how they run Spark for batch jobs to process 1-4 PiB of data across 200 B to 1 T events/day, saving more than 60% in costs.
CMP304 - T2: From Startups to Enterprise, Performance for a Low Cost
In this session, customers learn more about the T2 instance type and the performance and cost savings it can bring to startups, SMBs, and enterprises. Customers will share best practices and tips for how they use T2 instances across workloads including development and test, production web servers, continuous integration and more.
CMP305 - Serverless to 32 XLarge: A Unified Security Approach To AWS Compute
Each of the four distinct compute services available from AWS bring unique advantages to your application's design: Amazon EC2, AWS Elastic Beanstalk, Amazon ECS, and AWS Lambda. However, your security responsibilities change with each of these services. For example, with Amazon EC2 and AWS Elastic Beanstalk, you need a plan to lock down the operating system, your applications, and your data. Containers running in Amazon ECS provides additional flexibility and mobility but also introduces new challenges as most security techniques were not designed with containers in mind. AWS Lambda requires a completely new approach to security design at the function level. Making sure that the diversity available in AWS compute services doesn't increase your operational burden is a significant challenge. But with the right strategy, you can use a unified approach and reduce team impact. In this session, we look at an overall security strategy for AWS compute services pulled from real-world experiences of some top global companies. Strategy in hand, we apply the strategy to each of the available compute services in order to create a unified security view for workloads, regardless of their scale. Session sponsored byTrend Micro, Inc. AWS Competency Partner
CMP306 - Apache Spark on EC2 History, Best Practices with Customer Use Cases
Apache Spark is well known across industries, use cases and businesses of all sizes for its speed and ease of use in sophisticated analysis of large datasets. In this session, learn from Ion Stoica who co-led the Apache Spark project at the AMPLab (UC Berkeley) and co-founder of Databricks, about some of the latest innovations in Spark 2.0, a new open source tool Earnest to choose the optimal cluster configuration for your job, and how and why Databricks choose EC2 to run Spark. Well also take a look at how Amazon EC2 and the latest enhancements enable Sparks as a data processing platform, along-with best practices and cost optimization techniques for using Spark with AWS.
CMP307 - Save up to 90% and Run Production Workloads on Spot - Featuring IFTTT and Mapbox
Amazon EC2 allows you to bid for and run spare EC2 capacity, known as Spot instances, in a dynamically priced market. On average, customers save 80% to 90% compared to On Demand prices by using Spot instances. Achieving these savings has historically required time and effort to find the best deals while managing compute capacity as supply and demand fluctuate. In this session, we dive into how customers who have designed scalable, cloud friendly application architectures can leverage new Spot features to realize immediate cost savings while maintaining availability. Attendees will leave with practical knowledge of how, via well architected applications, they can run production services on the Spot instances just like IFTTT and Mapbox.
CMP311 - The Future of Cloud: Building with Stateless Infrastructure on Amazon EC2
ProtectWise has been hailed as one of the top 10 coolest startups of 2016, disrupting the security industry by providing cutting edge cloud-based network threat detection, forensics and analytics for Fortune 2000 customers. By developing a scalable, stateless platform, they are able to quickly right-size Amazon EC2 instances to optimize performance and ingest gigabytes of network packet capture at a fraction of the cost. Learn how their cloud-first approach towards infrastructure and application development has enabled them to be more agile, grow faster, and save money. In this session, attendees gain practical best practices for how to build cloud-ready applications and see real-life examples of how this hot startup applied those to common workloads like Cassandra.
CMP312 - Powering the Next Generation of Virtual Reality with Verizon
In six months, Verizon has built a best-in-class Augmented Reality and Virtual Reality (AR/VR) platform that streams HD video and game experiences using Amazon EC2 GPU Accelerated instances and CloudFront. Verizon will share their reference architecture and configuration best practices that enabled them to develop a massively scalable VR architecture that scales to support for 100K simultaneous HD video streams to customers around the globe.
CMP313 - Revolutionizing Car Buying with 3D Rendering on Amazon EC2
AWS GPU computing capabilities have allowed hot start-ups like Zerolight to revolutionize the car buying experience for clients like Audi, by providing advanced 3D car rendering simulations for their customers interested in personalizing cars in real-time at dealer showrooms and online. Zerolight chose to build their company from the ground up on AWS because of GPU-compute capabilities, scale, elasticity, and reliability, allowing customers to perform complex renderings quickly and in real-time. In this session, learn how Zerolight has implemented a scalable, cost-effective, highly responsive 3D rendering platform using Amazon EC2 G2 instances.
CMP314 - Bringing Deep Learning to the Cloud with Amazon EC2
Algorithmia is a startup with a mission to make state of the art machine learning discoverable by everyonethey offer the largest algorithm marketplace in the world, with over 2500 algorithms supporting tens of thousands of application developers. Algorithma is the first company to make deep learning, one of the most conceptually difficult areas of computing, accessible to any company via microservices. In this session, you learn how this startup has selected and optimized Amazon EC2 instances for various algorithms (including the latest generation of GPU optimized instances), to create a flexible and scalable platform. They also share their architecture and best practices for getting any computationally-intensive application started quickly.
CMP315 - Optimizing Network Performance for Amazon EC2 Instances
Many customers are using Amazon EC2 instances to run applications with high performance networking requirements. In this session, we provide an overview of Amazon EC2 network performance features (enhanced networking, ENA, placement groups, etc.), and discuss how we are innovating on behalf of our customers to improve networking performance in a scalable and cost-efficient manner. We share best practices and performance tips for getting the best networking performance out of your Amazon EC2 instances.
CMP316 - Learn How FINRA Aligns Billions of Time Ordered Events with Spark on EC2
FINRA is a leader in the Financial Services industry who sought to move toward real-time data insights of billions of time-ordered market events by migrating from SQL batch processes on-prem, to Apache Spark in the cloud. By using Apache Spark on Amazon EMR, FINRA can now test on realistic data from market downturns, enhancing their ability to provide investor protection and promote market integrity (FINRA enacts rules and provides guidance that securities exchanges brokers must follow). By using AWS Spot instances, FINRA has saved up to 50% from its on premises solution, increased elasticity/scalability, and accelerated reprocessing requests (from months to days). Learn best practices on how FINRA moves toward real-time data analytics with Spark and AWS, while managing production workloads in parallel, increasing performance and IT efficiency, reducing cost, and modernizing and scaling their infrastructure to prepare for real-time processing in the future.
CMP317 - Massively Parallel, Compute Intensive Workloads in the Cloud -- Choosing the right hardware accelerator and example use cases
Accelerated computing is on the rise because of massively parallel, compute-intensive workloads such as deep learning, 3D content rendering, financial computing, and engineering simulations. In this session, we provide an overview of our accelerated computing instances, including how to choose instances based on your application needs, best practices and tips to optimize performance, and specific examples of accelerated computing in real-world applications.
CMP318 - Building HPC Clusters as Code in the (Almost) Infinite Cloud
Every day, the computing power of high-performance computing (HPC) clusters helps scientists make breakthroughs, such as proving the existence of gravitational waves and screening new compounds for new drugs. Yet building HPC clusters is out of reach for most organizations, due to the upfront hardware costs and ongoing operational expenses. Now the speed of innovation is only bound by your imagination, not your budget. Researchers can run one cluster for 10,000 hours or 10,000 clusters for one hour anytime, from anywhere, and both cost the same in the cloud. And with the availability of Public Data Sets in Amazon S3, petabyte scale data is instantly accessible in the cloud. Attend and learn how to build HPC clusters on the fly, leverage Amazons Spot market pricing to minimize the cost of HPC jobs, and scale HPC jobs on a small budget, using all the same tools you use today, and a few new ones too.
CMP319 - NEW LAUNCH! Building Distributed Applications with AWS Step Functions
AWS Step Functions is a new, fully managed service that makes it easy to coordinate the components of distributed applications and microservices using visual workflows. Step Functions is a reliable way to coordinate components and step through the functions of your application. A graphical console helps you arrange and visualize the components of your application as a series of steps. Step Functions automatically triggers and tracks each step and retries when there are errors so that your application executes in order―and as expected―every time. This session shows how to use Step Functions to create, run, and debug multi-service applications in a matter of minutes. We also share how customers are using Step Functions to reliably build and scale multi-step applications such as order processing, report generation, and data transformation―and to innovate faster.
CMP320 - NEW LAUNCH! Delivering Powerful Graphics-Intensive Applications from the AWS Cloud
AWS provides unprecedented computational power for graphics-intensive applications in areas such as design, engineering simulations, and 3D content rendering. Together, Amazon EC2 Elastic GPUs and Amazon AppStream 2.0 provide the capabilities necessary for end users to access and run these applications. In this session, you learn more about Elastic GPUs and Amazon AppStream 2.0, and how you can run graphics-intensive applications on AWS. You also hear from ANSYS, a leader in engineering simulation software, and why they are moving the ANSYS Enterprise Cloud to Elastic GPUs and Amazon AppStream 2.0 to deliver a better experience for customers.
CMP321 - Building SaaS Offerings for Desktop Apps with Amazon AppStream 2.0
Amazon AppStream 2.0 is an application streaming service that provides users with instant access to their applications from anywhere, on any connected device. Delivering desktop apps from the cloud allows enterprises to manage user access easily and keep apps current; ISVs can reach more users across more devices. In this session, we cover best practices for streaming desktop applications from the AWS Cloud with Amazon AppStream 2.0, and show you how you can address advanced topics such as user authentication, security, and managing an application catalog.
CMP322 - Optimizing workloads in SAP HANA with Amazon EC2 X1 Instances
AWS and SAP have worked together closely to certify the AWS platform so that companies of all sizes can fully realize all the benefits of the SAP HANA in-memory database platform on the AWS cloud. By placing SAP systems in the cloud, organizations are achieving greater agility, flexibility, and cost efficiency while saving resources to focus on their core businesses. We will discuss recent SAP and AWS innovations including the Amazon EC2 X1 instance type that offers up to 2TB of RAM, and dive into features of the AWS platform that bring significant flexibility to SAP HANA deployments.
CMP323 - NEW LAUNCH! Introducing AWS Batch: Easy and efficient batch computing on Amazon Web Services
AWS Batch is a fully-managed service that enables developers, scientists, and engineers to easily and efficiently run batch computing workloads of any scale on AWS. AWS Batch automatically provisions compute resources and optimizes the workload distribution based on the quantity and scale of the workloads. With AWS Batch, there is no need to install or manage batch computing software, allowing you to focus on analyzing results and solving problems. AWS Batch plans, schedules, and executes your batch computing workloads across the full range of AWS compute services and features, such as Amazon EC2, Spot Instances, and AWS Lambda. AWS Batch reduces operational complexities, saving time and reducing costs. In this session, Principal Product Managers Jamie Kinney and Dougal Ballantyne describe the core concepts behind AWS Batch and details of how the service functions. The presentation concludes with relevant use cases and sample code.
CON301 - Operations Management with Amazon ECS
Managing the infrastructure behind your ECS cluster is an important and crucial task. In this session we will dive deep into best practices to setup and maintain your ECS cluster with tools such as CloudFormation and Auto Scaling. We will discuss how to setup monitoring and logging infrastructure for your cluster and containers. We will also detail how to tighten the security of your container instances with tools such as SELinux. We will do a live deployment of a complete microservices-based ecommerce website, using Weave Net and Weave Scope for container networking, automated service discovery, monitoring and real-time app visualization. Well use CloudFormation, Auto Scaling Groups and other services to show how easy it is to architect and deploy a real application on Amazon ECS with fewer moving parts. Well also visualize our application using a real-time map of the containers.
CON302 - Development Workflow with Docker and Amazon ECS
Keeping consistent environments across your development, test, and production systems can be a complex task. Docker containers offer a way to develop and test your application in the same environment in which it runs in production. You can use tools such as the ECS CLI and Docker Compose for local testing of applications; Jenkins and AWS CodePipeline for building and workflow orchestration; Amazon EC2 Container Registry to store your container images; and Amazon EC2 Container Service to manage and scale containers. In this session, you will learn how to build containers into your development workflow and orchestrate container deployments using Amazon ECS. You will hear how Okta runs 30,000 tests per developer commit and releases 10,000 new lines of code each week to production with a CI system based on 100% AWS services. We'll also discuss how Okta uses ECS for parallelized testing in CI and for production microservices in a multi-region, always on cloud service.
CON303 - Introduction to Container Management on AWS
Managing and scaling hundreds of containers is a challenging task. A container management solution takes care of these challenges for you, allowing you to focus on developing your application. In this session, we cover the role and tasks of a container management solution and we analyze how four common container management solutions - Amazon EC2 Container Service, Docker for AWS, Kubernetes, and Apache Mesos - stack against each other. We also see how you can easily get started with each of these solutions on AWS.
CON307 - NEW LAUNCH! Advanced Task Scheduling with Amazon ECS and Blox
As customers build and run production microservices architectures based on containers, having powerful tools to manage the placement and scheduling of these workloads is critical. Amazon ECS allows customers to focus on building their application and removes the need for managing the cluster management software entirely. This session will focus on the capabilities of the new ECS task placement engine, the built-in ECS schedulers, and explore how customers can build custom schedulers using the ECS event stream and Blox, a newly announced open source scheduling framework that gives developers additional choice and control in how they manage containers in production.
CON308 - Service Integration Delivery and Automation Using Amazon ECS
Through a combination of Amazon ECS and open source technologies, customers are able to build portable CI/CD pipelines on AWS. As container based deployments become more complex, they require additional rigging for integration. In this session, we show how popular Apache products like Kakfa, Storm, and Zookeeper are being deployed on top of Amazon ECS. We hear from HERE, a provider of mapping data, technologies, and services to the automotive, consumer, and enterprise sectors about an approach that leverages Consul from Hashicorp and Amazon ECS clusters for short-cycle deployments and tag-based environment promotion.
CON309 - Running Microservices on Amazon ECS
Running and managing large scale applications with microservices architectures is hard and often requires operating complex container management infrastructure. Amazon EC2 Container Service (ECS) is a highly scalable, high performance container management service that supports Docker containers and allows you to easily run applications on a managed cluster of Amazon EC2 instances. In this session, we will walk through a number of patterns used by our customers to run their microservices platforms. We will dive deep into some of the challenges of running microservices, such as load balancing, service discovery, and secrets management, and well see how Amazon ECS can help address them. We'll also hear from Instacart how they use a blue/green deployment process to deploy services to ECS and how they manage configuration with a RDS-based metadata service.
CON310 - Running Batch Jobs on Amazon ECS
Batch computing is a common way for developers, scientists and engineers to run a series of jobs on a large pool of shared compute resources, such as servers, virtual machines, and containers. Amazon ECS makes it easy to run and manage Docker-enabled applications across a cluster of Amazon EC2 instances. In this session will show you how to run batch jobs using Amazon ECS and together with other AWS services, such as AWS Lambda and Amazon SQS. We will see how you can leverage Amazon EC2 Spot Instances to power your ECS cluster and easily scale your batch workloads. You'll hear from Mapbox on how they use ECS to power their entire batch processing architecture to collect and process over 100 million milesof sensor data per daythat they use for powering their maps. Mapbox will also discuss how they optimize their batch processing framework on ECS using Spot Instances and demo their open source framework that will help you get up and running with ECS in minutes.
CON311 - Operations Automation and Infrastructure Management with Amazon ECS
At Capital One, we are using Docker and container technologies to advance microservices adoption, increase efficiencies of cloud resources, and decouple the application layer from the underlying infrastructure. Capital One is a federated organization with a you build it, you own it culture that provides autonomy and speed for delivery teams. Each federated team runs and operates their container management stack. In order for the federated teams to accelerate their cloud and container-based apps adoption, we created self-service automation tools for creation and operations management of container management stack. In this session, we explore our push-button automation tool that includes capabilities such as the creation and management of Amazon ECS clusters, an Application Load Balancer for dynamic and context-based routing and provides a user interface via a Jenkins Job or a AWS Lambda function. Our tooling also includes a home-grown dynamic service discovery and routing for applications requiring two-way mutual SSL authentication. We talk through how Capital One regularly updates AMIs with the latest patches and software versions using an automated solution that leverages AWS Lambda to rehydrate the Amazon ECS compute cluster with the latest AMI without causing any downtime. We also discuss how we created a sophisticated canary deployment automation using AWS Lambda and application services, where users can specify how to migrate to a new version of containerized apps and manage the deployment. AWS empowers enterprise Docker deployment with Amazon ECS and an ecosystem of cloud services and serverless architectures, making containerization in mission-critical environments easier than ever.
CON312 - Deploying Scalable SAP Hybris Clusters using Docker
Rent-A-Centers challenge was to architect, deploy, and manage a mission-critical SAP Hybris ecommerce platform that could scale to 2 million users a month. Together with Flux7, an AWS Advanced Consulting Partner, Rent-A-Center created an AWS-based approach that would help deliver the solution to market faster, in a secure, highly available, PCI-compliant fashion. In this session, we walk through the implementation details of this solution and its challenges, and explore how Rent-A-Center is now able to achieve ROI through agility, scalability, security, and cost savings.
CON313 - Netflix: Container Scheduling, Execution, and Integration with AWS
Customers from over all over the world streamed forty-two billion hours of Netflix content last year. Various Netflix batch jobs and an increasing number of service applications use containers for their processing. In this session, Netflix presents a deep dive on the motivations and the technology powering container deployment on top of Amazon Web Services. The session covers our approach to resource management and scheduling with the open source Fenzo library, along with details of how we integrate Docker and Netflix container scheduling running on AWS. We cover the approach we have taken to deliver AWS platform features to containers such as IAM roles, VPCs, security groups, metadata proxies, and user data. We want to take advantage of native AWS container resource management using Amazon ECS to reduce operational responsibilities. We are delivering these integrations in collaboration with the Amazon ECS engineering team. The session also shares some of the results so far, and lessons learned throughout our implementation and operations.
CON314 - Workshop: Deploy a Deep Learning Framework on Amazon ECS
Deep learning is an implementation of machine learning that uses neural networks to solve difficult and complex problems, such as computer vision, natural language processing, and recommendations. Due to the availability of deep learning libraries and frameworks, developers have the ability to enhance the capabilities of their applications and projects. In this workshop, you learn how to build and deploy a powerful deep learning framework called MXNet on containers. Theportability and resource management benefit of containersmeans developers can focus less on infrastructure and more on building. The labs start bydemonstrating the automation capabilities of AWS CloudFormation tostand upcore infrastructure; as an added bonus, you use Spot Fleet to leverage the cost benefits of using Spot Instances, especially for developer environments. Then, you walk throughcreatinganMXNet container in Docker anddeploying itwith Amazon ECS. Finally, you walk through an image classificationdemo of MXNet to validate that everything is working as expected. All you need to participate is a laptop and AWS account.
CON315 - Workshop: Deploy a Swift Web Application on Amazon ECS
In this workshop, you learn how to develop a mobile front-end using Swift, and develop a Swift microservices-based web application to be deployed on Amazon ECS. Workshop Requirements: Mac Laptop XCode 7 or 8 AWS Account IAM admin privileges for AWS account with full access to EC2, ECS, ECR, Code Pipeline, Mobile Hub and Device Farm. (If you choose to create a new account, please do it 24 hrs ahead.)
CON316 - State of the Union: Containers
Join us to learn about the latest developments from Amazon ECS and the container ecosystem. Deepak Singh, General Manager of AWS Container Services, discusses the evolution of containers on AWS and shares our vision for continued innovation in this space. You also hear about how other companies are using the AWS container platform to innovate and build new businesses.
CON401 - Amazon ECR Deep Dive on Image Optimization
Are you struggling with bulky images or slow push and pull times? In this session we will walk through the anatomy of a Docker image and provide techniques you can use to optimize images for faster pushes and pulls and reduce your overall storage footprint. We will discuss Docker image building (build containers versus runtime containers to remove unnecessary software), Docker image composition (minimizing the number of layers), the Docker Remote API (optimizing how images are pushed and pulled), and CI/CD Integration (automate building, versioning, and deploying images to production). Well also examine the tools that ECR provides to make Docker image management easier so that you can focus on building your application. Finally, we'll hear from Pinterest about how they use ECR and Docker, some valuable experiences gained along the way, and best practices for using ECR with Apache Mesos.
CON402 - Securing Container-Based Applications
Containers have had an incredibly large adoption rate since Docker was launched, especially from the developer community, as it provides an easy way to package, ship, and run applications. Securing your container-based application is now becoming a critical issue as applications move from development into production. In this session, you learn ways to implement storing secrets, distributing AWS privileges using IAM roles, protecting your container-based applications with vulnerability scans of container images, and incorporating automated checks into your continuous delivery workflow.
CTD201 - Architecting a 24x7 Live Linear Broadcast for 100% availability on AWS
Traditional broadcast infrastructures had to be built with on-premises architectures to provide five nines of reliability and to ensure the quality of service expected by premium content owners. However, robust live linear streaming services built on AWS are now feasible and in fact can be 100% available. Learn how broadcast media workflows with Elemental Cloud can provide ingest of heterogeneous video sources, fault tolerance across multiple Availability Zones, time synchronization of video streams, and sustained peak workloads in 24/7 applications.
CTD202 - Deliver and Monetize Your Content with Video Center Operations on AWS
The flexibility of AWS provides many advantages in the processing and delivery of bursty video applications such as video on demand and live events. However, full channel playout and advanced monetization have been the domains of traditional on-premises workflows. Learn how Elemental and AWS are architecting media-centric playout, monetization, and delivery services on top of AWS with live and file playlist integration, static and motion graphic overlays, ad content transcoding, and personalized, server-side ad insertion to bring additional video center operations to the cloud.
CTD203 - Journeys to the Cloud: Different Experiences in Video
In this session, youll hear different customers experiences bringing video workloads to AWS. Customer speakers from both traditional broadcasters and cloud-first technology companies will discuss what was same and what was different as they moved both video on demand asset creation and live feed production to AWS for the first time, and how their platforms have evolved with experience using Elastic Transcoder and Elemental Cloud.
CTD204 - Offload Security Heavy-lifting to the AWS Edge
Whether you are building a secure ecommerce application or developing games, security is a key consideration when architecting your application. In this session, you will learn about edge termination of your end user requests and will dive deep into advanced protocols and ciphers, enforcing end-to-end HTTPS connections with AWS Certificate Manager, access control with AWS WAF.
CTD205 - Introduction to Amazon CloudFront
End users expect to be able to view static, dynamic, and streaming content anytime, anywhere, and on any device. Amazon CloudFront is a web service that accelerates delivery of your websites, APIs, video content, or other web assets to end users around the globe with low latency, high data transfer speeds, and no commitments. In this session, learn what a content delivery network (CDN) such as Amazon CloudFront is and how it works, the benefits it provides, common challenges and needs, performance, recently released features like HTTP/2 and IPV6 support, pricing, and examples of how customers are using CloudFront.
CTD206 - NEW LAUNCH! Bringing AWS Lambda to the Edge
This session introduces Lambda@Edge, a new AWS Lambda feature that allows developers to perform simple computations at AWS edge locations in response to CloudFront events. This will be of interest to developers who want to build low-latency, customized web experiences. We cover product functionality and details of the programming model, and we walk through potential use cases.
CTD301 - Amazon CloudFront Flash Talks: Best Practices on Configuring, Securing and Monitoring your Distribution
In this series of 15-minute technical flash talks you will learn directly from Amazon CloudFront engineers and their best practices on debugging caching issues, measuring performance using Real User Monitoring (RUM), and stopping malicious viewers using CloudFront and AWS WAF.
CTD302 - Taking DevOps to the AWS Edge
In this session, we dive deep into how you can integrate Amazon CloudFront and related services into your application, be agile in developing and adapting the application, and follow best practices when configuring the services to improve security and performance, all while reducing costs. Attend this session and learn how to avoid needless forwarding of headers and cookies, test your application when making changes to the origin, version your configuration changes, monitor usage and automate security, create templates for new distributions, configure SSL/TLS certificates, and more.
CTD303 - Design Patterns for High Availability: Lessons from Amazon CloudFront that You Can Replicate on AWS
At AWS, the availability of our services is non-negotiable. While building our own services, such as Amazon CloudFront, we learn from and develop our own design patterns for high availability. In this session, we review several of these design patterns, and we show how you can implement the patterns in your own services or applications built on top of AWS using services such as Amazon Kinesis, AWS Elastic Beanstalk, or AWS Lambda.
CTD304 - How Mapbox Uses the AWS Edge to Deliver Fast Maps for Mobile, Cars, and Web Users Worldwide
Ian Ward, Platform and Security Engineer from Mapbox, discusses how the AWS global edge network helps improve the availability and performance of delivering hundreds of billions of map tiles to hundreds of millions of end users across the globe on mobile devices, in cars, and over the web. In this session, Ian shares insights on how Mapbox manages day-to-day edge operations using Amazon CloudFront logs, dashboards, and ad hoc queries, and how Mapbox has configured CloudFront with dozens of behaviors and origins to customize their content delivery. Mapbox has grown from using a single AWS region to using several regions, so Ian also explains how his team uses Amazon Route 53 and open source tools to simplify complexity around regional failover, and how Mapbox leverages AWS WAF to deter attacks and abuse.
CTD305 - Media Delivery from the Cloud: Integrated AWS Solutions for Premium Over the Top (OTT) Content
Learn from AWS and Sony DADC how the transformation in content aggregation, management, and delivery are optimized with Sony Ven.ue; a Media as a Service (MaaS) offering. In this session, you learn about the Ven.ue architecture, which uses end-to-end AWS services for content preparation, protection, management, and global content delivery. We dive deep on the integrations with AWS that make Amazon CloudFront a core component of the Ven.ue service. You also hear from a customer of Sony Ven.ue, Funimation, on how they benefit from this platform.
DAT201 - Cross-Region Replication with Amazon DynamoDB Streams
Under Armour implementedcross-region replication withAmazon DynamoDB Streams.Come listen as they share the keys to success.
DAT202 - Migrating Your Data Warehouse to Amazon Redshift
Amazon Redshift is a fast, simple, cost-effective data warehousing solution, and in this session, we look at the tools and techniques you can use to migrate your existing data warehouse to Amazon Redshift. We will then present a case study on Scholastics migration to Amazon Redshift. Scholastic, a large 100-year-old publishing company, was running their business with older, on-premise, data warehousing and analytics solutions, which could not keep up with business needs and were expensive. Scholastic also needed to include new capabilities like streaming data and real time analytics. Scholastic migrated to Amazon Redshift, and achieved agility and faster time to insight while dramatically reducing costs. In this session, Scholastic will discuss how they achieved this, including options considered, technical architecture implemented, results, and lessons learned.
DAT203 - Getting Started with Amazon Aurora
Amazon Aurora is a MySQL-compatible relational database engine with the speed, reliability, and availability of high-end commercial databases at one-tenth the cost. This session introduces you to Amazon Aurora, explores the capabilities and features of Aurora, explains common use cases, and helps you get started with Aurora. Debanjan Saha, general manager for Aurora, explains how Aurora differs from other commonly available databases while staying compatible with MySQL and providing a high-end, cost-effective alternative to commercial and open-source database engines. In addition, Linda Xu, data architect at Ticketmaster, walks you through Ticketmaster's journey to Amazon Aurora, starting with evaluation through production migration of a critical Ticketmaster database to Amazon Aurora. Ticketmaster is one of the world's top 10 e-commerce companies and the global market leader in ticketing. In this session, Linda discusses how Aurora lets Ticketmaster provide better services to their fans, customers, and clients, and helps reduce the cost and operational burden while giving greater flexibility to support heavy traffic spikes.
DAT204 - How Thermo Fisher Is Reducing Mass Spectrometry Experiment Times from Days to Minutes with MongoDB AWS
Mass spectrometry is the gold standard for determining chemical compositions, with spectrometers often measuring the mass of a compound down to a single electron. This level of granularity produces an enormous amount of hierarchical data that doesn't fit well into rows and columns. In this talk, learn how Thermo Fisher is using MongoDB Atlas on AWS to allow their users to get near real-time insights from mass spectrometry experimentsa process that used to take days. We also share how the underlying database service used by Thermo Fisher was built on AWS.
DAT205 - Relational and NoSQL Databases on AWS: NBC, MarkLogic, and FileMaker Perspectives on Data Management for Enterprise Consumer Apps
Learn how the AWS Marketplace brings together customers who have challenges with ISVs who have solutions to those challenges. See how to use relational and NoSQL technologies on AWS to build enterprise and consumer apps. NBC used MarkLogic to deliver an award-winning app that can handle high traffic levels and unexpected usage spikes. NBCs popular, Emmy-winning, SNL 40was launched to celebrate the 40th anniversary of Saturday Night Live, and delivers four decades of sketches and performances. Hosted on AWS, the app as well as a browser-based platform are powered by the MarkLogic Enterprise NoSQL database. Come learn from the team who collaborated on this project how to run your own database on AWS, and how to integrate with Amazon RDS and other data stores.A world-recognized automotive brand needed to deliver real-time response about their worldwide fleet vehicles. You will learn how they used a combination of AWS services and FileMaker Cloud, (an Apple subsidiary, procured through AWS Marketplace) to deliver high-scale dealer-facing applications.
DAT206 - NEW LAUNCH! Introducing PostgreSQL compatibility for Amazon Aurora
After we launched Amazon Aurora, a cloud-native relational database with region-wide durability, high availability, fast failover, up to 15 read replicas, and up to five times the performance of MySQL, many of you asked us whether we could deliver the same features - but with PostgreSQL compatibility. We are now delivering a preview of Amazon Aurora with this functionality: we have built a PostgreSQL-compatible edition of Amazon Aurora, sharing the core Amazon Aurora innovations with the object-oriented capabilities, language interfaces, JSON compatibility, ANSI:SQL:2008 compliance, and broad functional richness of PostgreSQL. Amazon Aurora will provide full PostgreSQL compatibilitywhile delivering more than twice the performance of the community PostgreSQL database on many workloads. At this session, we will be discussing the newest addition to Amazon Aurora in detail.
DAT207 - How Citus Enables Scalable PostgreSQL on AWS
Join the principal engineer of Citus Cloud for a brief overview of Citus, best use cases for it, and a drill down into how it's run and managed as a hosted service on top of AWS. The orchestration of Citus is homegrown, but comes from years of experience of running millions of PostgreSQL databases on top of AWS. Even if you aren't looking to leverage Citus to help you scale out, in this session you'll gain insights applicable to running and managing your stateful services on top of AWS. Citus is a PostgreSQL extension that transforms the database into a distributed, horizontally scalable database. Companies like Cloudflare use Citus to process 40 TB per day. With Citus MX, applications can take advantage of every node in the cluster for writes and yielding near-linear write scaling. Citus MX provide up to 500,000 durable writes per second.
DAT208 - Capturing Windows of Opportunity: Real-Time Analytics for Less Than $1000?
For many organizations, capturing a window of opportunity is their differentiator in the market. They invest in real-time systems to understand and respond to events. With so many use cases, it can be challenging for people to know how to invest in the right platform, and how to build in a cost-optimized way. In this session, we look at how some AWS customers are using real-time analytics to capture windows of opportunity: a telco with a major promotion, an advertising retargeter with global demands, and a personal IOT provider with a lifestyle solution. We dig deeper into their architecture and look for common patterns that can be used to build a real-time analytics platform in a cost-optimized way. We even see how a light-load, real-time analytics system can be built for less than $1000.
DAT301 - Amazon Aurora Best Practices: Getting the Best Out of Your Databases
Amazon Aurora is a fully managed relational database engine that provideshigher performance, availability and durability than previously possible using conventional monolithic database architectures.After launching a year ago, we continued adding many new features and capabilities to Aurora. In this session AWS Aurora experts will discuss the best practices that will help you putthese capabilities to thebest use. You will also hear from Amazon Aurora customer Intercom on the best practices they adopted for moving live databases with over two billion rows to a new datastore in Amazon Aurora with almost no downtime or lost records. Intercom was founded to provide a fundamentally new way for Internet businesses to communicate with customers at scale.For growing startups like Intercom, its natural for the load on datastores to grow on a weekly basis. The usual solutionto this problem is toget a bigger box from AWS. But very soon you reach a point wherebigger boat is not an option anymore. You will learn about the benefits of moving to such a datastore, the problems it introduced, and all about the new ability for scaling that was not there before.
DAT302 - Best Practices for Migrating from Commercial Database Engines to Amazon Aurora or PostgreSQL
You can significantly reduce database licensing and operational costs by migrating from commercial database engines to Amazon Aurora or Amazon RDS for PostgreSQL. In addition to cost reduction you also gain flexibility and operational efficiency by avoiding the frustrating usage constraints that the commercial databases licenses come with. Amazon Aurora and Amazon RDS for PostgreSQL are fully managed database services so you no longer need to worry about complex database management tasks. You can launch a single database instance or thousands of them in just a few minutes, and pay only for what you use. In this session we will dive deep into how AWS Database Migration Service and AWS Schema Conversion Tool help you migrate your commercial databases like Oracle and Microsoft SQL Server to Amazon Aurora or Amazon RDS for PostgreSQL easily and securely with minimal downtime. In this session you will also learn from Aaron Carerras, Senior Data Architect at FINRA, the approach FINRA has used to migrate their databases to Amazon RDS for PostgreSQL. As a financial regulator, FINRA has strict uptime SLAs and data security requirements and will be using RDS to meet them. The Financial Industry Regulatory Authority (FINRA) is one of the largest independent securities regulators in the United States, tracking up to 75 billion stock market events per day.
DAT303 - Deep Dive on Amazon Aurora
Amazon Aurora is a fully managed relational database engine that combines the speed and availability of high-end commercial databases with the simplicity and cost-effectiveness of open source databases. It is purpose-built for the cloud using a new architectural model and distributed systems techniques to provide far higher performance, availability and durability than previously possible using conventional monolithic database architectures. Amazon Aurora packs a lot of innovations in the engine and storage layers. In this session, we will do a deep-dive into some of the key innovations behind Amazon Aurora, new improvements to Aurora's performance, availability and cost-effectiveness and discuss best practices and optimal configurations.
DAT304 - Deep Dive on Amazon DynamoDB
Explore Amazon DynamoDB capabilities and benefits in detail and learn how to get the most out of your DynamoDB database. We go over best practices for schema design with DynamoDB across multiple use cases, including gaming, AdTech, IoT, and others. We explore designing efficient indexes, scanning, and querying, and go into detail on a number of recently released features, including JSON document support, DynamoDB Streams, and more. We also provide lessons learned from operating DynamoDB at scale, including provisioning DynamoDB for IoT.
DAT305 - Deep Dive on Amazon Relational Database Service
Amazon RDS allows customers to launch an optimally configured, secure and highly available database with just a few clicks. It provides cost-efficient and resizable capacity while managing time-consuming database administration tasks, freeing you up to focus on your applications and business. Amazon RDS provides you six database engines to choose from, including Amazon Aurora, Oracle, Microsoft SQL Server, PostgreSQL, MySQL and MariaDB. In this session, we take a closer look at the capabilities of RDS and all the different options available. We do a deep dive into how RDS works andthe best practises to achive the optimal perfomance, flexibility, and cost saving for your databases.
DAT306 - ElastiCache Deep Dive: Best Practices and Usage Patterns
In this session, we provide a peek behind the scenes to learn about Amazon ElastiCache's design and architecture. See common design patterns with our Redis and Memcached offerings and how customers have used them for in-memory operations to reduce latency and improve application throughput. During this session, we review ElastiCache best practices, design patterns, and anti-patterns.
DAT307 - Introduction to Managed Database Services on AWS
Which database is best suited for your use case? Should you choose a relational database or NoSQL or a data warehouse for your workload? Would a managed service like Amazon RDS, Amazon DynamoDB, or Amazon Redshift work better for you, or would it be better to run your own database on Amazon EC2? FanDuel has been running its fantasy sports service on Amazon Web Services (AWS) since 2012. You will learn best practices and insights from FanDuels successful migrations from self-managed databases on EC2 to fully-managed database services.
DAT308 - Fireside chat with Groupon, Intuit, and LifeLock on solving Big Data database challenges with Redis
Redis Labs' CMO is hosting a fireside chat with leaders from multiple industries including Groupon (e-commerce ), Intuit (Finance ), and LifeLock (Identity Protection ). This conversation-style session will cover the Big Data related challenges faced by these leading companies as they scale their applications, ensure high availability, serve the best user experience at lowest latencies, and optimize between cloud and on-premises operations. The introductory level session will appeal to both developer and DevOps functions. They will hear about diverse use cases such as recommendations engine, hybrid transactions and analytics operations, and time-series data analysis. The audience will learn how the Redis in-memory database platform addresses the above use cases with its multi-model capability and in a cost effective manner to meet the needs of the next generation applications.Session sponsored by Redis Labs.
DAT309 - How Fulfillment by Amazon (FBA) and Scopely Improved Results and Reduced Costs with a Serverless Architecture
Well share an overview of leveraging serverless architectures to support high performance data intensive applications. Fulfillment by Amazon (FBA) built the Seller Inventory Authority Platform (IAP) using Amazon DynamoDB Streams, AWS Lambda functions, Amazon Elasticsearch Service, and Amazon Redshift to improve results and reduce costs. Scopely will share how they used a flexible logging system built on Kinesis, Lambda, and Amazon Elasticsearch to provide high-fidelity reporting on hotkeys in Memcached and DynamoDB, and drastically reduce the incidence of hotkeys. Bothof these customers areusingmanaged services and serverless architecture to build scalable systems that can meet the projected business growth without a corresponding increase in operational costs.
DAT310 - Building Real-Time Campaign Analytics Using AWS Services
Quantcast provides its advertising clients the ability to run targeted ad campaigns reaching millions of online users. The real-time bidding for campaigns runs on thousands of machines across the world. When Quantcast wanted to collect and analyze campaign metrics in real-time, they turned to AWS to rapidly build a scalable, resilient, and extensible framework. Quantcast used Amazon Kinesis streams to stage data, Amazon EC2 instances to shuffle and aggregate the data, and Amazon DynamoDB and Amazon ElastiCache for building scalable time-series databases. With Elastic Load Balancing and Auto Scaling groups, they are able to set up distributed microservices with minimal operation overhead. This session discusses their use case, how they architected the application with AWS technologies integrated with their existing home-grown stack, and the lessons they learned.
DAT311 - How Toyota Racing Development Makes Racing Decisions in Real Time with AWS
Toyota Racing Development (TRD) developed a robust and highly performant real-time data analysis tool for professional racing. In this talk, learn how we structured a reliable, maintainable, decoupled architecture built around Amazon DynamoDB as both a streaming mechanism and a long-term persistent data store. In racing, milliseconds matter and even moments of downtime can cost a race. You'll see how we used DynamoDB together with Amazon Kinesis and Kinesis Firehose to build a real-time streaming data analysis tool for competitive racing.
DAT312 - How DataXu scaled its Attribution System to handle billions of events per day with Amazon DynamoDB
Attribution is the marketing term of art for allocating full or partial credit to individual advertisements that eventually lead to a purchase, sign up, download, or other desired consumer interaction. We'll share how we use DynamoDB at the core of our attribution system to store terabytes of advertising history data. The system is cost effective and dynamically scales from 0 to 300K requests per second on demand with predictable performance and low operational overhead.
DAT313 - 6 Million New Registrations in 30 Days: How the Chick-fil-A One App Scaled with AWS
Chris leads the team providing back-end services for the massively popular Chick-fil-A One mobile app that launched in June 2016. Chick-fil-A follows AWS best practices for web services and leverages numerous AWS services, including Elastic Beanstalk, DynamoDB, Lambda, and Amazon S3. This was the largest technology-dependent promotion in Chick-fil-A history. To ensure their architecture would perform at unknown and massive scale, Chris worked with AWS Support through an AWS Infrastructure Event Management (IEM) engagement and leaned on automated operations to enable load testing before launch.
DAT315 - Streaming ETL for RDS and DynamoDB
During this session Greg Brandt and Liyin Tang, Data Infrastructure engineers from Airbnb, will discuss the design and architecture of Airbnb's streaming ETL infrastructure, which exports data from RDS for MySQL and DynamoDB into Airbnb's data warehouse, using a system called SpinalTap. We will also discuss how we leverage Spark Streaming to compute derived data from tracking topics and/or database tables, and HBase to provide immediate data access and generate cleanly time-partitioned Hive tables.
DAT316 - How Telltale Games migrated its story analytics from Apache CouchDB to Amazon DynamoDB
Every choice made in Telltale Games titles influences how your character develops and how the world responds to you. With millions of users making thousands of choices in a single episode, Telltale Games tracks this data and leverages it to build more relevant stories in real time as the season is developed.In this session, youll learn about Telltale Games migration from Apache CouchDB to Amazon DynamoDB, the challenges of adjusting capacity to handling spikes in database activity, and how it streamlined its analytics storage to provide new perspectives of player interaction to improve its games.
DAT317 - Learn how IFTTT uses ElastiCache for Redis to predict events and index terabytes of logs
IFTTT is a free service that empowers people to do more with the services they love, from automating simple tasks to transforming how someone interacts with and controls their home. IFTTT uses ElastiCache for Redis to store transaction run history and schedule predictions as well as indexes for log documents on S3. Join this session to learn how the scripting power of Lua and the data types of Redis allowed them to accomplish something they would not have been able to elsewhere.
DAT318 - Migrating from RDBMS to NoSQL: How PlayStation™Network Moved from MySQL to Amazon DynamoDB
In this session, you will learn the key differences between a relational database management service (RDBMS) and non-relational (NoSQL) databases like Amazon DynamoDB. You will learn about suitable and unsuitable use cases for NoSQL databases. You'll learn strategies for migrating from an RDBMS to DynamoDB through a 5-phase, iterative approach. See how Sony migrated an on-premises MySQL database to the cloud with Amazon DynamoDB, and see the results of this migration.
DAT320 - AWS Database State of the Union
Raju Gulabani, vice president of AWS Database Services (AWS), discusses the evolution of database services on AWS and the new database services and features we launched this year, and shares our vision for continued innovation in this space. We are witnessing an unprecedented growth in the amount of data collected, in many different shapes and forms. Storage, management, and analysis of this data requires database services that scale and perform in ways not possible before. AWS offers a collection of such database and other data services like Amazon Aurora, Amazon DynamoDB, Amazon RDS, Amazon Redshift, Amazon ElastiCache, Amazon Kinesis, and Amazon EMR to process, store, manage, and analyze data. In this session, we provide an overview of AWS database services and discuss how our customers are using these services today.
DAT321 - Workshop: Using the Database Migration Service (DMS) for Database Consolidation, Data Distribution and Replication
It can help you do much more. You can use DMS to consolidate multiple databases into a single database or split a single database into multiple databases. You can also use DMS for data distribution to multiple systems. For both of these use cases your source database can be outside of AWS (on premises) or in AWS (EC2 or RDS). DMS can also be used for near real-time replication of data. Replication can be done to one or more targets within AWS, in the same region or across regions. You can also replicate data from databases within AWS to databases outside of AWS. In this session we will discuss all these usage patterns and help you try them out yourselves. Prerequisites: You should have good database knowledge and at least some experience with Amazon RDS or Amazon Aurora. Participants should have an AWS account established and available for use during the workshop. Please bring your own laptop.
DAT322 - Workshop: Stretching Scalability: Doing more with Amazon Aurora
Easy scalability is a powerful feature of Amazon Aurora. Scalability in its actual definition refers to being able to get larger or smaller depending on the need. Amazon Aurora allows you to easily achieve this by scaling the database instance up or down and adding or removing read replicas. Scaling across regions brings additional resilience to your architectures and could boost your application performance due to geographic proximity. You can perform all of these scaling operations through the Aurora console. You can also automate instance and read scaling using lambda function or scripts based on the usage pattern you define. You can extend the automation by feeding your database usage data from Aurora enhanced monitoring into Machine Learning to provide more sophisticated predictive patterns to drive your automation. In this session we will do a deep dive into how scalability works in Aurora and how to make the best use of it to reduce your cost, increase application performance and architect resilient applications. You should have good database knowledge and at least some experience with Amazon RDS or Amazon Aurora and should bring your own laptop.
DAT323 - Workshop: Converting Your Oracle or Microsoft SQL Server Database to an Open Source Amazon Aurora or PostgreSQL Database Using AWS SCT and AWS DMS
In this workshop, you migrate a sample sporting event and ticketing database from Oracle or Microsoft SQL Server to Amazon Aurora or Postgre SQL using the AWS Schema Conversion Tool (AWS SCT) and AWS Database Migration Service (AWS DMS). The workshop includes the migration of tables, indexes, procedures, functions, constraints, views, and more. We run SCT on a Amazon EC2 Windows instance--bring a laptop with Remote Desktop (or some other method of connecting to the Windows instance). Ideally, you should be familiar with relational databases, especially Oracle or SQL Server and PostgreSQL or Aurora, to get the most from this session. Additionally, attendees should be familiar with SCT and DMS. Familiarity with SQL Developer and pgAdmin III will be helpful but is not required. Prerequisites: Participants should have an AWS account established and available for use during the workshop. Please bring your own laptop.
DCS201 - Open-Source Resources
This session discusses the open-source community, its vital place within the AWS ecosystem, and how AWS works to provide seamless integration points. Our speakers share their experiences building and deploying cloud-based open-source projects while also reviewing some of today's most popular and relevant open-source platforms and solutions.
DCS202 - AWS Training Opportunities
Currently, a breadth of AWS training opportunities are available worldwide, both led by AWS and through community-driven training platforms. In this session, community leaders sort through the different training resources, discuss the resources they used to help them become AWS experts, and explain how different training solutions can complement one another.
DCS203 - Building and Growing a Successful AWS User Group
Our panel of experts lead AWS user groups in San Francisco, Yokohama, Munich, and Shanghai. In this session they share the stories behind how their groups began, list best practices for sustaining a technical meetup over time, and offer advice for AWS enthusiasts who are considering starting up a new user group in their city.
DCS204 - Developer Lightning Talks and Happy Hour
Don't miss our fast and furious conclusion to the Developer Community Summit. Enjoy food and drinks while listening to 6 community leaders, each of whom shares the latest developer tools, trends, and events taking place in their region in five minutes or less.
DCS205 - Workshop: Building Serverless Bots on AWS - Botathon
In this session we will learn about building Serverless Bots using AWS Services. We will explore contemporary examples of Bots and Serverless architectural building blocks needed to build a Bot. Taking few sample Bots, we will dive deeper into the AWS Services used to build them. The focus will be on Serverless architectural components such as Lambda, API Gateway, Alexa Skills Kit, etc. We will explore different interfaces, voice (eg. Alexa) and text (eg. Slack). We will also discuss about building intelligent Bots. In the hands on session, the participants will build a Serverless Bot. Participants are free to choose any theme for their Bot, although guidance will be provided on few starter ideas. The participants will be grouped into teams and will have access to starter code, libraries so that they can focus on building their unique bot rather than the underlying undifferentiated heavy lifting. At the end of the hour, the Bots get voted on and the winning Bot is demoed. Through this workshop, the participants will get a deeper understanding of Serverless AWS Services and how to use them to build a Serverless Bot. Prerequisites: Participants should have an AWS account established and available for use during the workshop. Please bring your own laptop.
DEV201 - DevOps on AWS: Accelerating Software Delivery with the AWS Developer Tools
Todays cutting edge companies have software release cycles measured in days instead of months. This agility is enabled by the DevOps practice of continuous delivery, which automates building, testing, and deploying all code changes. This automation helps you catch bugs sooner and accelerates developer productivity. In this session, well share the processes followed by Amazon engineers and discuss how you can bring them to your company by using AWS CodePipeline and AWS CodeDeploy, services inspired by Amazon's internal developer tools and DevOps culture.
DEV202 - From Dial-Up to DevOps - AOL’s Migration to the Cloud
AOL originally provided dial-up service to millions of people. Today, AOL powers advertising and media experiences for the webs top destinations. How do you maintain observability and reliability to both business and technical teams for high-traffic services in a dynamic infrastructure? Join us as we discuss AOLs DevOps journey. We will dive into its engineering culture, automation, and monitoring best practices that have allowed AOL to successfully reinvent their infrastructure, as they moved from globally distributed data centers to the AWS Cloud. Session sponsored by Datadog. AWS Competency Partner
DEV204 - How A Federal Agency Transformed Work and Adopted DevOps with GitHub
In this session, youll hear from GitHub and Accenture Federal Services, a trusted advisor to the US government, on why they have continued to invest in the adoption of and transition to cloud services. After migrating to AWS cloud, one agency deployed GitHub, the cloud-hosted, distributed version control and collaboration platform, as the backbone of its DevOps program. Now, thousands of users on software development teams at the agency collaborate both internally and with other agencies faster and more efficiently than ever before. Learn how they decreased duplicative work, raised the quality of their code, and greatly increased delivery velocity. Our Accenture Federal Services speaker will share details on what its like to run GitHub Enterprise on AWS for a federal agency, including the unique challenges and solutions that stem from running an appliance in the cloud, and advice for others considering this path. Session sponsored by GitHub. AWS Competency Partner
DEV205 - Monitoring, Hold the Infrastructure: Getting the Most from AWS Lambda
Just as we got a hang of monitoring our server-based applications, they take away the server. How do you monitor something that doesnt exist? Which metrics matter most in a serverless world? In this session, we will look at how applications are different in an AWS Lambda-based world and how to monitor them. Join us as we work our way through the stack and demonstrate how to capture the health and performance of your services. The focus of this session is not tool-specific. Attendees will learn production-tested lessons and leave with frameworks they can implement with their serverless workloads, no matter which platforms and tools they use. This session sponsored by Datadog. AWS Competency Partner
DEV206 - Scaling Your Web Applications with AWS Elastic Beanstalk
AWS Elastic Beanstalk provides an easy way for you to quickly deploy, manage, and scale applications in the AWS Cloud. Through interactive demos and code samples, this session will teach you how to deploy your code using Elastic Beanstalk, provision and use other AWS services (Amazon SNS, Amazon SQS, and Amazon DynamoDB), use your applications health metrics to tune performance, scale your application to handle millions of requests, perform zero-downtime deployments with traffic routing, and keep the underlying application platform up-to-date with managed updates. Code samples for demos will be available to all session attendees.
DEV207 - NEW LAUNCH! Developing Serverless C# Applications
Just as serverless application development is rapidly becoming the most popular way to bring highly scalable applications to the cloud, .NET has undergone radical changes with .NET Core to become a premier development platform for the cloud. In this session, you will learn how to use the newly launched C# support for .NET Core with AWS Lambda to create highly scalable serverless applications that target platforms from the traditional desktop to mobile devices. We will demonstrate how to write, test, and deploy C# code to AWS Lambda and see how we can leverage our serverless back end from mobile applications.
DEV209 - Another Day in the Life of a Netflix Engineer
Netflix is big. Really big. You just won't believe how vastly, hugely, mind-bogglingly big it is. Netflix is a large, ever changing, ecosystem system serving million of customers across the globe through cloud-based systems and a globally distributed CDN. This entertaining romp through the tech stack serves as an introduction to how we think about and design systems, the Netflix approach to operational challenges, and how other organizations can apply our thought processes and technologies. Well talk about: The Bits - The technologies used to run a global streaming company Making the Bits Bigger - Scaling at scale Keeping an Eye Out - Billions of metrics Break all the Things - Chaos in production is key DevOps - How culture affects your velocity and uptime
DEV211 - Automated DevOps and Continuous Delivery
In the digital economy, the fast development and deployment of applications is critical to success. To thrive in this application-oriented business environment, IT organizations are acting now to change their tools and processes to better support agile development methodologies. This session will cover performance benchmarking, benefits of migrating existing workloads, use of key services like Amazon RDS and AWS CloudHSM, and demonstrate how to deploy applications securely and at scale. Session sponsored by Cisco.
DEV301 - Amazon CloudWatch Logs and AWS Lambda: A Match Made in Heaven
In this session, we cover three common scenarios that include Amazon CloudWatch Logs and AWS Lambda. First, you learn how to build an Elasticsearch cluster from historical data using Amazon S3, Lambda, and CloudWatch Logs. Next, you learn how to add details to CloudWatch alarm notifications using Amazon SNS and Lambda. Finally, we show you how to bring Elastic Load Balancing logs to CloudWatch Logs using S3 bucket triggers from Lambda.
DEV302 - Automated Governance of Your AWS Resources
AWS CloudTrail, Amazon CloudWatch Events, AWS Identity Access Management (IAM), Trusted Advisor, AWS Config Rules, other services? In this session, we will help you use existing and recently launched services to automate configuration governanceso that security is embedded in the development process. We outline four easy steps(Control, Monitor, Fix, and Audit) and demonstrate how different services can be used to meet your governance needs. We will showcase real-life examples and you can take home a blog post with code examples and the full source code for scripts and tooling that AWS professional services have built using these services.
DEV303 - Deploying and Managing .NET Pipelines and Microsoft Workloads
In this session, well look at the AWS services that customers are using to build and deploy Microsoft-based solutions that use technologies like Windows, .NET, SQL Server, and PowerShell. Well start by showing you how to build a Windows-based CI/CD pipeline on AWS using AWS CodeDeploy, AWS CodePipeline, AWS CloudFormation, and PowerShell using an AWS Quick Start. Well also cover best practices for how you can create templates that let you automatically deploy ready-to-use Windows products by leveragingservices and tools like AWS CloudFormation, PowerShell, and Git. Woot, an online retailer for electronics, will share how it moved from using acomplex mix of custom PowerShell code for its DevOps processes to using services like Amazon EC2 Simple Systems Manager (SSM), AWS CodeDeploy, and AWS Directory Service. This migration eliminated the need for complex PowerShell scripts and reduced the operational complexity of performing operational tasks like renaming servers, joining domains, and securely handling keys.
DEV304 - Building the Future of DevOps with Amazon Web Services
At Dynatrace, we challenged ourselves to build a virtual team member to help operations teams run large-scale cloud infrastructures. Think J.A.R.V.I.S. from Iron Man, but for operations. We built our cloud infrastructure on Amazon EC2, Elastic Load Balancing load balancers, and Auto Scaling groups for real-time scalability, Amazon Route 53 for instant customer access, Amazon Echo and Alexa for voice interaction, AWS Lambda for fast prototyping of the human-interaction layer, and Amazon DynamoDB for handling complex conversations. In this session, we will also discuss how we extend the service by using Amazon Machine Learning and AWS IoT to more naturally integrate our virtual assistant into the real world. Session sponsored by Dynatrace. This session sponsored by Dynatrace. AWS Competency Partner
DEV305 - Configuration Management in the Cloud
To ensure that your application operates in a predictable manner in both your test and production environments, you must vigilantly maintain the configuration of your resources. By leveraging configuration management solutions, Dev and Ops engineers can define the state of their resources across their entire lifecycle. In this session, we will show you how to use AWS OpsWorks, AWS CodeDeploy, and AWS CodePipeline to build a reliable and consistent development pipeline that assures your production workloads behave in a predictable manner.
DEV306 - How News UK Centralized Cloud Governance Through Policy Management
When you run a complex AWS environment with thousands of Amazon EC2 instances, more than half a petabyte of object storage, and support the largest daily newspapers in the UK, you need a world-class cloud management strategy. For companies like News Corp, implementing policies that automate infrastructure schedules, right-size workloads, and manage and modify reservations is critical. As you scale your cloud infrastructure, defining centralized governance rules while enabling decentralized management is key to running an optimized cloud. This session is designed for advanced operations, infrastructure, and engineering teams to improve/deploy optimization strategies. It covers the five best cloud management practices, including automating Reserved Instance modifications, setting policies to ensure proper tagging, and scheduling lights-on/lights-off policies.Session sponsored by CloudHealth Technologies.
DEV307 - How to Scale and Operate Elasticsearch on AWS
Elasticsearch has quickly become the leading open source technology for scaling search and building document services on. Many software providers have come to rely on it to serve the needs of high-performance, production applications. In this talk, well go deep on lessons learned from three years in production scaling from a few shards to more than 100 spread across 100s of nodes on AWS--to serve real-time queries against 100s of millions of documents. Attendees will learn: * How to capacity plan for ES on AWS * How to scale and reshard on AWS with zero downtime * What AWS and ES metrics to collect and alert on * Tips on day to day ES operations Session sponsored by SignalFx.
DEV308 - Chalice: A Serverless Microframework for Python
Chalice is a new Python serverless microframework for AWS. In this session, we'll show how to use it to create serverless REST APIs that use Amazon API Gateway and AWS Lambda. We'll deploy our first REST API in less than 30 seconds. Several live demos will show how to build on this example to leverage other Chalice capabilities, including advanced URL routing, automatic policy generation, and Swagger exports.
DEV309 - Turbocharge Your Microsoft .NET Developments with AWS
In this session, you will discover how to integrate the AWS developer tools into your development process. We will demonstrate how to leverage AWS services, the .NET SDK, and the Visual Studio Toolkit to simplify and streamline your development processes. This session is targeted at development teams using Microsoft Visual Studio and the Microsoft ecosystem of products. Most of the presentation will be in Visual Studio.
DEV310 - DevOps on AWS: Choosing the Right Software Deployment Technique
Weve seen companies like fast-growing startups, large enterprises, and even Amazon.com itself adopt and evolve strategies to optimize their delivery of cloud applications. Some AWS customers perform in-place updates across their servers. Some perform blue-green deployments to newly provisioned servers. In this session, well share the advantages of each approach and talk about the scenarios in which you should choose one over the other. We will also demonstrate how to perform each deployment technique in an agile and reliable manner.
DEV311 - Multi-Region Delivery Netflix Style
Netflix rapidly deploys services across multiple AWS accounts and regions over 4,000 times a day. Weve learned many lessons about reliability and efficiency. Whats more, weve built sophisticated tooling to facilitate our growing global footprint. In this session, youll learn about how Netflix confidently delivers services on a global scale and how, using best practices combined with freely available open source software, you can do the same.
DEV312 - Cloud Monitoring: Change is the New Normal- New Relic Gannett Discuss their Measurement Journey
Dynamic applications require dynamic resources and dynamic infrastructure. AWS provides many resources for applications to build highly scaled, highly available, dynamic applications, services, and microservices. However, managing and tracking these resourcesand making sure they are operating as expectedis a challenge. In this session, we discuss how to monitor and manage the dynamic resources that make up your applications, and learn how to tell when a resource is causing your application problems. Designed for people already acquainted with basic dynamic resource allocation techniques, such as effectively using Auto Scaling, this session helps you take your resource management to the next level. Session sponsored by New Relic. AWS Competency Partner
DEV313 - Infrastructure Continuous Delivery Using AWS CloudFormation
In this session, we will review ways to manage the lifecycle of your dev, test, and production infrastructure using CloudFormation. Learn how to architect your infrastructure through loosely coupled stacks using cross-stack references, tightly coupled nested stacks and other best practices. Learn how to use CloudFormation to provision and manage a continuous deployment pipeline for your infrastructure-as-code. Automate deployment of new development environments as your infrastructure evolves, promote your new architecture for testing, and deploy changes to production.
DEV315 - How Bleacher Report Gains Competitive Edge with AWS Elastic Beanstalk and Docker
When you're one of the top sports media sites on the internet, you deal with scale like no other. Bleacher Report and its Team Stream app enable millions of users to access their own personalized view of sports. In this session, we'll talk about how we broke up the monolith into microservices and how Elastic Beanstalk empowered us to move quickly. Learn how a small Ops team provided a world-class build/release pipeline by standing on the shoulder of giants (AWS Elastic Beanstalk, Jenkins, and Docker). This session is designed for those who want to get up and running as quickly as possible, but are uncompromising in their ownership of infrastructure. We'll discuss the reasoning behind our switch from fully managing our own infrastructure to a managed service, including some advanced customizations made possible through AWS CloudFormation and AWS Elastic Beanstalk configuration files (.ebextensions).
DEV316 - NEW LAUNCH! Introduction to AWS X-Ray
Analyzing and debugging production distributed applications built using a service-oriented or microservices architecture is a challenging task. In this session, we will introduce AWS X-Ray, a new service that makes it easier to identify performance bottlenecks and errors, pinpoint issues to specific service(s) in your application, identify the impact of issues on users of your application, and visualize a request call graph and service call graph for your applications. We will show interactive demos, and code samples for the demo will be available to all session attendees.
DEV317 - Automating and Scaling Infrastructure Administration with AWS Management Tools
In this session, well show how customers can use management tools to standardize the creation of AWS resources and then govern these resources through the lifecycle. By using AWS CloudFormation and AWS Service Catalog to provision resources at scale, AWS Config to audit any changes to the configuration of these resources, Amazon CloudWatch to monitor the health of these resources, and AWS CloudTrail to audit who or what made API calls to these resources, customers can automate and scale the administration of their infrastructure on AWS. They can even go one step further and automate compliance checking and remediation by using AWS Config rules and Amazon CloudWatch Events. We will demo how this is possible by looking at some common use cases.
DEV319 - Automating Cloud Management and Deployment for a Diverse Enterprise Application Portfolio
Building scalable automation tools that work across heterogeneous application environments is challenging and can inhibit enterprise cloud migration efforts. Learn how the DevOps team at Infor, one of the worlds leading ISVs, manages dozens of enterprise applications built with a variety of technologies and application architectures. In addition to Infors approach to deployment and management automation, this session will cover the core tools theyve developed on top of native AWS services such as AWS CloudFormation, AWS CodeDeploy, and AWS Lambda.
DEV320 - What hundreds of companies show us about how to build a DevOps Culture
Adam Jacob, the CTO of Chef, will walk through what a DevOps Culture really means, explaining the most important concepts through examples and stories from the trenches. We'll talk about how to build empowered organizations, design delightful products, run effective operations, and more.
DEV321 - Enabling DevOps for an Enterprise with AWS Service Catalog: The John Wiley Sons Journey with AWS Professional Services
As incumbent enterprises move to the cloud, questions arise how transform the legacy IT culture to maximize the agility and flexibility AWS provides. Speed and dexterity must be implemented in a consistent manner, minimizing the impact to the organizational structure, but taking into account the existing skill sets and knowledge base. With AWS Service Catalog, you can manage commonly deployed AWS CloudFormation template versions, enable controlled self-provisioning, and leverage those same products in your automated deployment pipelines to AWS. In this session, developers, operations leads, architects, and IT managers learn how to leverage AWS Service Catalog and AWS CloudFormation to transform IT culture to maximize the agility, flexibility, and value that the AWS platform provides. Additionally, John Wiley Sons, a 200-year-old enterprise, demonstrates how AWS Professional Services helped them balance the velocity achieved by moving to AWS with a structured governance model to deploy their cloud infrastructure and application code.
DEV401 - Automating Workflows for Analytics Pipelines
Learn how to leverage new workflow management tools to simplify complex data pipelines and ETL jobs spanning multiple systems. In this technical deep dive from Treasure Data, company founder and chief architect walks through the codebase of DigDag, our recently open-sourced workflow management project. He shows how workflows can break large, error-prone SQL statements into smaller blocks that are easier to maintain and reuse. He also demonstrates how a system using last good checkpoints can save hours of computation when restarting failed jobs and how to use standard version control systems like Github to automate data lifecycle management across Amazon S3, Amazon EMR, Amazon Redshift, and Amazon Aurora. Finally, you see a few examples where SQL-as-pipeline-code gives data scientists both the right level of ownership over production processes and a comfortable abstraction from the underlying execution engines. This session is sponsored by Treasure Data. AWS Competency Partner
DEV402 - The Effective AWS CLI User
Understanding the internals of the AWS CLI will make you a more effective user. This talk provides a deep dive into the architecture, debugging techniques, advanced usage patterns, and some of its more advanced features. We recommend this talk for people who are already familiar with the AWS CLI because the topics will be advanced. By the end of the talk, audience members will have a deeper understanding of the AWS CLI, allowing them to tackle current or future applications from a different perspective.
DEV403 - DevOps on AWS: Advanced Continuous Delivery Techniques
Continuous delivery makes teams more agile and quickens the pace of innovation. Too often, though, teams adopt continuous delivery without putting the right safety mechanisms in place. In this talk, we'll transform a simple but typical software release process into one that is safe. We'll use DevOps techniques like continuous integration, a variety of non-production testing stages, rollbacks, machine redundancy, Availability Zone redundancy, canary deployments, canary tests, and dashboards. We'll use AWS Lambda, AWS CloudFormation, AWS CodePipeline, AWS CodeDeploy, Amazon CloudWatch alarms and dashboards, and AWS Elastic Beanstalk.
DEV404 - Develop, Build, Deploy, and Manage Containerized Services and Applications on the AWS Cloud
In this session, we walk through the simple process of deploying and managing your own Linux-based application in the cloud and also discuss key use-cases and benefits to automated configuration, deployment, and administration of application stacks. Session sponsored by Red Hat.
ENT201 - Deploying Amazon WorkSpaces at Enterprise Scale to Deliver a New Desktop Experience
Your enterprise has decided it is exiting the traditional desktop business and migrating to Amazon WorkSpaces. Your challenge: how do you provide end users a high quality experience using DaaS while integrating key enterprise services? Focusing on user adoption and simplified operational management DaaS offers significant benefits over traditional physical desktops and VDI solutions. These benefits include hourly consumption pricing, sizing flexibility, linear scalability, and simplified management. We will dive into the methodology and design decisions that Informa, a global leader in business intelligence and part of the Information Economy, made to migrate to Amazon WorkSpaces. The goal of this strategy for Informa was to provide their end users with a Windows 10 desktop that could scale across three AWS regions. Following this session you will be equipped with a methodology for migrating to Amazon WorkSpaces and key design decision points that will assist your organization in rolling out this service for your end users. Best practices will be presented and there will be discussion on the capabilities required to build an enterprise-scale solution around WorkSpaces. All of this will include the fundamental value proposition AWS has brought to market with their industry-leading DaaS offering. Session sponsored by AHEAD.
ENT202 - Driving AWS Cost Efficiency at Your Company
In this session, youll learn how large, scaled organizations control and optimize their AWS costs. Youll understand how you can grow usage in a cost-effective way across all business units or product teams. Attendees will walk away with new strategies and tools for driving cost-efficient decision-making at every level of the company. Topics include: - How to align your teams around key cost and business measures - How to best collaborate between Finance and Ops around cost management - How to allocate costs and drive efficient use of resources - How cost management practices change in a serverless world of Lamba functions, Docker containers, and other services. Session sponsored by Cloudability.
ENT203 - Enterprise Fundamentals: Design Your Account and VPC Architecture for Enterprise Operating Models
Customers looking to migrate to AWS often ask the following questions at the beginning of their journey What is the right AWS account structure and VPC design for me? and How do I minimize the impact to my IT operations? This session discusses the different account structures, VPC design patterns and network deployment architectures to align to different Enterprise IT Operating Models. We will also discuss the implications each pattern has with respect to Security, Finance and Operations. Nielsen, a global enterprise providing an understanding of what consumers watch and what consumers buy, will discuss the decision making process for their AWS account and VPC design. Nielsen will also discuss the implications of their decision. This session will benefit IT architects, managers and technology strategists to: Understand the 4 common Enterprise IT Operating Models and the VPC design patterns associated with each The security, operational and financial considerations associated with each design pattern Ask the correct questions to determine how best to set up their AWS accounts
ENT204 - Large-scale AWS Migrations
As you begin to move out of your data center and develop a cloud-first strategy, you'll need support for large-scale migrations to AWS. In this session, CSC shares details about the journey to AWS by some of our largest enterprise customers. We provide best practices for planning your large-scale migrations and focus on business processes in addition to technology. We show how CSC used this approach to migrate to AWS as part of our separation last year into two publicly traded companies: CSC and CSRA. In less than six months, CSC took our 56-year-old company and broke it into two companies, one of which was brand new and without any infrastructure or enterprise applications. We explain how we leveraged the AWS Partner ecosystem to achieve this incredible IT challenge. Session sponsored by CSC. AWS Competency Partner
ENT206 - Lift and Evolve – Saving Money in the Cloud is Easy, Making Money Takes Help
Every enterprise knows by now that it can save money by simply lifting and shifting workloads to the cloud, but many are missing the larger opportunity to also make money by moving. While quick costs savings are good for the bottom line, they do little to move the top line numbers. To achieve both savings and earnings, corporate thinking about technologies must change in order to enable faster processes leveraged enterprise-wide. In this session we will explore multiple customer success stories where the customers have evolved from leveraging basic compute and storage products (EC2 and S3) to integrating new services into operations by leveraging Lambda, DynamoDB, CodeDeploy, etc. Once this is achieved, enterprises are enabled to manage and deploy code rapidly in a programmatic and elastic secure network, ensuring governance and security standards across the globe. We will look at the migration process trusted by hundreds of clients as well as how to cope with the process and people components that are so important to enable agility, while focusing heavily on the technology. The Coca-Cola Company (TCCC) and 2nd Watch story will dive deep into the technology that allows TCCC to manage hundreds of AWS Accounts, hundreds of workloads, thousands of instances, and hundreds of business partners around the globe. TCCCs Configuration Management System has Puppet at the core and relies on over a dozen core and emerging AWS products across accounts, availability zones and regions. This complex and globally-available system ensures all of TCCCs workloads in AWS meet corporate policies but also allows for rapid scale of both consumer and enterprise workloads. Session sponsored by 2nd Watch. AWS Competency Partner
ENT207 - Leading Enterprise Innovation at Startup Speed
Incumbent enterprises face dramatically competitive landscapes, with threats from almost every direction. Protecting your core business and innovating for the future is a delicate balancing act. Innovating as fast as a startup becomes a core competency, but failed new product innovation wastes time and resources. In this session, IT managers and professionals learn how running a lean enterprise can be a powerful framework for leading enterprise-scale innovation as effectively and fast as a startup.
ENT209 - Cost Optimization at Scale
Organizations where cloud adoption has matured into broader enterprise deployment are facing the need to better manage and control their costs and expenditures. Cost optimization at scale is a process that involves a number of changes across the business, including technical, organizational and cultural transformation. In this session, you will learn the fundamentals of cost optimization and how this can be used to help your organization drive costs down and still being able to meet capacity, demand and organizational requirements. Key topics being discussed are right sizing services, optimizing purchase models and implementing a culture of cost management.
ENT210 - Deep Dive on AWS Cloud Data Migration Services
When evaluating and planning migrating your data from on premises to the Cloud, you might encounter physical limitations. Amazon offers a suite of tools to help you surmount these limitations by moving data using networks, roads, and technology partners. In this session, we discuss how to move large amounts of data into and out of the Cloud in batches, increments, and streams.
ENT211 - Getting Started with the Hybrid Cloud: Enterprise Backup and Recovery
This sessions is for architects and storage admins seeking simple and non-disruptive ways to adopt cloud platforms in their organizations. You will learn how to deliver lower costs and greater scale with nearly seamless integration into your existing Backup ad Recovery processes to achieve fast, simple wins that demonstrate the scale and flexibility of cloud services for storage. Services mentioned: S3, Glacier, Snowball, 3rd party partners, Storage Gateway, and cloud data migration services.
ENT212 - Preparing for a Large-Scale Migration to AWS
Many organizations rely on a TCO calculation to build a business case for migration to the Cloud. Unfortunately, this approach overlooks organizational readiness and wider economic factors, such as the total cost and the benefits of migration. This session provides executives with a toolkit for measuring readiness and developing a compelling business case.
ENT213 - Best practices for running enterprise workloads on AWS
Fortune 500 companies are increasingly using cloud services to run enterprise workloads to improve security, increase agility, and enable scale. Learn how OpenEye is running their AWS-native platform and workflow engine to support collaboration and data sharing at large pharmaceutical companies like Pfizer. In this session, OpenEye will share cloud best practiced around security controls, cross-departmental collaboration across the enterprise, and agility at scale. Attendees will gain practical tips for using AWS in the enterprise and healthcare industries.
ENT214 - The Cloud Is Enterprise Ready: How the University of Arizona Is Migrating Core Enterprise Systems to the Cloud
As enterprises seek to fully realize the scalability, availability, security, and agility of cloud computing, many are learning that they must migrate core business applications, such finance and human resource systems, not just public-facing websites. The University of Arizona shares how they addressed the challenge of creating a more efficient and agile IT operation. After initial success in migrating their Research Administration system, UA became the first major university to migrate their PeopleSoft application environments—Learning (EL), Human Resources (HR), and Student (SA)—to AWS. Learn about the university's enterprise migration roadmap, lessons learned, plans and budgets, and the tools available to support major enterprise migrations. See how to create a roadmap to cloud computing, transform application development, deploy mission-critical workloads at scale, and retrain staff.
ENT215 - Building Enterprise Cloud Operations As a Service with T-Mobile
After launching several thousand EC2 instances in the cloud, we've learned that the key to running an IT fleet successfully on AWS is enforcing operational and economic discipline. As AWS service consumption grows, operational costs and overhead shouldn't grow linearly. Instead, IT should encourage new tenants that migrate from data centers to AWS to slowly shift toward a self-service delivery model and adopt the DevOps operations model. Creating and offering an AWS Cloud operations service catalog enables organizations to efficiently take full advantage of AWS' flexibility and modularity. T-Mobile, whose journey to AWS Cloud management started more than 2 years ago, uses a service catalog to enforce operational discipline in the Cloud. Their catalog is custom crafted for each Cloud-based IT workload. This session provides insight into the AWS Cloud operations strategy and its transformation, the creation of a Cloud operations service catalog, and how this approach supports reliable engineering on AWS. Sponsored by Accenture. AWS Competency Partner
ENT216 - Workshop: Migrating Microsoft Applications to AWS
In this workshop, we will explore the different approaches to migrating Microsoft applications to AWS. Well walk through the concerns and considerations to take into account while planning a migration, and learn how to develop and implement a migration plan to move applications from on-premises (or traditional hosting) to AWS. This session will use a case study format to dive deep into the details of how to successfully plan an application migration. To keep it real, teams will work through planning a SharePoint migration that integrates in with an existing Active Directory.
ENT217 - Migrating Enterprise Messaging to the Cloud
Enterprises rely on messaging to integrate services and applications and to exchange information critical to running their business. However, managing and operating dedicated message-oriented middleware and underlying infrastructure creates costly overhead and can compromise reliability. In this session, enterprise architects and developers learn how to improve scalability, availability, and operational efficiency by migrating on-premises messaging middleware to a managed cloud service using Amazon SQS. Hear how Capital One is using SQS to migrate several core banking applications to the cloud to ensure high availability and cost efficiency. We also share some exciting new SQS features that allow even more workloads to take advantage of the cloud.
ENT218 - Simplify Cloud Migration with AWS Server Migration Service
Data center migrations are often long and complex projects that involve thousands of servers. Without automated tools, these migrations can consume hundreds of hours. The recently launched AWS Server Migration Service speeds and simplifies the migration of on-premises servers to the AWS Cloud. SMS automates incremental replication of live server volumes, speeding up the migration process and significantly reducing the manual tasks required for server migration. This session for IT managers introduces key features and benefits of SMS, and includes a demo of the SMS end-to-end workflow.
ENT304 - Fueling Migration: Shortcutting your Application Portfolio Assessment
For many large enterprises, gathering data on your applications and developing a detailed business case for an all-in migration to AWS can be a time consuming activity. Instead of waiting on a full analysis to be completed, a global healthcare company decided to take a shortcut by reinventing the portfolio assessment process in a brilliantly simple manner. This new streamlined approach enabled them to immediately take advantage of the increased speed and innovative capabilities of AWS. In this session, IT directors and enterprise architects can learn these techniques to quickly identify which of their own applications fit cloud-friendly patterns; including, how to build a prioritized and multi-threaded migration queue that maximizes velocity and business value.
ENT305 - Setting the Stage for Instant Success: Getting the Most Out of Your AWS Deployment
Ticketmaster will share their playbill to managing and optimizing their AWS deployment in real-time. Learn how this ticket-sales giant was able to deploy, train and obtain utilization of the Cloudyn solution for their significant AWS deployment - and get rapid results. Ticketmaster relies on a large and dynamic AWS workload of thousands of instances across multiple regions and availability zones. As their reliance on the cloud increases, so does their need for actionable insights. Using Cloudyn, Ticketmaster is now able to confidently grow their cloud and focus efforts on expanding their core business. In this session, presenters will also demonstrate how to use big data analytics to manage and optimize cloud investments, and leverage best practices to realize cloud potential. Attendees will also hear about cloud industry trends, the latest developments in optimization and how to plan for 2017. Session sponsored by Cloudyn.
ENT306 - Transforming BP’s Cloud Environment at Industrial Scale
In this talk, we will describe the journey and approach to moving BP to the Cloud. The IT team at BP was faced with a huge legacy estate, a highly conservative risk posture, and the most complex and heavily outsourced environment of any of the Oil Gas super-majors. Working with Cloudreach and AWS, we have been on a three-year change journey to design a global Cloud migration program. We will describe how our strategy has evolved, our 3rd generation approach and operating model for brokering Cloud services at scale, creating a controls framework for Cloud security, developing the business case to transition from a large fixed cost base and, finally, developing an at-scale migration approach. The scale and complexity of our business has thrown up many difficult operational issues and we hope that this talk will provide valuable insight, to other organisations on a similar journey, into how BP has tackled these challenges. This session is sponsored by Cloudreach. AWS Competency Partner
ENT307 - Saving at Scale with Reserved Instances
As your Reserved Instance portfolio grows, tracking your ROI becomes more critical. Youll need the right tools, metrics, and strategies in place to make sure youre maximizing RI savings and minimizing unused RI hours. In this session, youll learn how scaled organizations like yours track, report and optimize their Reserved Instance savings. Attendees walk away with new strategies and tools for choosing Reserved Instances and properly account for your RIs after theyve been purchased. Topics will include: - Accounting for Reserved Instances and their upfront costs - Centralizing and streamlining the RI buying process - Choosing the right RIs to buy and modify Session sponsored by Cloudability.
ENT308 - Building a Solid Business Case for Cloud Migration
Learn how to create a compelling business case for a large-scale migration to AWS. We present a framework and tools for creating your business case, and guidelines for using AWS services to maximize value and optimize cost for migrations to the AWS Cloud. Learn a new way of thinking about cost that includes automation, new technologies, organizational change, and other factors.
ENT310 - Choosing the Right Partner for Your AWS Journey
You have decided to shut down a data center and move your applications to AWS, and you know youll need some help. Whether you need just tools, boots on the ground to move those applications, or someone to teach you how, you can find help among the group of elite partners that the AWS Partner Network has assembled to assist customers with meeting their business objectives. In this session, you will discover the variety of services our partners provide, learn how to select the appropriate partner for your needs, and hear about the experiences of customers who used partners to accomplish big objectives on AWS.
ENT311 - Identifying Your Migration Options: the 6 Rs
Learn about the 6 Rs of migrating, migration patterns, and how to map your portfolio to the patterns. Discover how your business drivers and overall IT strategy determine the migration patterns you need, and learn best practices for selecting the appropriate Rs.
ENT312 - Develop Your Migration Toolkit
Learn about some of the most useful and popular tools that you can leverage at various stages of a migration project. These tools will allow your teams to focus on coordinating the migration and automating as many migration activities as possible.
ENT313 - FINRA in the Cloud: the Big Data Enterprise
Large-scale enterprise migration can be a complex undertaking, especially for organizations that re-architect solutions to leverage the benefits of the Cloud. FINRA, which regulates US equities and options markets, recently completed a 2.5-year migration and re-architecture of its Big Data platform. Their platform consumes billions of market events every day. FINRA has developed scalable platforms and services on AWS that enable migrating enterprise applications and business functions to the Cloud quickly. Their data management platform takes advantage of AWS storage and compute products. In this session, IT influencers and decision makers will learn lessons from FINRAs migration, including how to create an enterprise-class Cloud architecture and which technology skills are required for transitioning to the Cloud. We also share examples of the business value FINRA has realized.
ENT314 - Insurance and Financial Services' Journey to Cloud: Designing a Hybrid Cloud and Migrating Regulated Workloads to AWS
Migrating regulatory workloads can be complex and time consuming. Come learn how Guardian Life Insurance migrated their regulated workloads to the AWS Cloud. Guardian Life's hybrid cloud journey started with moving web and application tiers into AWS, while keeping legacy repositories on premises. Key planning activities included vendor selection, migration planning, hybrid architecture design, compliance ensurance, and AWS and partner tools selection. Guardian Life shares how they defined security, compliance, and data management controls and solutions, which enabled them to migrate entire applications to AWS. The session also covers lessons learned and key considerations for hybrid cloud architectures.
ENT315 - Start Your Cost Optimization Program: Learning from Intuit’s Cloud Journey
How do you optimize costs when migrating user-facing services from a proprietary data center to the Cloud? When Intuit faced this thorny question, they formed a cost-optimization team, built visibility into cost and usage data, decided when to use automation, and developed a culture of cost consciousness. Who do you select for this cross-functional team? How do you empower the team to reach across business units? Which metrics should you focus on first? How do you align your organization with this cause? Which roadblocks and issues could you face along the way? How will you create an internal chargeback model? Hear from the Intuit team about their journey, learn from the challenges that they overcame, and hear about the results they've achieved.
ENT317 - VMware and AWS Together - VMware Cloud on AWS
VMware CloudTMon AWS brings VMwares enterprise class Software-Defined Data Center software to Amazons public cloud, delivered as an on-demand, elastically scalable, cloud-based VMware sold, operated and supported service for any application and optimized for next-generation, elastic, bare metal AWS infrastructure. This solution enables customers to use a common set of software and tools to manage both their AWS-based and on-premises vSphere resources consistently. Further virtual machines in this environment have seamless access to the broad range of AWS services as well. This session will introduce this exciting new service and examine some of the use cases and benefits of the service. The session will also include a VMware Tech Preview that demonstrates standing up a complete SDDC cluster on AWS and various operations using standard tools like vCenter.
ENT318 - Enterprise Fundamentals: Use AWS to Secure Your DevOps Pipeline Like a Bank Would
Continuous delivery can be challenging, especially for enterprises that deal with strict compliance requirements, like those in the financial services sector. AWS and Stelligent frequently work together with many large financial services enterprises to build solutions that enable customers to run their business faster and more safely on AWS. Together, we help customers ensure the security of the source code used to trigger builds, insert strict business controls at run time, and continuously inspect running infrastructure to ensure compliance. In this session, we share highly effective techniques that you can incorporate into your continuous delivery system to provide bank-level controls and security, and faster deployments. We explore a strong encryption pattern for handling build artifacts in a continuous delivery pipeline, a simple process for inspecting AWS CloudFormation templates to ensure that business rules are in compliance before a template makes AWS API calls, and a runtime inspector that uses AWS Lambda and AWS Config rules to ensure that running infrastructure is always in compliance.
ENT320 - 20k in 20 Days - Agile Genomic Analysis
AstraZeneca is a global science-led biopharmaceutical company developing innovative medicines used by millions of patients worldwide. With AWS, AstraZeneca processed more exomes in 20 days than during the previous 3 years, enabling scientists to receive results more quickly, develop medicines faster, and treat more patients sooner. AstraZeneca was able to identify ~20% more patients with actionable variants of major cancer types by using VarDict, an internally developed, open source contributed variant caller, at scale in the Cloud. Learn how AstraZeneca used the AWS Cloud Services to rapidly develop and scale an asynchronous architecture to meet an urgent business opportunity, accelerating the speed of scientific discovery in a cost effective manner.
ENT321 - Enterprise IT as a Service: Empowering the Digital Experience for Broadspectrum
Join Broadspectrum as they share how they achieve their business goals using a cloud-first IT strategy and AWS for as a Service deployments. To support new customer projects, Broadspectrum frequently needs to set up new sites or offices. This often requires setting up infrastructure for a specific site for only the duration of the project. Learn how Broadspectrum leverages AWS and Wipro's Boundary Less Data Center Solution to enable on-demand provisioning of site-in-a-box. Gard Little, analyst from IDC, Stephen Orban, AWS Head of Enterprise Strategy, and Ramesh Nagarajan, SVP of Integrated Services at Wipro, join the discussion. Session sponsored by Wipro.
ENT401 - Unlocking the Four Seasons of Migrations and Operations: Enterprise Grade, Cloud Assured with Infosys and AWS
To keep pace with todays hybrid IT landscape, enterprises have to innovate faster while ensuring compatibility with legacy IT systems. They also need to reduce costs and increase business value through responsive infrastructure backed by flexible, intelligent, and context-aware agile IT processes. To do this, they need service assurance in four distinct areas: Security Operations management Workload migration Enterprise IT integration The challenge lies in achieving this assurance not through traditional methods but through extreme automation, driven by enterprise-centric business and IT policies, while embracing the ability to learn continuously and be proactive with respect to the environments context. As an Advanced and Premier consulting partner in the AWS Partner Network (APN), Infosys leverages its AWS investments to jointly develop solutions that help clients drive greater business value faster. As a bonus, youll also hear from one of our key clients, who shares their experience of working with Infosys on their AWS migration needs. This session is sponsored by Infosys. AWS Competency Partner
FIN301 - Fraud Detection with Amazon Machine Learning on AWS
In this session, we provide programmatic guidance on building tools and applications to detect and manage fraud and unusual activity specific to financial services institutions. Payment fraud is an ongoing concern for merchants and credit card issuers alike and these activities impact all industries, but are specifically detrimental to Financial Services. We provide a step-by-step walkthrough of a reference solution to detect and address credit card fraud in real time by using Apache Apex and Amazon Machine Learning capabilities. We also outline different resource and performance optimization options and how to work data security into the fraud detection workflow.
FIN302 - Disaster Recovery and Business Continuity for Systemically Important Financial Institutions
Modern financial services organizations rely heavily on technology and automated systems to run business-as-usual. However, if this technology were interrupted by natural disasters or other events, there could be a devastating impact on investors and market participants, and in turn your reputational brand. In this session, we provide a step-by-step disaster recovery solution employed by a major exchange. This solution leverages Amazon EC2 Container Service to provide Docker containers, Weave Net to support a multicast overlay network that enables high volume multicast feeds in a cloud environment, and AWS CloudFormation for the ability to easily create and manage AWS assets. The session also covers the importance of redundancy (not just operationally, but for SEC compliance reasons as well) and how financial services organizations can increase geographical diversification of their primary and disaster recovery data centers. We dive deep into each major component of the solution.
FIN303 - Use AWS to Secure Your DevOps Pipeline Like a Bank
Continuous Delivery can be challenging, especially for enterprises that deal with strict compliance requirements like those in the financial services sector. AWS and Stelligent frequently work together with many large Financial Services Enterprises to help incorporate their capabilities securely on the Cloud. From security of the source code used to trigger builds, to the insertion of strict business controls at run time, and out to the continuous inspection of the running infrastructure to ensure compliance, we are helping to build capabilities that are enabling them to run their business faster and safer on AWS. In this breakout session we will share some very effective techniques that can be incorporated into your Continuous Delivery capability that bring bank level controls and security while enabling faster deployments.We will be looking at a strong encryption pattern for handling Build Artifacts in a Continuous Delivery Pipeline, a simple process for inspecting CloudFormation to ensure business rules are in compliance prior to AWS api calls, and a runtime inspector with Lambda and AWS Config Rules that can be leveraged to ensure that running infrastructure is always in compliance. Attendees will benefit from real-world scenarios and code examples of how to incorporate these techniques into their capabilities.
FIN304 - Grid Computing for Risk Management on AWS
In this session, we explain how Financial Services organizations can leverage AWS grid computing capabilities to perform large-scale calculations for risk management purposes. Numerous financial services companies face the same basic challenge: modeling multiple scenarios with different risk factors simultaneously or in quick succession in order to make informed decisions that maximize gains and minimize financial loss. Some examples of these workloads include Monte Carlo simulations, price model validation and back-testing, and risk calculations for hedging and capital optimization strategies. We provide programmatic guidance around what AWS services to use when running a grid computing cluster that requires thousands of cores and specific industry use cases and key benefits around speed and costs that the AWS platform, auto-scaling capabilities, and various compute services can help achieve.
GAM301 - How EA Leveraged Amazon Redshift and AWS Partner 47Lining to Gather Meaningful Player Insights
In November 2015, Capital Games launched a mobile game accompanying a major feature film release. The back end of the game is hosted in AWS and uses big data services like Amazon Kinesis, Amazon EC2, Amazon S3, Amazon Redshift, and AWS Data Pipeline. Capital Games will describe some of their challenges on their initial setup and usage of Amazon Redshift and Amazon EMR. They will then go over their engagement with AWS Partner 47lining and talk about specific best practices regarding solution architecture, data transformation pipelines, and system maintenance using AWS big data services. Attendees of this session should expect a candid view of the process to implementing a big data solution. From problem statement identification to visualizing data, with an in-depth look at the technical challenges and hurdles along the way.
GAM302 - Sony PlayStation: Breaking the Bandwidth Barrier using Soft State and ELB Properties
As systems and user bases grow, a once abundant resource can become scarce. While scaling out PlayStation services to millions of users at over a 100,000 requests/second, network throughput became a precious resource to optimize for. Alex and Dustin talk about how the microservices that power Playstation achieved low latency interactions while conserving on precious network bandwidth. These services powered by Amazon Elastic Load Balancing and Amazon DynamoDB benefitted from soft-state optimizations, a pattern that is used in complex interactions such as searching through a users social graph in sub 100 ms, or a users game library in 7 ms. As a developer utilizing Amazon Web services, you will discover new patterns and implementations which will better utilize your network, instances, and load balancers in order to deliver personalized experiences to millions of users while saving costs.
GAM303 - Develop Games Using Lumberyard and Leverage AWS with GameLift, Cloud Canvas, and Twitch
This session focuses on the cloud-connected features of Amazon Lumberyard. We introduce Amazon GameLift core concepts, complete a step-by-step walkthrough of deploying a multiplayer game to the cloud, and demonstrate how to rapidly scale a game based on player demand. We also cover Amazon Lumberyards Cloud Canvas, which empowers developers to leverage services such as databases, data storage, and methods to execute game logic within AWS without needing to configure or maintain servers. Finally we discuss integrating Twitch to your Lumberyard game with Twitch ChatPlay and JoinIn.
GAM401 - Riot Games: Standardizing Application Deployments Using Amazon ECS and Terraform
This session covers how the data team at Riot Games utilizes ECS to consolidate and improve disparate deployment and hosting strategies across a wide range of applications deployed via Docker containers. The team will share how cluster management and container orchestration through ECS enables the team to quickly adopt and evolve shared service hosting solutions as Riot continues its journey towards becoming a multi-game studio.
HLC301 - Data Science and Healthcare: Running Large Scale Analytics and Machine Learning on AWS
Working with Amazon Web Services AWS and 1Strategy, an Advance AWS Consulting partner; the Cambia Health Data Sciences teams have been able to deploy HIPAA compliant and secured AWS Elastic Map Reduce (EMR) data pipelines on the cloud. In this session, we will dive deep into the architectural components of this solution and you will learn how utilizing AWS services has helped Cambia decrease processing time for analytics, increase application flexibility and accelerate speed to production. The second part of the session is going to cover machine learning and its role in reducing cost and improving quality of care. The healthcare community must rely on advanced analytics and machine learning toanalyzemultiple facets of healthcare data and processit at scale to gain insights on things that matter. You will learn why AWS is a well suited platform for machine learning. We will take you through the steps of building a machine learning model using Amazon ML for a real world problem of predicting patient readmissions.
HLC302 - AWS Infrastructure for a Global Population Health Platform, Cerner’s HealtheIntent
This talk will feature Cerners HealtheIntent, a programmable platform for facilitating analytics and care delivery at the scale of a population. The population health platform utilizes AWS services for storage, compute, networking and databases to maintain business continuity and provide on-demand disaster recovery. You will learn how Cerner has used AWS Snowball to migrate petabytes of data into AWS S3, leverages AWS Direct Connect to replicate incremental updates and utilizes AWS CloudFormation to automate reactive and dynamic provisioning of EC2 instances and other resources. You will also hear first-hand about how AWS enables Cerner to expand HealtheIntent into new global markets.
HLC303 - Embracing DevSecOps while Improving Compliance and Security Agility and Posture
This session will demonstrate how to embrace DevSecOps to improve your security and compliance agility and posture within the highly regulated HIPAA environment. We will cover compliance frameworks, data decoupling strategies to fully utilize AWS, and best practices learned from the industry most active cloud adopters.
HLC304 - Building IoT Applications with AWS and Amazon Alexa
Alexa, what is the Internet of Things? Now that technology is small enough to be embedded in everyday devices, Healthcare has an opportunity to exploit the extraordinary potential of connecting ordinary devices. In this presentation, we explain how to rapidly build an IoT system and how to drive the Cloud with your voice on an Amazon Echo. In addition to describing how to use Alexa, we explore using AWS IoT, Lambda, Amazon SNS, and DynamoDB.
IOT201 - NEW LAUNCH! Introducing AWS Greengrass
AWS has launched AWS Greengrass, a platform that extends the AWS Cloud onto your devices so they can act locally on the data they generate, while still taking advantage of the cloud. In this session we will talk about how Greengrass works and what you can do with it. You will also hear from early customers who will discuss their use cases for Greengrass and how it fits into their overall IoT strategy.
IOT202 - Internet of Things (IoT) Edge and Device Services
AWS IoT edge and device services make it easy to get started and scale quickly along with your business needs. Medical equipment, industrial machinery, building automation, and simple device to trigger services, are just a few physical-world use cases that are benefiting from elastic cloud computing while meeting the local execution requirements and real time responsiveness. This session covers the intersection between the device and cloud industries, and the way AWS and our customers will shape the future of those industries together. We will showcase how our customers are using AWS IoT Button, the IoT Device SDKs, and other AWS services to improve the existing business models, invent new way of working, and balance the benefits of the cloud services with the need for local execution.
IOT203 - 1-Click Enterprise Innovation with the AWS IoT Button
Enterprises can use the AWS IoT button to provide unique innovative experiences to their consumers, simplifying the customer interface, and increasing engagement and brand loyalty. Enterprises and developers alike can use this single-purpose device to build fully customized IoT applications, or select from a list of predefined blueprints to instantly configure the device for unique use cases.
IOT204 - Introduction to AWS IoT in the Cloud
Whats new with AWS IoT? This is an Introduction to the AWS IoT Platform and an overview of new features. Join us for a discussion on the features launched over the last year, and the best practices on how to use the AWS IoT Platform to get your device data into the cloud.
IOT205 - IoT Analytics: Insights for a Connected World
Establishing connectivity between devices and the cloud is the first step to building great IoT applications. Generating insights with these data streams, and acting on it in real-time, that creates value for your business. Join us and learn how to expand the operational picture of your IoT solution with AWS IoT. We will show you how to use metrics to enable new possibilities in generating insights and decision making engines. And we will connect to Salesforce to leverage your CRM data and empower service agents in this IoT context, using a solar panel monitoring and maintenance as a demonstration of AWS IoT features paired with Salesforce IoT Cloud.
IOT302 - IoT Security: The New Frontiers
Only year ago we launched AWS IoT, and at re:Invent we showed how AWS IoT makes it easy to secure millions of connected devices. However, we have learned from our customers that a number of unique security challenges for the Internet of Things (IoT) exist. In this presentation we will go through our updated vision on device provisioning, firmware updates, securing communication between devices and AWS IoT endpoints, updating policies to control the actions devices can make, isolating the control and data planes in firmware, and securely integrating AWS IoT with broader AWS services ecosystem. In addition to discussing basic AWS IoT provisioning patterns, we will concentrate on new features, such as Bring Your Own Certificate, support for ECC certificates and cyphers, and provisioning devices in constrained environments using Just-In-Time Registration. This session will demonstrate how customers can identify and protect their IoT infrastructure from rogue devices or targeted attacks.
IOT303 - Innovation After Installation: Establishing a Digital Relationship with AWS IoT
In 2004, approximately 400 billion fax pages were sent. Twelve years later, its about 4% of that number. The pace of technological change is rapid, but most devices live in the field for 10 to 15 years. Its hard to maintain competitive value in the face of constant technology improvement, but IoT is changing that. Well examine the architectures that allows AWS IoT customers like Pitney Bowes to connect devicesto the cloud and enrich the client experience though personalized analytics andrecommendations,automated supplies replenishment, and just-in-time self-service.
IOT304 - IoT and Beyond: Building IoT Solutions for Exploring the Final Frontier
Jet Propulsion Laboratory is a well-known innovator in outer space, particularly in its search for life out there. JPL is now innovating in the physical space to improve life here. AWS IoT is critical to their innovations. See a re:Invent preview about how JPL, as an early adopter of AWS IoT, has prototyped voice control to ask questions of the room, the budget, or the system. Theyve also used it for controlling lights and sound to detect cyber security threats, rapid prototyping of robots, low-cost virtual windows to the outside, and much more. The results have been excellent. JPL will demonstrate and talk about these prototypes, including what worked and what didnt. They will also share the promise integrated serverless computing holds.
IOT306 - IoT Visualizations and Analytics
In this workshop, we focus on visualizations of IoT data using ELK, Amazon Elasticsearch, Logstash, and Kibana orAmazon Kinesis.We will dive into how these visualizations can give you new capabilites and understanding when interacting with your device data from the context they provide on the world around them.
IOT307 - IoT State of the Union
In this general session, AWS IoT experts will present an in-depth look at the current state of the Internet of Things. Learn about trends and industry use cases. Hear how other organizations are using AWS IoT to connect devices to the cloud. Explore some of the most recent IoT announcements as we kick off the IoT re:Source Mini Con.
IOT308 - Smart Home Solution with Amazon Alexa
Simply connecting the things that were never connected before as part of the Internet-of-Things is leading to new data insights that translate into meaningful change. AWS and Intel are working together to provide a secure, scalable edge-to-cloud solution for IoTapplications. Intel gateways utilize Windriver Helix Device Cloud to authenticate the device with AWS IoT and initiate secure data transport as well as provide a framework for active edge device management and over the air software and security updates. With AWS and Intel, you can implement an IoT solution quickly and with minimal upfront investments, seamless connectivity, and deliver enhanced security from device to network to cloud, then use AWS Big Data services to drive business insight. In this workshop you will learn how to use the Amazon Alexa Skills Kit SDK along with the Windriver Helix Device Cloud API to teach Alexa how to pass requests and data between Alexa and an Intel IoT Gateway based solution how to provision IoT edge devices, sensors and smart switches as part of an End-to-End SmartHome solution. We will guide you through the scripting and code snippets required to do this. We will provide you with a real live experience of making Amazon Alexa and Windriver Helix Device Cloud device management work together and testing the flexible Smart Home reference framework.
IOT309 - NEW LAUNCH! Lambda Everywhere
You can now execute Lambdas almost anywhere originating in the cloud,and on connected devices with AWS Greengrass. This advanced technical session explores Lambda Functions and what it means to use them across these diverse environments. We will treat the cloud as the brain, using local Lambdas for local executions. This way devices can react instinctively, much like the autonomic nervous system, operating in the periphery and responsible for collecting and filtering information, implementing simple and time-sensitive local actions reflexively.
IOT401 - Serverless IoT Back Ends
This session will cover common customer implementations and patterns for building connected/smart home implementations with AWS IoT. This includes the end-user experience for onboarding a smart home appliance and then integrating it with the AWS ecosystem (for targeted push notifications, predictive maintenance, and so on). iRobot will join us to discuss their smart home integrations with the Roomba 980 and AWS IoT.
IOT402 - IoT Blueprints: Optimizing Supply for Smart Agriculture from the Farm to the Table
30% of global food produce is wasted in the supply chain: storage, movement, and delivery. By using AWS IOT to enable sensors to manage the supply chain and big data to understand patterns, industrial companies can gain efficiencies in electricity and transportation.
IOT403 - Robots: The Fading Line Between Real and Virtual Worlds
Experience how live virtual 3D worlds rendered with Amazon Lumberyard, a free, cross-platform, 3D game engine, interconnect with IoT devices in the real world. This session will illustrate how AWS IoT can be used to remotely control animate objects with Bluetooth, such as a Sphero robot. By using AWS Lambda and AWS IoT, we will create a bi-directional communication between moving robots that detect collisions and a virtual world rendered in Amazon's game engine. By using AWS IoT with the Alexa Skills Kit and the Amazon Echo, we will learn how to control the physical and virtual robots through voice.
LFS301 - Building a Platform for Collaborative Scientific Research on AWS
This session discusses the architecture, formation, and usage of a collaborative HPC/big data scientific research and analysis environment on AWS.The pharmaceutical industry trend toward joint ventures and collaborations has created a need for new platforms in which to work together. We'll dive into architectural decisions for building collaborative systems. Examples include how such a platform allowed Human Longevity, Inc. to accelerate software deployment to production in a fast-paced research environment, and how Celgene uses AWS for research collaboration with outside universities and foundations.
LFS302 - Continuous Compliance in the AWS Cloud for Regulated Life Sciences Applications within Merck
Life sciences organizations running regulated workloads in the cloud can move from point-in-space testing of their environment to near real-time testing to achieve continuous compliance with the mandates of auditors and regulation entities. Get deep insights into some of the AWS services used to accomplish continuous compliance such as Amazon CloudTrail, Amazon CloudWatch, AWS Config, Amazon VPC, Amazon S3, and Amazon EC2. Get real-world use cases of how heavily regulated environments within Merck maintain governance and control over a shared environment. We also discuss the automated tools used by Merck to eliminate manual processes and streamline IT management.
LFS303 - How to Build a Big Data Analytics Data Lake
For discovery-phase research, life sciences companies have to support infrastructure that processes millions to billions of transactions. The advent of a data lake to accomplish such a task is showing itself to be a stable and productive data platform pattern to meet the goal.We discuss how to build a data lake on AWS, using services and techniques such as AWS CloudFormation, Amazon EC2, Amazon S3, IAM, and AWS Lambda. We also review a reference architecture from Amgenthat uses a data lake to aid in their Life ScienceResearch.
LFS304 - Large-Scale, Cloud-Based Analysis of Cancer Genomes: Lessons Learned from the PCAWG Project
The PanCancer Analysis of Whole Genomes (PCAWG) project is a large-scale, highly distributed research collaboration designed to identify common patterns of mutations across 2,800 cancer genomes.The use of public and private clouds were instrumental in analyzing this dataset using current best practice containerized pipelines. This session describes the technical infrastructure built for the project, how we leveraged cloud environments to perform the core analysis, and the lessons learned along the way.
MAC201 - Getting to Ground Truth with Amazon Mechanical Turk
Jump-start your machine learning project by using the crowd to build your training set. Before you can train your machine learning algorithm, you need to take your raw inputs and label, annotate, or tag them to build your ground truth. Learn how to use the Amazon Mechanical Turk marketplace to perform these tasks. We share Amazon's best practices, developed while training our own machine learning algorithms, and walk you through quickly getting affordable and high-quality training data.
MAC202 - Deep Learning in Alexa
Neural networks have a long and rich history in automatic speech recognition. In this talk, we present a brief primer on the origin of deep learning in spoken language, and then explore todays world of Alexa. Alexa is the AWS service that understands spoken language and powers Amazon Echo. Alexa relies heavily on machine learning and deep neural networks for speech recognition, text-to-speech, language understanding, and more. We also discuss the Alexa Skills Kit, which lets any developer teach Alexa new skills.
MAC203 - NEW LAUNCH! Introducing Amazon Rekognition
This session will introduce you to Amazon Rekognition, a new service that makes it easy to add image analysis to your applications. With Rekognition, you can detect objects, scenes, and faces in images. You can also search and compare faces. Rekognitions API lets you easily build powerful visual search and discovery into your applications. With Amazon Rekognition, you only pay for the images you analyze and the face metadata you store. There are no minimum fees and there are no upfront commitments. To get started with Rekognition, simply log in to the Rekognition console to try the service with sample photos or your own photos. Join this session and learn more about Amazon Rekognition!
MAC204 - NEW LAUNCH! Introducing Amazon Polly
This session will introduce you to Amazon Polly, a new deep learning service that turns text into lifelike speech. Polly enables existing applications to speak as a first class feature and creates the opportunity for entirely new categories of speech-enabled products from mobile apps and cars, to devices and appliances. Polly includes 47 lifelike voices and support for 24 languages, so you can select the ideal voice and distribute your speech-enabled applications in many geographies. Polly is easy to use you just send the text you want converted into speech to the Polly API, and Polly immediately returns the audio stream to your application so you can play it directly or store it in a standard audio file format, such as MP3. Polly supports Speech Synthesis Markup Language (SSML) tags like prosody so you can adjust the speech rate, pitch, or volume. Polly is a secure service that delivers all of these benefits at high scale and at low latency. You can cache and replay Pollys generated speech at no additional cost. Polly lets you convert 5M characters per month for free during the first year. Pollys pay-as-you-go pricing, low cost per request, and lack of restrictions on storage and reuse of voice output make it a cost-effective way to enable speech synthesis everywhere. Join this session to learn more and find out how you get can started with Amazon Polly, today!
MAC205 - Deep Learning at Cloud Scale: Improving Video Discoverability by Scaling Up Caffe on AWS
Deep learning continues to push the state of the art in domains such as video analytics, computer vision, and speech recognition. Deep networks are powered by amazing levels of representational power, feature learning, and abstraction. This approach comes at the cost of a significant increase in required compute power, which makes the AWS cloud an excellent environment for training. Innovators in this space are applying deep learning to a variety of applications. One such innovator, Vilynx, a startup based in Palo Alto, realized that the current pre-roll advertising-based models for mobile video werent returning publishers' desired levels of engagement. In this session, we explain the algorithmic challenges of scaling across multiple nodes, and what Intel is doing on AWS to overcome them. We describe the benefits of using AWS CloudFormation to set up a distributed training environment for deep networks. We also showcase Vilynxs contributions to video discoverability, and explain how Vilynx uses AWS tools to understand video content. This session is sponsored by Intel.
MAC206 - Machine Learning State of the Union Mini Con
With the growing number of business cases for artificial intelligence (AI), machine learning (ML) and deep learning (DL) continue to drive the development of cutting edge technology solutions. We see this manifested in computer vision, predictive modeling, natural language understanding, and recommendation engines. During this full afternoon of sessions and workshops, learn how you can develop your own applications to leverage the benefits of these services. Join this State of the Union presentation to hear more about ML and DL at AWS and see how Motorola Solutions is leveraging these state-of-the-art technologies to solve public safety challenges, and how Ohio Health intends to inject AI into the medical system.
MAC301 - Transforming Industrial Processes with Deep Learning
Deep learning has revolutionized computer vision by significantly increasing the accuracy of recognition systems. This session will discuss how the Amazon Fulfillment Technologies Computer Vision Research team has harnessed deep learning toidentify inventory defectsin Amazons warehouses. Beginning with a brief overview ofhow orders on Amazon.com are fulfilled, the talk will describe a combination of hardware and software that usescomputer vision and deep learning that visually examine bins of Amazon inventorytolocatepossiblemismatches between the physical inventory and inventory records. With thegrowth ofdeep learning, the emphasis of new system design shifts from clever algorithms toinnovativewaystoharnessavailable data.
MAC302 - Leveraging Amazon Machine Learning, Amazon Redshift, and an Amazon Simple Storage Service Data Lake for Strategic Advantage in Real Estate
The Howard Hughes Corporation partnered with 47Lining to develop a managed enterprise data lake based on Amazon S3. The purpose of the managed EDL is to fuse relevant on-premises and third-party data to enable Howard Hughes to answer its most valuable business questions. Their first analysis was a lead-scoring model that uses Amazon Machine Learning (Amazon ML) to predict propensity to purchase high-end real estate. The model is based on a combined set of public and private data sources, including all publicly recorded real estate transactions in the US for the past 35 years. By changing their business process for identifying and qualifying leads to use the results of data-driven analytics from their managed data lake in AWS, Howard Hughes increased the number of identified qualified leads in their pipeline by over 400% and reduced the acquisition cost per lead by more than 10 times. In this session, you will see a practical example of how to use Amazon ML to improve business results, how to architect a data lake with Amazon S3 that fuses on-premises, third-party, and public data sets, and how to train and run an Amazon ML model to attain predictive accuracy.
MAC303 - Zillow Group: Developing Classification and Recommendation Engines with Amazon EMR and Apache Spark
Customers are adopting Apache Spark ‒ an open-source distributed processing framework ‒ on Amazon EMR for large-scale machine learning workloads, especially for applications that power customer segmentation and content recommendation. By leveraging Spark ML, a set of machine learning algorithms included with Spark, customers can quickly build and execute massively parallel machine learning jobs. Additionally, Spark applications can train models in streaming or batch contexts, and can access data from Amazon S3, Amazon Kinesis, Amazon Redshift, and other services. This session explains how to quickly and easily create scalable Spark clusters with Amazon EMR, build and share models using Apache Zeppelin and Jupyter notebooks, and use the Spark ML pipelines API to manage your training workflow. In addition, Jasjeet Thind, Senior Director of Data Science and Engineering at Zillow Group, will discuss his organization's development of personalization algorithms and platforms at scale using Spark on Amazon EMR.
MAC304 - NEW LAUNCH! Introducing Amazon Lex
Amazon Lex is a service for building conversational interfaces into any applications using voice and text. With Lex, the same deep learning engine that powers Amazon Alexa is now available to any developer, enabling you to build sophisticated, natural language chatbots into your new and existing applications. Amazon Lex provides the deep functionality and flexibility of natural language understanding (NLU) and automatic speech recognition (ASR) to allow you to build highly engaging user experiences with lifelike, conversational interactions. In this introductory session, find out how Lex provides deep functionality and flexibility to empower you to define entirely new categories of products that are made possible through conversational interfaces.
MAC306 - Using MXNet for Recommendation Modeling at Scale
For many companies, recommendation systems solve important machine learning problems. But as recommendation systems grow to millions of users and millions of items, they pose significant challenges when deployed at scale. The user-item matrix can have trillions of entries (or more), most of which are zero. To make common ML techniques practical, sparse data requires special techniques. Learn how to use MXNet to build neural network models for recommendation systems that can scale efficiently to large sparse datasets.
MAC307 - Predicting Customer Churn with Amazon Machine Learning
In this session, we take a specific business problempredicting Telco customer churnand explore the practical aspects of building and evaluating an Amazon Machine Learning model. We explore considerations ranging from assigning a dollar value to applying the model using the relative cost of false positive and false negative errors. We discuss all aspects of putting Amazon ML to practical use, including how to build multiple models to choose from, put models into production, and update them. We also discuss using Amazon Redshift and Amazon S3 with Amazon ML.
MAC308 - NEW LAUNCH! Workshop: Hands on with Amazon Lex, Amazon Polly, and Amazon Rekognition
Amazon AI services bring natural language understanding (NLU), automatic speech recognition (ASR), visual search and image recognition, text-to-speech (TTS), and machine learning (ML) technologies within reach of every developer. Amazon Lex make it easy to build sophisticated text and voice chatbots, powered by Alexa; Amazon Rekognition provides deep learning-based image recognition; and Amazon Polly turns text into lifelike speech. In this workshop, you'll get a chance to use each of the new deep learning services. We'll see you there!
MAC401 - Scalable Deep Learning Using MXNet
Deep learning continues to push the state of the art in domains such as computer vision, natural language understanding and recommendation engines. One of the key reasons for this progress is the availability of highly flexible and developer friendly deep learning frameworks. During this workshop, members of the Amazon Machine Learning team will provide a short background on Deep Learning focusing on relevant application domains and an introduction to using the powerful and scalable Deep Learning framework, MXNet. At the end of this tutorial youll gain hands on experience targeting a variety of applications including computer vision and recommendation engines as well as exposure to how to use preconfigured Deep Learning AMIs and CloudFormation Templates to help speed your development. Module 1. Background on Deep Learning Module 2. Walkthrough on Setting up AMIs, CloudFormation Templates and other Deep Learning Frameworks on AWS Module 3. A Peek Under the MXNet Hood (MXNet Internals) Module 4. Hands on with MXNet: NDArrays, Symbols, And Mechanics of Training Deep Neural Networks Module 5. Hands on with MXNet: Application Examples Targeting Compute Vision and Recommendation Engines Prerequisites: 1. An AWS account isnt required (but you should get one anyway, theyre free..) 2. Familiarity with Python and Jupyter Notebooks 3. Some experience in machine Learning would be helpful 4. Bring your laptop.
MAC403 - Automatic Grading of Diabetic Retinopathy through Deep Learning
Diabetic retinopathy (DR) is the leading cause of blindness for people aged 20 to 64, and afflicts more than 120 million people worldwide. Fortunately, vigilant monitoring greatly improves the chance to preserve ones eyesight. This work used deep learning to analyze images of the retina and fundus for automated diagnosis of DR on a grading scale from 0 (normal) to 4 (severe). We achieved substantial improvement in accuracy compared to traditional approaches and continued advances by using a small auxiliary dataset that provided low-effort, high-value supervision. Data for training and testing, provided by the 2015 Kaggle Data Science Competition with over 80,000 high resolution images (4 megapixels), required Amazon EC2 scalability to provide the GPU hardware needed to train a convolutional network with over 2 million parameters. For the competition, we focused on accurately modeling the scoring system, penalizing bad mistakes more severely, and combatting the over-prevalence of grade-0 examples in the dataset. We explored ideas first at low resolution on low-cost single-GPU instances. After finding the best methodology, we showed it could be scaled to equivalent improvements at high resolution, using the more expensive quad-GPU instances more effectively. This prototype model placed 15 out of 650 teams across the world with a kappa score of 0.78. Weve now advanced the model via a new architecture that integrates the prototype and a new network specialized in finding dot hemorrhages, critical to identifying early DR. By annotating a small set of 200 images for hemorrhages, the performance jumped to a kappa of 0.82. We believe strategies that employ a bit more supervision for more effective learning are pivotal for cracking deep learnings greatest weakness: its voracious appetite for data.
MAE301 - Accelerating the Transition to Broadcast and OTT Infrastructure in the Cloud: Spotlight on Building Media Services on AWS and Elemental
In this session, we show how to seamlessly transition VOD, live, and other advanced media workflows from on-premises deployments to the cloud. Cinpolis will provide an overview of their transcoding solution on AWS and how they have seamlessly expanded the solution increasing their customer reach. We'll show real world examples of the API calls used to configure and control all elements of the workflow including compression and origination. And how standard AWS services can be media-optimized with Elemental Technologies to form a robust live solution.
MAE302 - Turner's cloud native media supply chain for TNT, TBS, Adult Swim, Cartoon Network, CNN
As Turner continues to make the transition from a traditional broadcast organization to a consumer-centric, data-driven media company, we are being challenged to re-think our approach to content supply. There is a need to achieve new levels of agility, flexibility and scalability to meet the rapidly evolving demands of our top media brands- including TBS, TNT, Cartoon Network, Adult Swim and CNN. To that end, we are transitioning the infrastructure that acquires, processes and distributes media for consumer-facing systems to the cloud. At the core of this environment is our Supply Chain Management application. The SCM app provides business and technical process management via anHTML based UI framework,State Machine, Rules Engine, Cost Model, Forms Service. We took advantage of several AWS specific services, includingLambda, S3,Dynamo DB, SNS,Elastic, Cloud Formation and Code Commit. The entire system is instance-less with all application code running in either the browser or within Lambda's. To ease development and debugging we created a method to run all JS libraries in the browser,switchingto Lambda when we deploy with Code Commit. Cloud media processing infrastructure is BEING created on demand via an integration with SDVI. The SDVI andSCM apps exchange events and datavia SNS and S3.
MAE303 - Discovery Channel's Broadcast Workflows and Channel Origination on AWS
Media delivery requirements are continually changing, driven by accelerating mobile, tablet, smart TV, and set-top technology advances. Broadcasters need agile solutions to the changing media and entertainment landscape that don't require multiyear projects with large upfront investments. In this session, we walk through Discovery Communications' migration of its broadcast playout and channel origination to AWS. Discovery Communications is a leader in nonfiction media, reaching more than 3 billion cumulative viewers in 220 countries and territories. Traditionally, broadcast origination for content delivered to telecommunications companies, cable TV, and satellite has existed only in on-premises data centers. In this session, we walk through Discovery's migration of broadcast playout supporting hundreds of channels worldwide to AWS. We show how Discovery has not only reduced their TCO but also has improved their agility by launching new channels on demand. We also walk through how channel origination is being deployed in a secure, automated fashion, and with a level of high availability that exceeds what is possible in a traditional data center.
MAE304 - High Performance Cinematic Production in the Cloud
The process of making a film is highly complex, and comprises of multiple workflows across story development, pre-production, production, post-production and final distribution. Given the size and amount of media and assets associated with each stage, high performance infrastructure is often essential to meeting deadlines. In this session we will take a deeper dive at running a full cinematic production in the cloud, with a focus on solutions for each of the production stages. We will also look at best practices around design, optimization, performance, scheduling, scalability and low latency utilizing AWS technologies such as EC2, Lambda, Snowball, Direct Connect, and Partner Solutions.
MBL201 - AWS Mobile State of the Union - Serverless, New User Experiences, Auth, and More
AWS provides a range of services and tools to help you create industry leading, cloud-enabled mobile apps that can securely scale to millions of users globally. Join Amit Patel, GM of AWS Mobile, to hear our vision for mobile apps and the cloud, industry trends, recent product launches, and success stories directly from our customers. We'll walk through and demo the AWS Mobile offerings for building compelling cloud-enabled mobile apps and for engaging your app users. Youll learn how to use these offerings (serverless API Gateway/Lambda, Cognito, and new services) to make it easy to develop both your iOS and Android frontend, as well as your mobile backend.
MBL202 - Taking Data to the Extreme
As GoPro expands into content networks and launches new products, new challenges have appeared. One of the most critical challenges facing GoPro during this period of rapid growth is their ability to make effective use of massive amounts of data. Every day, GoPro collects increasing amounts of data generated by internet connected consumer devices (smart cameras, smart drones), GoPro mobile apps, GoPro content networks, GoPro e-commerce sales, and social media. This data ranges from raw camera logs to refined and well-structured e-commerce datasets. In the past, it took GoPro months to understand new inbound data and determine how to transform or augment it for analysis. To streamline this process and bridge the gap between tech-savvy engineers and data-savvy analysts, GoPro is creating an analysis loop, which informs product usage trends and product insights. This analysis loop serves a large ecosystem of GoPro executives, product managers, engineers, data scientists, and business analysts through an integrated technology pipeline consisting of Apache Kafka, Apache Spark Streaming, Clouderas distribution of Hadoop, and Tableaus Data Visualization Software as the end user analytical tool. Session sponsored by Tableau Software.
MBL204 - How Netflix Achieves Email Delivery at Global Scale with Amazon SES
Companies around the world are using Amazon Simple Email Service (Amazon SES) to send millions of emails to their customers every day, and scaling linearly, at cost. In this session, you learn how to use the scalable and reliable infrastructure of Amazon SES. In addition, Netflix talks about their advanced Messaging program, their challenges, how SES helped them with their goals, and how they architected their solution for global scale and deliverability.
MBL205 - Test and Debug Your Mobile Apps on Real Android and iOS Devices in the Cloud with AWS Device Farm
With the widespread and accelerating adoption of mobile devices, delivering high quality mobile experiences became essential for success in today's marketplace In this session, learn how you can deliver top quality apps by using AWS Device Farm for testing and debugging on hundreds of unique Android and iOS devices in the cloud. A Device Farm customer, Gannett, describes their journey from a time-consuming manual activity to fast and repeatable automation activity when using Chef, Scalr, AWS, and Jenkins. The new cloud infrastructure allows Gannett to scale on-demand to build and test 100+ Android native apps on Amazon EC2 instances and Device Farm to speed up their continuous integration pipeline with efficiency.
MBL206 - AWS Customers Saving Lives with Mobile and IoT Technology
Join us for a compelling session to learn how technological innovation can save lives. Hear from a customer that uses AWS very differently and has the ability to move rapidly to address pressing societal problems with the aid of AWS. Learn how mobile and IoT technology allows them to scale and accelerate their impact. Speakers talk about why they selected specific AWS services, and how that allowed their teams to be agile and focus more on their solutions instead of undifferentiated plumbing. The ShotSpotter Flex gunfire data service uses acoustic sensors to detect gunshots in the field. Data is then transmitted to the AWS Cloud, and further processed to validate signatures and, if necessary, alert emergency dispatch teams.
MBL304 - NEW LAUNCH! How to Enable Real-Time Mobile App Engagement with Amazon Pinpoint
Amazon Pinpoint, a new AWS service, makes it easy to run targeted campaigns to improve user engagement. Pinpoint helps you understand app user behavior, define who to target, what push notification to send, when to deliver the notifications, and track results.
MBL305 - Developing Mobile Apps and Serverless Microservices for Enterprises using AWS
In this session, we dive deep into how to apply the serverless microservices approach to developing mobile and web applications on AWS, for both business-to-consumer and business-to-employee enterprise applications. We discuss use cases, scenarios, best practices, and design patterns around user authentication, authorization, backend microservices, API management, analytics, user engagement, security, and integration with enterprise apps and data sources. We go into details about how to use AWS Lambda, Amazon API Gateway, Amazon Cognito, and other AWS services to develop mobile applications and backend microservices.
MBL306 - Serverless Authentication and Authorization: Identity Management for Serverless Architectures
By leveraging serverless architectures, startups and enterprises are building and running modern applications and services with increased agility and simplified scalabilityall without managing a single server. Many applications need to manage user identities and support sign-in/sign-up. In this session, we dive deep on how to support millions of user identities, as well as how to integrate with social identity providers (such as Google and Facebook) and existing corporate directories. You learn the real-world design patterns that AWS customers use to implement authentication and authorization. By combining Amazon Cognito identity pools and user pools with API Gateway, AWS Lambda, and AWS IAM, you can add security without adding servers.
MBL307 - NEW LAUNCH! Enhance Your Mobile Apps with AI Using Amazon Lex
Amazon Echo and Alexa have shown that voice interfaces provide significant benefits to users interactions are easy, fast, and context-driven. In this hands-on session, youll see how to add compelling voice and chat interfaces to your mobile apps, using Amazon Lexfor processing conversations and triggering corresponding actions in your backend systems, all without having to manage any infrastructure. Youll leave knowing how to build apps that can Find me a nearby hotel or Reorder supplies for the copier.
MBL310 - Add User Sign-In, User Management, and Security to your Mobile and Web Applications with Amazon Cognito
Secure user sign-up and sign-in is critical for many mobile and web applications. Amazon Cognito is the easiest way to secure your mobile and web applications by providing a comprehensive identity solution for end user management, registration, sign-in, and security. In this product deep dive, we will walk through Cognitos feature set, which includes serverless flows for user management and sign-in, a fully managed user directory, integrations with existing corporate directories, and many other features. In addition, we will cover key use cases and discuss the associated benefits.
MBL402 - Re-imagining Insurance Processes with AWS Mobile Services
Insurance processes are fragmented and manual in nature. For the next generation of digitally savvy consumers, New York Life Labs uses modern technology and analytics toengage with and provide best-in-class customer service. This session will discuss how New York Life Labs is mobilizing key insurance technology components including customer service, underwriting, policy administration, and claims processing. This session willfocus on how New York Life Labs is taking a mobile-first approach in redesigning claims processing with responsive design, automated workflows, and AWS mobile services. Some of the services discussed will be Amazon Cognito, Amazon SNS, AWS Lambda, Amazon DynamoDB, and Amazon SWF, and how they all come together.
MBL403 - Deep Dive: Building and Delivering Mobile Apps for the Enterprise Using AWS Mobile Hub
This session takes a practical approach to developing real-world enterprise applications (business-to-consumers and business-to-employees) using a serverless backend that can scale to virtually unlimited users without any infrastructure to manage. Learn how to develop enterprise apps using AWS Mobile Hub, Amazon Cognito, Amazon API Gateway, and AWS Lambda to implement best practices for authentication/authorization, cloud logic, and secure integration of existing enterprise resources and user directories. Finally, see how to employ a robust application lifecycle (build, test, and deliver) and implement a multi-stage rollout to production.
MBL404 - Deep-Dive: Native, Hybrid and Web patterns with Serverless and AWS Mobile Services
In this deep-dive session, we outline how to leverage the appropriate AWS services for sending different types and sizes of data, such as images or streaming video. We'll cover common real-world scenarios related to authentication/authorization, access patterns, data transfer and caching for more performant Mobile Apps. You learn when you should access services such as Amazon Cognito, Amazon DynamoDB, Amazon S3, or Amazon Kinesis directly from your mobile app, and when you should route through Amazon API Gateway and AWS Lambda instead. Additionally, we cover coding techniques across the native, hybrid, and mobile web using popular open-source frameworks to perform these actions efficiently, and with a smooth user experience.
NET201 - Creating Your Virtual Data Center: VPC Fundamentals and Connectivity Options
In this session, we walk through the fundamentals of Amazon VPC. First, we cover build-out and design fundamentals for VPC, including picking your IP space, subnetting, routing, security, NAT, and much more. We then transition into different approaches and use cases for optionally connecting your VPC to your physical data center with VPN or AWS Direct Connect. This mid-level architecture discussion is aimed at architects, network administrators, and technology decision-makers interested in understanding the building blocks that AWS makes available with Amazon VPC and how you can connect this with your offices and current data center footprint.
NET202 - DNS Demystified: Getting Started with Amazon Route 53, featuring Warner Bros. Entertainment
Whether youre running a simple website, a mobile app, or a suite of business applications, DNS is a fundamental part of any architecture in the cloud. In this mid-level architecture session, well cover everything you need to get started with Amazon Route 53, AWSs highly-available DNS service. Youll learn how to use public DNS, including routing techniques such as weighted round-robin, latency-based routing, and geo DNS; how to configure DNS failover using health checks; how and when to use private DNS within your Virtual Private Cloud (VPC); and how Amazon Route 53 interacts with Amazon EC2s DNS for instance naming and DNS resolution across your network. We will conclude the session with a real-world migration example. Warner Bros. Entertainment recently completed a full DNS migration to Route 53. Vahram Sukyas, Vice President, Application Infrastructure Operations at Warner Bros. Entertainment, will share details on his team's architecture, migration strategy, and lessons learned which are useful for enterprises and startups alike.
NET203 - From EC2 to ECS: How Capital One uses Application Load Balancer Features to Serve Traffic at Scale
Capital One began moving to AWS just two years ago. Every day, the amount of traffic we serve from the cloud continues to grow. With development teams having the freedom to choose their own technology stacks, many teams have quickly started moving applications to Docker. In this session, learn how Capital One uses a combination of the Elastic Load Balancing service along with Application Load Balancer features to increase deployment speed and reliability.
NET204 - NEW LAUNCH IPv6 in the Cloud: Protocol and AWS Service Overview
Recently, AWS announced support for Internet Protocol version 6 (IPv6) for several AWS services, providing significant capabilities for applications and systems that need IPv6. This session provides an overview of IPv6 and covers key aspects of AWS support for the protocol. We discuss Amazon S3 and S3 Transfer Acceleration, Amazon CloudFront and AWS WAF, Amazon Route 53, AWS IoT, Elastic Load Balancing, and the virtual private cloud (VPC) environment of Amazon EC2. The presentation assumes solid knowledge of IPv4 and those AWS services.
NET205 - Future-Proofing the WAN and Simplifying Security On Your Journey To The Cloud
You can leverage the agility and scale of cloud services and consolidate your data centers, but performance and security are only as good as your WAN. Is this session, learn best practices for connecting cloud, data center, and branch sites through public and private networking to maximize performance, minimize costs, and simplify security. Session sponsored by Level 3.
NET301 - Cloud agility and faster connectivity with AT NetBond and AWS
Learn how the AT MPLS VPN with the network of tomorrows virtualized network functions and Software Defined Networking (SDN) will help you create and deliver agile workloads for your Enterprise. Youll also learn how AT combines trending viability of open standards-based software for broader network applications. Additionally, youll see how the AT NetBond API integration with AWS Direct Connect removes the complexity and enables on-demand, private connection within minutes via a self-service portal. AT NetBond connects your people, your data, and your business directly to your AWS services. This fast, highly secure, scalable, private network connection increases performance, while improving control and delivering a better ROI for your enterprise applications. Join us for an informative session on how you can enhance your cloud connectivity with AT and AWS.Session sponsored by AT
NET302 - Global Traffic Management with Amazon Route 53 Traffic Flow
As companies grow and expand their global footprint, it becomes increasingly critical to make systems highly available while also improving responsiveness to end-users. Companies are choosing to place their applications closer to end-users to improve performance, which introduces the complications of how to route end-user traffic to the most appropriate endpoints and how to most efficiently route traffic within internal systems. In this session, learn how customers are using Route 53's Traffic Flow service for global traffic management, improving performance and availability for end users while reducing IT management cost. We will walk through how to use Traffic Flow to manager traffic to your applications' globally-distributed endpoints to optimize for constraints such as endpoint load, the health of your resources, geographic restrictions, and Internet latency. We'll demonstrate how you can configure multiple routing policies and take advantage of code control and versioning for easier management of your DNS and traffic management configuration.
NET303 - NextGen Networking: New Capabilities for Amazon’s Virtual Private Cloud
Amazons Virtual Private Cloud (Amazon VPC) lets you provision a logically isolated section of the Amazon Web Services (AWS) cloud and gives you complete control over your virtual networking environment. Amazon VPC continues to evolve with new capabilities and enhancements. These features give you increasingly greater isolation, control, and visibility at the all-important networking layer. In this session, we review some of the latest changes, discuss their value, and describe their use cases.
NET304 - Moving Mountains: Netflix's Migration into VPC
Netflix was one of the earliest very large AWS customers. By 2014, we were running hundreds of applications in Amazon EC2. That was great, until we needed to move to VPC. Given our scale, uptime requirements, and the decentralized nature of how we manage our production environment, the VPC migration (still ongoing) presented particular challenges for us and for AWS as it sought to support our move. In this talk, we discuss the starting state, our requirements and the operating principles we developed for how we wanted to drive the migration, some of the issues we ran into, and how the tight partnership with AWS helped us migrate from an EC2-Classic platform to an EC2-VPC platform.
NET305 - Extending Datacenters to the Cloud: Connectivity Options and Considerations for Hybrid Environments
Many enterprises on their journey into the cloud require consistent and highly secure connectivity between their existing data center and AWS footprints. In this session, we walk through the different architecture options for establishing this connectivity using AWS Direct Connect and VPN. With each option, we evaluate the considerations and discuss risk, performance, encryption, and cost. As we walk through these options, we try to answer some of the most common questions that typically arise from enterprises that tackle design and implementation. You'll learn how to make connectivity decisions that are suitable for your workloads, and how to best prepare against business impact in the event of failure.
NET307 - NEW LAUNCH IPv6 in the Cloud: Virtual Private Cloud Deep Dive
In this session, we explore AWS support for IPv6 for full end-to-end connectivity for EC2 instances inside of a VPC. IPv6 on EC2 instances introduces new capabilities and interesting new wrinkles into the VPC model. Customer VPCs receive IPv6 addresses from an Amazon address block, and existing features such as security groups, network ACLs, route tables, peering, and gateways have been enhanced to support IPv6. Finally, we look at some future capabilities planned for IPv6 in VPC.
NET308 - How Gree Launched New Games Faster and More Securely with AWS Marketplace and Amazon VPC
Gree is a gaming company with the distinction of launching the worlds first mobile social game in 2007 and later went on to develop a rich portfolio of original games. With an active acquisition and development strategy, Gree is transforming the way games are built. Learn how Gree leveraged Amazon VPCs and Aviatrix Enterprise cloud-defined networking in AWS Marketplace to rapidly integrate acquisitions and develop new games to maintain competitive differentiation in this fast moving segment. Hear about the use case, requirements and lessons learned. Learn how you can transitioned to Amazon VPCs in the middle of the development process.
NET401 - Another Day, Another Billion Packets
In this session, we walk through the Amazon VPC network presentation and describe the problems we were trying to solve when we created it. Next, we walk through how these problems are traditionally solved, and why those solutions are not scalable, inexpensive, or secure enough for AWS. Finally, we provide an overview of the solution that we've implemented and discuss some of the unique mechanisms that we use to ensure customer isolation, get packets into and out of the network, and support new features like VPC endpoints.
NET402 - Deep Dive: AWS Direct Connect and VPNs
As enterprises move to the cloud, robust connectivity is often an early consideration. AWS Direct Connect provides a more consistent network experience for accessing your AWS resources, typically with greater bandwidth and reduced network costs. This session dives deep into the features of AWS Direct Connect and VPNs. We discuss deployment architectures and demonstrate the process from start to finish. We show you how to configure public and private virtual interfaces, configure routers, use VPN backup, and provide secure communication between sites by using the AWS VPN CloudHub.
NET403 - Elastic Load Balancing Deep Dive and Best Practices
Elastic Load Balancing automatically distributes incoming application traffic across multiple Amazon EC2 instances for fault tolerance and load distribution. In this session, we go into detail about Elastic Load Balancing configuration and day-to-day management, as well as its use in conjunction with Auto Scaling. We explain how to make decisions about the service and share best practices and useful tips for success.
NET404 - Making Every Packet Count
Many applications are network I/O bound, including common database-based applications and service-based architectures. But operating systems and applications are often untuned to deliver high performance. This session uncovers hidden issues that lead to low network performance, and shows you how to overcome them to obtain the best network performance possible.
NET405 - Amazon s2n: Cryptography and Open Source at AWS
Launched in June of 2015, s2n is an AWS open-source implementation of the TLS and SSL network security protocols, which focus on security, simplicity, and performance. With development led by engineers from Amazon EC2, Amazon S3, Amazon CloudFront, and AWS security and cryptography services, s2n is a unique opportunity to observe how we develop and test security and availability for critical software at AWS. Learn how we iterate and code, how we automate software verification beyond the usual code reviews, and how open source works at Amazon.
SAC201 - Lessons from a Chief Security Officer: Achieving Continuous Compliance in Elastic Environments
Does meeting stringent compliance requirements keep you up at night? Do you worry about having the right audit trails in place as proof? Cengage Learnings Chief Security Officer, Robert Hotaling, shares his organizations journey to AWS, and how they enabled continuous compliance for their dynamic environment with automation. When Cengage shifted from publishing to digital education and online learning, they needed a secure elastic infrastructure for their data intensive and cyclical business, and workload layer security tools that would help them meet compliance requirements (e.g., PCI). In this session, you will learn why building security in from the beginning saves you time (and painful retrofits) later, how to gather and retain audit evidence for instances that are only up for minutes or hours, and how Cengage used Trend MicroDeep Securityto meet many compliance requirements and ensured instances were instantly protected as they came online in a hybrid cloud architecture.Session sponsored by Trend Micro, Inc. AWS Competency Partner
SAC202 - Workshop: Secure Your Web Application with AWS WAF and Amazon CloudFront
In this workshop, we help you understand how you can help protect your web applications from threats cost effectively by using AWS WAF and Amazon CloudFront. As attacks and attempts to exploit vulnerabilities in web applications become more sophisticated and automated, having an effective web request filtering solution becomes key to keeping your users'data safe. We will cover common attack vectors and what you can do to mitigate them. You will learn how to leverage AWS WAF in conjunction with Amazon CloudFront to detect unwanted traffic and block it using simple configurations and automations. Prerequisites: Participants should have an AWS account established and available for use during the workshop. Please bring your own laptop.
SAC303 - Become an AWS IAM Policy Ninja in 60 Minutes or Less
Are you interested in learning how to control access to your AWS resources? Have you ever wondered how to best scope down permissions to achieve least privilege permissions access control? If your answer to these questions is yes, this session is for you. We take an in-depth look at the AWS Identity and Access Management (IAM) policy language. We start with the basics of the policy language and how to create and attach policies to IAM users, groups, and roles. As we dive deeper, we explore policy variables, conditions, and other tools to help you author least privilege policies. Throughout the session, we cover some common use cases, such as granting a user secure access to an Amazon S3 bucket or to launch an Amazon EC2 instance of a specific type.
SAC304 - Predictive Security: Using Big Data to Fortify Your Defenses
In a rapidly changing IT environment, detecting and responding to new threats is more important than ever. This session shows you how to build a predictive analytics stack on AWS, which harnesses the power of Amazon Machine Learning in conjunction with Amazon Elasticsearch Service, AWS CloudTrail, and VPC Flow Logs to perform tasks such as anomaly detection and log analysis. We also demonstrate how you can use AWS Lambda to act on this information in an automated fashion, such as performing updates to AWS WAF and security groups, leading to an improved security posture and alleviating operational burden on your security teams.
SAC305 - How AWS Automates Internal Compliance at Massive Scale using AWS Services
Is your IT environment getting bigger and more complex than your compliance team can handle? Get a peek under the hood of how the AWS Compliance team manages and automates security assurance and compliance in the AWS environment. Well tell you what were doing to automate controls, match up huge data sets to validate compliance, how we perform game day simulations of entire region outages, and how we manage our ever-present external audits. With each example, well give you some ideas on how to use AWS services to manage the security and compliance of your AWS and on-prem environments. In this session, Chad Woolf, Director of Risk and Compliance for AWS, and Sara Duffer, Director of Security Assurance Automation discusses how the AWS Compliance team uses AWS services like Amazon Inspector, Amazon CloudWatch Logs, AWS CloudTrail, and AWS Config to manage risk, compliance, and audit in the massive scale of the AWS IT environment.
SAC306 - Encryption: It Was the Best of Controls, It Was the Worst of Controls
Encryption is a favorite of security and compliance professionals everywhere. Many compliance frameworks actually mandate encryption. Thoughencryption is important,itis also treacherous. Cryptographic protocols are subtle, and researchers are constantly finding new and creative flaws in them. Using encryption correctly, especially over time, also is expensive because you have to stay up to date. AWS wants to encrypt data. And our customers, including Amazon, want to encrypt data. In this talk, welook at some of the challenges with using encryption, how AWS thinks internally about encryption, and how that thinking has informed the services we have built, the features we have vended, and our own usage of AWS.
SAC307 - The Psychology of Security Automation
Historically, relationships between developers and security teams have been challenging. Security teams sometimes see developers as careless and ignorant of risk, while developers might see security teams as dogmatic barriers to productivity. Can technologies and approaches such as the cloud, APIs, and automation lead to happier developers and more secure systems? Netflix has had success pursuing this approach, by leaning into the fundamental cloud concept of self-service, the Netflix cultural value of transparency in decision making, and the engineering efficiency principle of facilitating a paved road. This session explores how security teams can use thoughtful tools and automation to improve relationships with development teams while creating a more secure and manageable environment. Topics include Netflixs approach to IAM entity management, Elastic Load Balancingand certificate management, and general security configuration monitoring.
SAC308 - Hackproof Your Cloud: Responding to 2016 Threats
CloudCheckr Co-Founders Aaron Newman and Aaron Klein will highlight effective strategies and tools that AWS users can employ to improve their security posture. Specific emphasis will be placed upon leveraging native AWS services and the talk will include concrete steps that users can begin employing immediately.Session sponsored by CloudCheckr. AWS Competency Partner
SAC309 - You Can’t Protect What You Can’t See: AWS Security Monitoring Compliance Validation
Ensuring security and compliance across a globally distributed, large-scale AWS deployment requires a scalable process and a comprehensive set of technologies. This session will deep-dive into the AWS native monitoring and security services and some Splunk technologies leveraged globally to perform security monitoring across a large number of AWS accounts. You will learn about the collection plumbing including components of S3, Kinesis, CloudWatch, SNS, Dynamo DB and Lambda, as well as the tooling and processes used at Adobe to deliver scalable monitoring without managing an unwieldy number of API keys and input stanzas. Session sponsored by Splunk. AWS Competency Partner
SAC310 - Securing Serverless Architectures, and API Filtering at Layer 7
AWS serverless architecture components such as Amazon S3, Amazon SQS, Amazon SNS, CloudWatch Logs, DynamoDB, Amazon Kinesis, and Lambda can be tightly constrained in their operation. However, it may still be possible to use some of them to propagate payloads that could be used to exploit vulnerabilities in some consuming endpoints or user-generated code. This session explores techniques for enhancing the security of these services, from assessing and tightening permissions in IAM to integrating tools and mechanisms for inline and out-of-band payload analysis that are more typically applied to traditional server-based architectures.
SAC311 - Evolving an Enterprise-Level Compliance Framework with Amazon CloudWatch Events and AWS Lambda
Johnson Johnson is in the process of doing a proof of concept to rewrite the compliance framework that they presented at re:Invent 2014. This framework leverages the newest AWS services and abandons the need for continual describes and master rules servers. Instead, Johnson Johnson plans to use a distributed, event-based architecture that not only reduces costs but also assigns costs to the appropriate projects rather than central IT.
SAC312 - Architecting for End-to-End Security in the Enterprise
This session tells how our most mature, security-minded Fortune 500 customers adopt AWS while improving end-to-end protection of their sensitive data. Learn about the enterprise security architecture decisions made during actual sensitive workload deployments as told by the AWS professional services and the solution architecture team members who lived them. In this very prescriptive, technical walkthrough, we share lessons learned from the development of enterprise security strategy, security use-case development, security configuration decisions, and the creation of AWS security operations playbooks to support customer architectures.
SAC313 - Enterprise Patterns for Payment Card Industry Data Security Standard (PCI DSS)
Professional services has completed five deep PCI engagements with enterprise customers over the last year. Common patterns were identified and codified in various artifacts. This session introduces the patterns that help customers address PCI requirements in a standard manner that also meets AWS best practices. Hear customers speak about their side of the journey and the solutions that they used to deploy a PCI compliance workload.
SAC314 - GxP Compliance in the Cloud
GxP is an acronym that refers to the regulations and guidelines applicable to life sciences organizations that make food and medical products such as drugs, medical devices, and medical software applications. The overall intent of GxP requirements is to ensure that food and medical products are safe for consumers and to ensure the integrity of data used to make product-related safety decisions. The term GxP encompasses a broad range of compliance-related activities such as Good Laboratory Practices (GLP), Good Clinical Practices (GCP), Good Manufacturing Practices (GMP), and others, each of which has product-specific requirements that life sciences organizations must implement based on the 1) type of products they make and 2) country in which their products are sold. When life sciences organizations use computerized systems to perform certain GxP activities, they must ensure that the computerized GxP system is developed, validated, and operated appropriately for the intended use of the system. For this session, co-presented with Merck, services such as Amazon EC2, Amazon CloudWatch Logs, AWS CloudTrail, AWS CodeCommit, Amazon Simple Storage Service (S3), and AWS CodePipeline will be discussed with an emphasis on implementing GxP-compliant systems in the AWS Cloud.
SAC315 - Scaling Security Operations and Automating Governance: Which AWS Services Should I Use?
This session enables security operators to automate governance and implement use cases addressed by AWS services such as AWS CloudTrail, AWS Config Rules, Amazon CloudWatch Events, and Trusted Advisor. Based on the nature of vulnerabilities, internal processes, compliance regimes, and other priorities, this session discusses the service to use when. We also show how to detect, report, and fix vulnerabilities, or gain more information about attackers. We dive deep into new features and capabilities of relevant servicesand use an example from an AWS customer,Siemens AG, about how to best automate governance and scale. A prerequisite for this session is knowledge of security and basic software development using Java, Python,or Node.
SAC316 - Security Automation: Spend Less Time Securing Your Applications
As attackers become more sophisticated, web application developers need to constantly update their security configurations. Static firewall rules are no longer good enough. Developers need a way to deploy automated security that can learn from the application behavior and identify bad traffic patterns to detect bad bots or bad actors on the Internet. This session showcases some of the real-world customer use cases that use machine learning and AWS WAF (a web application firewall) with automated incident response and machine learning to automatically identify bad actors. We also present tutorials and code samples that show how customers can analyze traffic patterns and deploy new AWS WAF rules on the fly.
SAC317 - IAM Best Practices to Live By
This session covers AWS Identity and Access Management (IAM) best practices that can help improve your security posture. We cover how to manage users and their security credentials. We also explain why you should delete your root access keysor at the very least, rotate them regularly. Using common use cases, we demonstrate when to choose between using IAM users and IAM roles. Finally, we explore how to set permissions to grant least privilege access control in one or more of your AWS accounts.
SAC318 - Life Without SSH: Immutable Infrastructure in Production
This session covers what a real-world production deployment of a fully automated deployment pipeline looks like with instances that are deployed without SSH keys. By leveraging AWS CloudFormation along with Docker and AWS CodeDeploy, we show how we achieved semi-immutable and fully immutable infrastructures, and what the challenges and remediations were.
SAC319 - Architecting Security and Governance Across a Multi-Account Strategy
Whether it is per business unit or per application, many AWS customers use multiple accounts to meet their infrastructure isolation and billing requirements.In this session, wediscuss considerations, limitations, and security patterns when building out a multi-account strategy.Weexplore topics such as identity federation, cross-account roles, consolidated logging, and account governance.Thomson Reuters also presents their approach to a multi-account strategy. At the end of the session, we present an enterprise-ready, multi-account architecture that you can start leveraging today.We encourage you to attend Deep Dive: Implementing Security and Governance Across a Multi-Account Strategy for a deeper dive into the tools and implementation of this strategy.
SAC320 - Deep Dive: Implementing Security and Governance Across a Multi-Account Strategy
Now that you have decided to move forward with an AWS multi-account strategy, how should you help secure and manage your accounts? This session is a deep dive into some of the tools and processes that you can use to help ensure that you have a hardened blast radius between your AWS accounts while still being able to manage your security baseline across them. Some of the topics we cover are: How to perform monitoring and logging. Continuous compliance. Incident response in your multi-account strategy. This session builds on the Introduction: Architecting Security and Governance Across a Multi-Account Strategy session, which helps you architect a multi-account strategy for your organization.
SAC321 - Cyber Resiliency – surviving the breach
In this session, youll learn how to setup your AWS environment to avoid a cyber security attack and how to build a cloud specific incident response plan if your organization is breached. Cyber security expert and founder of Alert Logic Misha Govshteyn will share lessons learned from organizations that have protected themselves in AWS and from those that have been breached but quickly resolved their issues and implemented strong controls as a result. Joining Misha will be Sven Skoog, Senior Manager IT Security at Monotype who will discuss the cyber security posture they implemented within the AWS cloud and how they have built a robust process to ensure ongoing protection.Session sponsored by Alert Logic. AWS Competency Partner
SAC322 - NEW LAUNCH! AWS Shield—A Managed DDoS Protection Service
At re:Invent 2016, we are launching AWS Shield, a managed DDoS protection service. With AWS Shield, you can help protect Amazon CloudFront, Elastic Load Balancing, and Amazon Route 53 resources from DDoS attacks. In additionto introducing AWS Shield, this session presents some of the things we do behind the scenes to detect and mitigate Layer 3/4 network attacks and highlights ways you can use this new service to protect against Layer 7 application attacks.
SAC323 - NEW SERVICE: Centrally Manage Multiple AWS Accounts with AWS Organizations
AWS Organizations is a new administrative capability, which allows you to control multiple AWS accounts centrally. With Organizations, you can hierarchically organize and manage your AWS accounts and apply organizational controls across these accounts to meet your business needs. In this session, we cover the capabilities of AWS Organizations and discuss best practices when managing multiple AWS accounts.
SAC326 - How Harvard University Improves Scalable Cloud Network Security, Visibility, and Automation
With constantly evolving threats across the Internet, Harvard University deployed a security network platform to mitigate cyber threats, current and future, to protect institutional and research data. By using multiple geographic locations, best of breed equipment, and network automation, Harvard provides visibility, availability, and multilayer protections for their cloud network. This talk discusses the benefits, considerations, and lessons learned from using their security network platform at the edge of the cloud. Learn how Harvard designed and deployed the platform, utilizing serverless architecture to orchestrate the solution from within to protect their most sensitive data and afford students, faculty, and staff the flexibility of cloud computing.
SAC327 - No More Ransomware: How Europol, the Dutch Police, and AWS Are Helping Millions Deal with Cybercrime
Come to this session to learn how Europol, the Dutch police, Intel Security, and Kaspersky Lab have come together in an unprecedented collaboration of government and private-sector organizations. Open source ransomware code makes it easier to lock victims computers and encrypt their data, resulting in an alarming increase of cyber ransom. In response www.nomoreransom.org was created with the additional cooperation of AWS and Barracuda Web Application Firewall. Learn what tools are available to retrieve encrypted data and take a peek under the hood of this mission-critical website in the fight against ransomware. Perhaps because the site opposes ransomware, it has already received a number of attacks. Learn how AWS and Intel worked to rebuff these persistent assaults.
SAC401 - 5 Security Automation Improvements You Can Make by Using Amazon CloudWatch Events and AWS Config Rules
This session demonstrates 5 different security and compliance validation actions that you can perform using Amazon CloudWatch Events and AWS Config rules. This session focuses on the actual code for the various controls, actions, and remediation features, and how to use various AWS services and features to build them. Thedemos in this session include CIS Amazon Web Services Foundations validation; host-based AWS Config rules validation using AWS Lambda, SSH, and VPC-E; automatic creation and assigning of MFA tokens when new users are created; and automatic instance isolation based on SSH logons or VPC Flow Logs deny logs. This session focuses on code and live demos.
SAC402 - The AWS Hero’s Journey to Achieving Autonomous, Self-Healing Security
We are all embarking on a journey in the cloud that can be frightening at times, thrilling at others, but at all times filled with pitfalls and scary monsters that threaten the security of our infrastructure, applications, and data. The ultimate reward for all our hard work is to achieve a state of autonomous, self-healing security within our environment--one that can withstand any threats, whether internal or external. In this session, wewalk you through the steps you need to be successful in your journey, just like Ellie Mae and many other enterprises and agencies. Your journey starts with security automation, and from there you will push outside of your security comfort zone, thanks to the gift of enhanced visibility and omniscience. Next we use CloudFormation Templates and custom signatures to move through our next security challenge with speed, and finally, webuild auto-remediation into our security strategy with AWS Lambda workflows that enable the system to self-correct when misconfigurations occur. This fast-paced session will be filled code, best practices to help you in your quest, and even a few surprises about the ultimate destination of your journey. Session sponsored by Evident.io. AWS Competency Partner
SEC301 - Audit Your AWS Account Against Industry Best Practices: The CIS AWS Benchmarks
Audit teams can consistently evaluate the security of an AWS account. Best practices greatly reduce complexity when managing risk and auditing the use of AWS for critical, audited, and regulated systems. You can integrate these security checks into your security and audit ecosystem. Center for Internet Security (CIS) benchmarks are incorporated into products developed by 20 security vendors, are referenced by PCI 3.1 and FedRAMP, and are included in the National Vulnerability Database (NVD) National Checklist Program (NCP). This session shows you how to implement foundational security measures in your AWS account. The prescribed best practices help make implementation of core AWS security measures more straightforward for security teams and AWS account owners.
SEC302 - Workshop: Adhere to the Principle of Least Privilege by Using AWS Identity and Access Management (IAM) and Amazon Virtual Private Cloud (VPC)
AWS IAM and Amazon VPC offer powerful tools that help you adhere to the principle of least privilege in your resource permissions and network security settings. This workshop will start with the fundamentals of IAM and VPC security techniques and will give you hands-on experience in writing, testing, applying, troubleshooting, and auditing progressively more tightly scoped IAM policies. You will also get experience building and monitoring VPC security groups that grant only the access required to perform tasks.
SEC303 - Get the Most from AWS KMS: Architecting Applications for High Security
AWS Key Management Service provides an easy and cost-effective way to secure your data in AWS. In this session, you learn about leveraging the latest features of the service to minimize risk for your data. We also review the recently released Import Key feature that gives you more control over the encryption process by letting you bring your own keys to AWS.
SEC304 - Reduce Your Blast Radius by Using Multiple AWS Accounts Per Region and Service
This session shows you how to reduce your blast radius by using multiple AWS accounts per region and service, which helps limit the impact of a critical event such as a security breach. Using multiple accounts helps you define boundaries and provides blast-radius isolation. Thoughmanaging multiple accounts can be difficult, we will present an upcoming AWS solution that will help automate the process for controlling cross- account access by managing roles across multiple accounts.
SEC305 - Scaling Security Resources for Your First 10 Million Customers
Cloud computing offers many advantages, such as the ability to scale your web applications or website on demand. But howdo you scale your security and compliance infrastructure along with the business? Join this session to understand best practices for scaling your security resources as you grow from zero to millions of users. Specifically, you learn the following: How to scale your security and compliance infrastructure to keep up with a rapidly expanding threat base. The security implications of scaling for numbers of users and numbers of applications, and how to satisfy both needs. How agile development with integrated security testing and validation leads to a secure environment. Best practices and design patterns of a continuous delivery pipeline and the appropriate security-focused testing for each. The necessity of treating your security as code, just as you would do with infrastructure. The services covered in this session include AWS IAM, Auto Scaling, Amazon Inspector, AWS WAF, and Amazon Cognito.
SEC306 - Workshop: Choose Your Own SAML Adventure: A Self-Directed Journey to AWS Identity Federation Mastery
AWS supports identity federation using SAML (Security Assertion Markup Language) 2.0. Using SAML, you can configure your AWS accounts to integrate with your identity provider (IdP). Your federated users then are authenticated and authorized by your organization's IdP, and they can use single sign-on (SSO) to access AWS. In this workshop, you choose your own path through the exercises to direct yourself to the technologies and use cases that matter to you. We start by guiding you through deploying an IdP and configuring SAML federation for AWS, including federated CLI access. We will then continue to walk you through a number of advanced SAML use cases, including how to: Write S3 bucket policies for specific federated users. Use SAML attributes to enforce additional authorization requirements. Automate federation configurations across a large number of AWS accounts. Implement other advanced SAML use cases for AWS.
SEC307 - Microservices, Macro Security Needs: How Nike Uses a Multi-Layer, End-to-End Security Approach to Protect Microservice-Based Solutions at Scale
Microservice architectures provide numerous benefits but also have significant security challenges. This session presents how Nike uses layers of security to protect consumers and business. We show how network topology, network security primitives, identity and access management, traffic routing, secure network traffic, secrets management, and host-level security (antivirus, intrusion prevention system, intrusion detection system, file integrity monitoring) all combine to create a multilayer, end-to-end security solution for our microservice-based premium consumer experiences. Technologies to be covered include Amazon Virtual Private Cloud, access control lists, security groups, IAM roles and profiles, AWS KMS, NAT gateways, ELB load balancers, and Cerberus (our cloud-native secrets management solution).
SEC308 - Securing Enterprise Big Data Workloads on AWS
Security of big data workloads in a hybrid IT environment often comes as an afterthought. This session discusses how enterprises can architect secure, big-data workloads on AWS. We cover the application of authentication, authorization, encryption, and additional security principles and mechanisms to workloads leveraging Amazon Elastic MapReduce andAmazon Redshift.
SEC309 - Proactive Security Testing in AWS: From Early Implementation to Deployment Penetration Testing
Attend this session to learn about security testing your applications in AWS. Effective security testing is challenging, but multiple features and services within AWS make security testing easier. This session covers common approaches to testing, including how we think about testing within AWS, how to apply AWS services to your test setup, remediating findings, and automation.
SEC310 - Mitigating DDoS Attacks on AWS: Five Vectors and Four Use Cases
Distributed denial of service (DDoS) attack mitigation has traditionally been a challenge for those hosting on fixed infrastructure. In the cloud, users can build applications on elastic infrastructure that is capable of mitigating and absorbing DDoS attacks. What once required overprovisioning, additional infrastructure, or third-party services is now an inherent capability of many cloud-based applications. This session explains common DDoS attack vectors and how AWS customers with different use cases are addressing these challenges. As part of the session, we show you how to build applications that are resilient to DDoS and demonstrate how they work in practice.
SEC311 - How to Automate Policy Validation
Managing permissions across a growing number of identities and resources can be time-consuming and complex. Testing, validating, and understanding permissions before and after policy changes are deployed is critical to ensuring that your users and systems have the appropriate level of access. This session walks through the tools that are available to test, validate, and understand the permissions in your account. We demonstrate how to use these tools and how to automate them to continually validate the permissions in your accounts. The tools demonstrated in this session help you answer common questions such as: Which users and roles have access to perform powerful actions? Which users and roles have access to critical resources such as Amazon S3 buckets? Who is able to launch instances in a specific region?
SEC312 - re:Source Mini Con for Security Services State of the Union
AWS CISO Steve Schmidt presents the state of the union for re:Source Mini Con for Security Services. He addresses the state of the security and compliance ecosystem; large enterprise customer additions in key industries; the vertical view: maturing spaces for AWS security assurance (GxP, IoT, CIS foundations); and the international view: data privacy protections and data sovereignty. The state of the union also addresses a number of new identity, directory, and access services, and closes by looking at what's on the horizon.
SEC313 - Automating Security Event Response, from Idea to Code to Execution
With security-relevant services such as AWS Config, VPC Flow Logs, Amazon CloudWatch Events, and AWS Lambda, you now have the ability to programmatically wrangle security events that may occur within your AWS environment, including prevention, detection, response, and remediation. This session covers the process of automating security event response with various AWS building blocks, taking several ideas from drawing board to code, and gaining confidence in your coverage by proactively testing security monitoring and response effectiveness before anyone else does.
SEC314 - Common Considerations for Data Integrity Controls in Healthcare
This session is intended for customers seeking to build out a comprehensive plan around data integrity in the cloud. Information governance, system validation, PHI and PII records come with their own set of regulatory considerations - and by using tools such as AWS CloudFormation, Amazon Virtual Private Cloud, Directory Service, AWS SDKs, RDS, AWS CloudTrail, and Amazon CloudWatch portions of the governance burden can be lifted, offloaded, or partnered around. A look at the top ten data integrity controls will be considered, including a customer presentation featuring an example of a regulated quality management system. Finally, the physical import/export tool Snowball's addition to the AWS Business Associate Agreement (BAA) program will be discussed, along with the practical application and security implications of chain of custody.
SEC401 - Automated Formal Reasoning About AWS Systems
Automatic and semiautomatic mechanical theorem provers are now being used within AWS to find proofs in mathematical logic that establish desired properties of key AWS components. In this session, we outline these efforts and discuss how mechanical theorem provers are used to replay found proofs of desired properties when software artifacts or networks are modified, thus helping provide security throughout the lifetime of the AWS system. We consider these use cases: Using constraint solving to show that VPCs have desired safety properties, and maintaining this continuously at each change to the VPC Using automatic mechanical theorem provers to prove that s2ns HMAC is correct and maintaining this continuously at each change to the s2nsource code Using semi-automatic mechanical theorem provers to prove desired safety properties of protocols and code
SVR201 - NEW LAUNCH! Serverless Apps with AWS Step Functions
AWS Step Functions is a new, fully-managed service that makes it easy to coordinate the components of distributed applications and microservices using visual workflows. Step Functions is a reliable way to connect and step through a series of AWS Lambda functions so that you can build and run multi-step applications in a matter of minutes. This session shows how to use AWS Step Functions to create, run, and debug cloud state machines to execute parallel, sequential, and branching steps of your application, with automatic catch and retry conditions. We share how customers are using AWS Step Functions to reliably scale multi-step applications such as order processing, report generation, and data transformationall without managing any infrastructure.
SVR202 - What’s New with AWS Lambda
In this session, you'll learn whats new and hot with AWS Lambda. Come on a tour with Dr. Tim Wagner, General Manager of AWS Lambda, to learn what weve been working on and what we are planning for the future. You'll get a hands-on demonstration of some our newest features which will provide you with a launching pad for some of the later sessions in the day.
SVR301 - Real-time Data Processing Using AWS Lambda
Serverless architecture can eliminate the need to provision and manage servers required to process files or streaming data in real time. In this session, we will cover the fundamentals of using AWS Lambda to process data in real-time from push sources such as AWS Iot and pull sources such as Amazon DynamoDB Streams or Amazon Kinesis. We will walk through sample use cases and demonstrate how to set up some of these real-time data processing solutions. We'll also discuss best practices and do a deep dive into AWS Lambda real-time stream processing. You also hear from speakers from Thomson Reuters, who discuss how the company leverages AWS for its Product Insight service. The service provides insights to collect usage analytics for Thomson Reuters products. The speakers walk through its architecture and demonstrate how they leverage Amazon Kinesis Streams, Amazon Kinesis Firehose, AWS Lambda, Amazon S3, Amazon Route 53, and AWS KMS for near real-time access to data being collected around the globe. They also outline how applying AWS methodologies benefited its business, such as time-to-market and cross-region ingestion, auto-scaling capabilities, low-latency, security features, and extensibility.
SVR302 - Optimizing the Data Tier in Serverless Web Applications
AWS Lambda empowers developers to build cloud-native web applications or platforms using microservices architecture. The three-tier web app has been a cornerstone architecture pattern for decades in this space. This session walks you through the process of identifying the presentation, logic, and data tiers required to build web applications with AWS Lambda at the core. By using AWS Lambda as your logic tier, you have a wide number of data storage options for your data tier. AWS offers a wide range of database services to fit your application requirements. We dive into methodologies for picking the right database technology based on your application requirements. You learn the benefits of SQL vs. NoSQL database architectures for the data tier of your serverless app. We demonstrate connecting your serverless app to various AWS database offerings including Amazon RDS, Amazon Aurora, Amazon DynamoDB, and Amazon ElastiCache, and elaborate on the setup of each option with AWS Lambda. We also provide guidelines and best practices for implementing this architecture pattern, such as setting up a VPC on Lambda to connect to private resources, leveraging local caching for opportunistic performance improvement, choosing the right language runtime for different workload patterns, managing database connections, and more.
SVR303 - Coca-Cola: Running Serverless Applications with Enterprise Requirements
The value of building serverless applications on AWS applies to businesses of all sizes. For enterprisesthat own hundreds or thousands of applications, the cost of managing and securing servers could equal millions of dollars per year. In this session, we will discuss how building serverless applications on AWS can eliminate many of those costs and still meet enterprise requirements across a breadth of areas: including identity and access control, network security, performance, availability, deployment and operations. Also, Coca-Cola North America will give a glimpse into the tools and strategies theyve developed in order to make their next generation of digital marketing applications completely serverless, and give a peak into the innovative ways theyre planning to use AWS Lambda in the future.
SVR304 - bots + serverless = ❤
Chatbots are the new apps. Businesses of all sizes, from startups to enterprises, are looking for new ways to connect with their users through natural, conversational interfaces. Developers have started using chatbots to improve the productivity and efficiency of their operations. In this session, we show how to use AWS Lambda and other serverless offerings from AWS to build chatbots quickly and efficiently. We share examples from our recently concluded AWS Slack Hackathon with a full walkthrough of building a conversational chatbot in an easy, fast, and fun way, along with helpful tools, tips, and techniques.
SVR305 - ↑↑↓↓←→←→ BA Lambda Start
Ever wished you had a list of cheat codes to unleash the full power of AWS Lambda for your production workload? Come learn how to build a robust, scalable, and highly available serverless application using AWS Lambda. In this session, we discuss hacks and tricks for maximizing your AWS Lambda performance, such as leveraging customer reuse, using the 500 MB scratch space and local cache, creating custom metrics for managing operations, aligning upstream and downstream services to scale along with Lambda, and many other workarounds and optimizations across your entire function lifecycle. You also learn how Hearst converted its real-time clickstream analytics data pipeline from a server-based model to a serverless one. The infrastructure of the data pipeline relied on Amazon EC2 instances and cron jobs to shepherd data through the process. In 2016, Hearst converted its data pipeline architecture to a serverless process that relies on event triggers and the power of AWS Lambda. By moving from a time-based process to a trigger-based process, Hearst improved its pipeline latency times by 50%.
SVR306 - Serverless Computing Patterns at Expedia
In the middle of 2015, Expedia started using AWS Lambda for serverless computing. We built boilerplate templates in Node.js, Java, and Python so development teams could build and deploy serverless applications into AWS. Currently, we have 300 AWS Lambda functions processing 40 million invocations per day. In this session, we will discuss how development teams use boilerplate templates to create serverless applications with Amazon API Gateway and AWS Lambda and how they deploy them to AWS. We will cover patterns, architectural design choices, and the benefits --- like cost, scale, availability, and operations --- of running serverless applications.
SVR307 - Application Lifecycle Management in a Serverless World
Amazon API Gateway and AWS Lambda provide a new way of building applications by removing servers from the picture. But what does the removal of servers mean to tasks like deployment, monitoring, and debugging? How should you set up blue-green deployments or set alarms? Come learn all this and more, including how to use AWS services and tools like AWS CodePipeline, AWS CloudFormation, and Amazon CloudWatch to manage your serverless applications at high quality.
SVR308 - Content and Data Platforms at Vevo: Rebuilding and Scaling from Zero in One Year
Vevo has undergone a complete strategic and technical reboot, driven not only by product, but also by engineering. Since November 2015, Vevo has been replacing monolithic, legacy content services with a modern, modular, microservices architecture, all while developing new features and functionality. In parallel, Vevo has built its data platform from scratch to power internal analytics as well as a unique music video consumption experience through a new personalized feed of recommendations all in less than one year. This has been a monumental effort that was made possible in this short time span largely because of AWS technologies. The content team has been heavily using serverless architectures and AWS Lambda in the form of microservices, taking a similar approach to functional programming, which has helped us speed up the development process and time to market. The data team has been building the data platform by heavily leveraging Amazon Kinesis for data exchange across services, Amazon Aurora for consumer-facing services, Apache Spark on Amazon EMR for ETL + Machine Learning, as well as Amazon Redshift as the core analytics data store. In this session, Miguel and Alan walk you through Vevo's journey, describing best practices and learnings that the Vevo team has picked up along the way.
SVR309 - Wild Rydes Takes Off – The Dawn of a New Unicorn
Wild Rydes (www.wildrydes.com) needs your help! With fresh funding from its seed investors, Wild Rydes is seeking to build the worlds greatest mobile/VR/AR unicorn transportation system. The scrappy startup needs a first-class webpage to begin marketing to new users and to begin its plans for global domination. Join us to help Wild Rydes build a website using a serverless architecture. Youll build a scalable website using services like AWS Lambda, Amazon API Gateway, Amazon DynamoDB, and Amazon S3. Join this workshop to hop on the rocket ship! To complete this workshop, you'll need: Your laptop AWS Account AWS Command Line Interface Google Chrome git Text Editor
SVR310 - All Your Chats are Belong to Bots: Building a Serverless Customer Service Bot
Bots are eating the world! Wild Rydes (www.wildrydes.com), a new startup that is building the worlds leading mobile/VR/AR unicorn transportation system, has decided to use serverless chatbots to staff its customer service department. As it scales to millions of users, Wild Rydes needed a scalable way to meet the customer service needs of its customers instead of relying on human customer service agents. Wild Rydes needs your help to implement its vision. In this workshop, you will help Wild Rydes launch the future of customer service. You will build a customer service bot for Facebook that runs on AWS Lambda and Amazon API Gateway. Youll enable the bot to respond intelligently to customers by building in Natural Language Processing (NLP). The bot will use an event-driven architecture in which Lambda functions trigger workflows that pull customer responses from a Knowledge Base of Q Youll also write a function to trigger a manual approval request to a Slack channel, so that Wild Rydes technical staff can approve or reject messages from the bot to the customer. Finally, youll also learn to use Amazon S3, Amazon DynamoDB, and Amazon Elasticsearch Service to log all incoming requests and create live analytical dashboards, such as for sentiment analysis, to track customer satisfaction.
SVR311 - The State of Serverless Computing
Join us to learn about the state of serverless computing from Dr. Tim Wagner, General Manager of AWS Lambda. Dr. Wagner discusses the latest developments from AWS Lambda and the serverless computing ecosystem. He talks about how serverless computing is becoming a core component in how companies build and run their applications and services, and he also discusses how serverless computing will continue to evolve.
SVR401 - Using AWS Lambda to Build Control Systems for Your AWS Infrastructure
Defining infrastructure resource policies in an organized manner can help your company better manage its infrastructure resources. This session familiarizes you with using AWS Lambda to process data and provide control logic for your infrastructure. You can use Amazon CloudWatch Events to monitor infrastructure resources in real-time, and you can use AWS Lambda to react to events based on a set of rules. We demonstrate how you can build a rules engine for creating, monitoring, and managing policies. This is all done using Alexa and Alexa Skills Kit.
SVR402 - Operating Your Production API
In this session, you learn how to monitor and manage your serverless APIs in production. We show you how to set up Amazon CloudWatch alarms, interpret CloudWatch logs for Amazon API Gateway and AWS Lambda, and automate common maintenance and management tasks on your service.
LD108 - Live Demo: Creating an AWS Data Lake on S3
See a real-time demo of AWS technology. Lots of info packed into this 10-minute, quickfire session.
LD109 - Live Demo: Using S3 Transfer Acceleration
See a real-time demo of AWS technology. Lots of info packed into this 10-minute, quickfire session.
LD121 - Live Demo: One Large Storage Pool - Clustering with Snowball Edge
See a real-time demo of AWS technology. Lots of info packed into this 10-minute, quickfire session.
LD132 - Live Demo: Using AWS Lambda from Snowball Edge
See a real-time demo of AWS technology. Lots of info packed into this 10-minute, quickfire session.
STG201 - Storage State of the Union
This session provides a foundational overview of the AWS storage portfolio, including block, file, object, and cloud data migration services. This session will touch on the significant new offerings, outline some of the most common use cases and prepare you for the individual deep dive sessions, customer sessions and new announcements.
STG202 - Deep Dive on Amazon Elastic File System
In this session, we fill you in about Amazon EFS, including an overview of this recently introduced service, its use cases, and best practices for working with it.
STG203 - Earth on AWS—Next-Generation Open Data Platforms
Making earth observation data available by using Amazon S3 is accelerating scientific discovery and enabling the creation of new products. Attend and learn how the scale and performance of Amazon S3 lets earth scientists, researchers, startups, and GIS professionals gather and analyze planetary-scale data without worrying about limitations of bandwidth, storage, memory, or processing power. Learn how AWS is being used to combine satellite imagery, social data, and telemetry data to produce new products and services. Learn also how Amazon S3 provides much more than storage, and how an open geospatial data lake on Amazon S3 can be used as the basis for planetary-scale applications built with Amazon EMR, Amazon API Gateway, and AWS Lambda. As part of this talk, AWS customer Digital Globe demonstrates how they use open data stored in S3 to distribute high-resolution satellite imagery to their customers around the world.
STG204 - Case Study: How Startups like Mapbox, Ring, Hudl, and Others Use AWS to Punch above Their Weight
Join us for this lightning-round showcase of hot new brands and startup companies that are using AWS to play a really big game. You'll hear from experts like Mapbox CIO Will White, Ring Senior Engineer Jason Gluckman, Hudl Engineering Director Rob Hruska,and many others as they explain how they thought about the problems they faced and how they solved them in this TED-style session packed with lots of creative thinking.
STG206 - Case Study: How Spokeo Improved Web Application Response Times with Amazon EFS
At Spokeo, we are running a fast, big data, and high-traffic website providing people search services. But at our scale, we started to reach limitations to how fast our conventional web stack could do things and concluded that a Ruby on Railsonly solution simply couldn't keep up. In this session, we cover some of the options we had to solve this problem and why we chose Amazon Elastic File System (Amazon EFS) as a central part of our solution with metrics and benchmarking. Using EFS, we were able to take response times down from 250 ms to below 70 ms. We look into the architecture of the solution and lessons we learned along the way. In the end, we find that faster response times are just the beginning of the benefits that we see.
STG207 - Case Study: How Atlassian Uses Amazon EFS with JIRA to Cut Costs and Accelerate Performance
At Atlassian, we create popular software tools to help every team unleash their full potential. We use our issue tracking tool JIRA to handle customer support issues from around the globe. As the business grew, we decided to move from a single server instance of JIRA to our JIRA Data Center product on AWS infrastructure to increase reliability, scale, and security. JIRA Data Center's requirement to use a shared file system caused us to try several in-house and third-party solutions prior to the availability of Amazon Elastic File System (Amazon EFS). We chose EFS for its performance, ease of use, automatic scaling, and out-of-the-box distribution capabilities. In this session, we discuss how Atlassian uses EFS to run JIRA Data Center. Topics include the technical architecture, how and why we chose EFS, our recommendations, and the lessons learned along the way.
STG208 - Case Study: How Monsanto Uses Amazon EFS with Their Large-Scale Geospatial Data Sets
At Monsanto, we build and use technologies that support our data and also BI efforts that facilitate intelligent, data-driven decisions. In the past year, we've embarked on large-scale efforts to modernize our geospatial platform and improve our analytic processing capabilities by building out new cloud and open-source based services. We found using Amazon Elastic File System (Amazon EFS) gave us the flexibility and performance we were seeking while saving us significant time, effort, and cost. In this session, we discuss how Monsanto uses the Amazon EFS service to run our large scaling geospatial data sets such as raster, and to perform highly parallelized analytics for data scientists and business users. Topics include the technical architecture, how and why we chose EFS for handling data sets that are terabytes in size, our recommendations, and the lessons learned along the way.
STG209 - Strategic Planning for Long-Term Data Archiving with Amazon Glacier
Without careful planning, data management can quickly turn complex with a runaway cost structure. Enterprise customers are turning to the cloud to solve long-term data archive needs such as reliability, compliance, and agility while optimizing the overall cost. Come to this session and hear how AWS customers are using Amazon Glacier to simplify their archiving strategy. Learn how customers architect their cloud archiving applications and share integration to streamline their organization's data management and establish successful IT best practices.
STG210 - Simplified Data Center Migration—Lessons Learned by Live Nation
As the global leader of live entertainment, Live Nation promotes and produces over 22,000 events annually, operates out of 37 countries, and cultivates over 530 million fans globally. To focus on the growth of the business and shed increasing infrastructure costs, the company made the strategic decision to get out of the data center business and go all in with the cloud. Using instrumental services like AWS Import/Export Snowball, VM Import/Export, AWS CloudFormation and AWS Identity and Access Management, VP Cloud Services Jake Burns quickly and efficiently migrated priority business and operational applications, allowing for immediate cost efficiencies. Learn how AWS offerings like Snowball played a decisive role in Live Nation's ability to easily migrate data and enable end users to quickly access applications to minimize operational impact.
STG211 - Case Study: Data-Heavy Healthcare: UPMCe’s Transformative Approach to Disrupting Healthcare
Today's health care systems generate massive amounts of protected health information (PHI) patient electronic health records, imaging, prescriptions, genomic profiles, insurance records, even data from wearable devices. In this session, UPMCe dives deep into two efforts: Their Data Liberation Projecta next-gen petabyte-scale software solution that provides responsible management of PHI within their own environments as well as externally, and Neutrino a real time medical document aggregator which utilizes natural language processing techniques to unlock hidden value from unstructured narratives. UPMC Enterprises (UPMCe), a division of University of Pittsburgh Medical Center, builds technology and invests in health care companies, from new startups to large established partners, with an eye toward revolutionizing healthcare. They embody the startup mentality with a focus on innovation and creating new data-heavy applicationsall in support of new spin-off companies, furthering economic development, and disrupting healthcare. Join us to learn how they do security management and governance using Amazon S3, Amazon EC2, AWS Config, AWS CloudTrail, and other Amazon services help UPMCe think big about healthcare data in the public sector.
STG212 - Three Customer Viewpoints: Private Equity, Managed Services, and Government – How These Customers Transformed Business Operations through Storage
Customers seeking to make transformations to their environment typically gain efficiencies in both operational and capital costs. This session presents three different case studies, each from a different category of operations Private Equity, Managed Services, and Government. In Government, General Dynamics will discuss how implementing AWS storage as part of their DevOps environment, saving $700k per year. Carlyle Group will show how replacing their on-premises storage with cloud based solutions brought new capabilities to their organization. Finally, Cobalt Iron will discuss how moving their customers to cloud backup targets allowed them to deliver 2M cloud backups per month.
STG213 - NEW SERVICE: Introducing the new AWS Storage Gateway: hybrid cloud storage with file, volume, and tape interfaces
AWS Storage Gateway seamlessly combines existing infrastructure investment with the AWS Cloud for backup, restore, disaster recovery, hybrid workloads, and tiered storage use cases. It has a new file interface that supports moving files in NFS-based file systems into Amazon S3. This session dives into the use cases and advantages of hybrid cloud storage, exploring this newly updated data transfer service with stories from real-life customer deployments.
STG214 - NEW LAUNCH! Introducing AWS Snowball Edge and AWS Snowmobile
At re:Invent 2016 AWS announced several revolutionary ways to move data into and out of the AWS Cloud in large batches, so you can retire servers, racks, and data centers and embark on new projects faster. Two of these new services are AWS Snowmobile and AWS Snowball Edge. Snowmobile cuts cloud data migration projects from years to weeks in time with a 45 foot truck with 100PB of capacity. AWS Snowball Edge is a petabyte-scale data transfer service with portable storage and compute capabilities. It transports up to 100TB of data in a ruggedized, weather resistant, rack mountable, shippable device with the same embedded cryptography and security as the original Snowball. Learn more about each of these services and hear from two customers on the impact these services have made on their data center operations.
STG215 - How Amazon S3 Storage Management Helps Optimize Storage at Scale, with Special Guest, Pinterest
Customers using Amazon S3 at large scale benefit greatly from storage management features. Storage lifecycle policies help them reduce storage costs. Cross-region replication makes it easier to copy data between AWS regions for compliance or disaster recovery. Event notifications allow automatic initiation of processes on objects as they arrive, or capture information about objects and log it for security purposes. In this session, you'll learn about these features, and also several new storage management features in Amazon S3 that give users unmatched visibility into what data they are storing and how that data is being used. These new features make it simpler to analyze usage by users, apps, or organizations, to highlight anomalies, and to optimize business process workflows. They also help identify opportunities to reduce costs, improve performance, and archive infrequently used data. In addition, they can provide insight into who is accessing data stored in S3. As part of this talk, AWS customer Pinterest shows how they have been able to leverage many of the new S3 storage management features to reduce their storage costs significantly by moving a large amount of their data from S3 Standard to S3 Standard Infrequent Access storage.
STG217 - How Aptean uses AWS Marketplace storage solutions to backup AWS production environments leveraging native snapshots
Aptean is a global enterprise software provider that uses AWS as the core of its infrastructure because its a solution that reliably backs up Apteans Amazon EC2 instances. Come to this session to learn what happened when Aptean needed a highly reliable, full backup solution that also allowed for ease of scale, automation, and instant recovery in case of a failure. In this session, youll learn how CPM helps take full advantage of AWS Snapshots, thus adding a management layer to control retention, automate recoveries, and allow live, application-aware backup of both Windows and Linux instances. Aptean will also why they selected N2W Cloud Protection Manager (CPM) in AWS Marketplace as its backup solution extending AWS services and supporting their diverse customer base. By the end of the session, youll have learned the details of how CPM helps Aptean properly manage its snapshots and recoveries, ensuring a resilient deployment that meets Apteans business continuity goals.
STG301 - Deep Dive on Amazon Elastic Block Store
In this popular session, you will learn about the latest features and use cases for Amazon EBS, including best practices, an overview of newly introduced features, and brand-new re:Invent announcements. In particular we will cover the expanded portoflio of volume types, including provisioned IOPS, cold storage, and throughput-optimized. This session will help database admins and application architects understand how to blend performance and cost with applicaitns for big data analytics, data warehousing, and transactional and NoSQL databases.
STG302 - Deep Dive on Amazon Glacier
Not just for archiving or compliance use cases, Amazon Glacier accommodates customers simply looking to replace their on-premises long term storage with a cost efficient, durable, cloud option, from which they can easily and quickly access their data when they need to. This session will introduce newly launched features for Amazon Glacier, review the current service feature set, and share the global data center shut down and storage strategy for Sony DADC New Media Solutions (NMS). NMS is Sonys digital servicing division providing global digital distribution, linear playout and white label OTT/Commerce solutions for clients such as BBC Worldwide, NBCUniversal, Sony Playstation, and Funimation Entertainment. Hear from Andy Shenkler, NMSs Chief Technology and Solutions Officer as he talks about the key factors that drove the organizations decision to move away from tape and go towards the cloud and out of the infrastructure business overall. Learn more about the impact and operational practices inside a world class digital supply chain as they were able to move over 20 petabytes of data, over 1M hours of video, to the cloud and never looked back.
STG303 - Deep Dive on Amazon S3
Come learn about new and existing Amazon S3 features that can help you better protect your data, save on cost, and improve usability, security, and performance. We will cover a wide variety of Amazon S3 features and go into depth on several newer features with configuration and code snippets, so you can apply the learnings on your object storage workloads.
STG305 - Reinventing Disaster Recovery Leveraging AWS Cloud Infrastructure, with special guest, Dow Jones
These days, EVERY workload is considered critical by someone in the organization. As a result, SLAs are shrinking. IT is challenged to meet these SLAs, but there isnt enough budget to provide services like disaster recovery (DR) using traditional methods and infrastructure. The good news is that public cloud platforms, like AWS, are becoming the de facto infrastructure choice for DR. However, workload portability solutions that simplify cross-platform or cloud recovery are required to meet most RTO RPO SLAs in the cloud. AWS provides the infrastructure we need to bring DR to tier 2 and tier 3 workloads that have never been able to afford it before. Now, we need orchestration and automation to make it scalable and reliable. In this session you will learn key considerations and practical steps for getting to the AWS cloud and how you can leverage Amazon S3 storage for cost-effective disaster recovery. Dow Jones will also share details on their migration to AWS Cloud, the benefits realized there, and what the future looks like. Session sponsored by Commvault.
STG306 - Tableau Rules of Engagement in the Cloud
You have billions of events in your fact table, all of it waiting to be visualized. Enter Tableau but wait: how can you ensure scalability and speed with your data in Amazon S3, Spark, Amazon Redshift, or Presto? In this talk, youll hear how Albert Wong and Srikanth Devidi at Netflix use Tableau on top of their big data stack. Albert and Srikanth also show how you can get the most out of a massive dataset using Tableau, and help guide you through the problems you may encounter along the way. Session sponsored by Tableau. AWS Competency Partner
STG307 - Case Study: How Prezi Built and Scales a Cost-Effective, Multipetabyte Data Platform and Storage Infrastructure on Amazon S3
Prezi has over 75 million registered users generating 900 GB of new data each day. They store XML describing over 260 million public presentations that have been viewed over 2 billion times. This created a scale problem that on-premises storage couldn't solve. In this session, you'll learn how Prezi leveraged the power and flexibility of Amazon S3 to turn their storage problem into an analytics opportunity and scale storage to meet the demands of their business without overspending. Prezi provides insight on how a small team of engineers accepted the challenge and succeeded using AWS with managed Hadoop (Amazon EMR) and Amazon Redshift. You'll see how Prezi runs and improves its data platform in a self-service data culture with this small team, without drowning in maintenance and support. Additionally, Prezi shows how the infrastructure team architected their services for storing and serving crucial data, user-generated and otherwise, using S3 and Amazon CloudFrontincluding pitfalls, best practices, cost considerations, and learnings along the way.
STG308 - Case Study: Analytics Without Limits. FINRA’s Scalable Big Data Architecture on S3
FINRA partnered with AWS product teams to leverage Amazon EMR and Amazon S3 extensively to build an advanced analytics solution. In this session, you'll hear how FINRA implemented a data lake on S3 to provide a single source for their big data analytics platform. FINRA ingests 75 billion records each day of stock market transactions, with an AWS storage footprint of 20 petabytes across S3 and Amazon Glacier. To deal with this workload, FINRA has architected a platform that separates storage from compute to manage capacity for each independently, leading to improved performance and cost effectiveness. You'll also learn how FINRA was able to leverage Hbase on Amazon EMR to achieve significant benefits over running Hbase on a fixed capacity cluster. FINRA was able to implement a system that seamlessly scales in response to data growth and can scale quickly in response to user traffic. By working with multiple clusters, FINRA can now isolate ETL and user query workloads and has achieved rapid, built-in disaster recovery capability by leveraging data storage on S3 to run from multiple AZs and across regions.
STG309 - Case Study: How Startups Like Smartsheet and Quantcast Accelerate Innovation and Growth with Amazon S3
Startups around the world use AWS services to access the power of the cloud to grow faster and more cost effectively. In this session, Smartsheet talks about how they were able to cost-effectively build their prototype for scale and avoid replatforming at different points in the adoption curve, and Quantcast discusses how they are running a high-performance analytics solution on AWS. They provide several tips and tricks for S3, and show how they removed a traditional MySQL data store from a distributed-image hosting application so that the only required data store is S3. They also show how to avoid common, cumbersome database practices by working with the eventually consistent nature of S3 objects and the fact that objects and directories share the same namespace.
STG311 - Case Study: How Videology and Zendesk Modernized Their Big Data Platforms on Amazon EBS
The companies Videology and Zendesk both had the same problemhow to rearchitect their big data processing platforms to scale to meet growing demand, while at the same time improving performance, availability, and cost structure? Videology provides a converged advertising solution that is screen-agnostic, ensuring unduplicated reach with the right frequency cadence to achieve guaranteed results. To achieve this, their big data platform ingests, processes, and analyzes a variety of logs. In this talk, Videology discusses how they migrated our platform to use Cloudera on a mix of m4 and r3 instances using the Amazon EBS Streaming Optimized HDD (st1) volume type. Zendesk provides a cloud-based customer support platform that allows quicker and easier interaction between businesses and customers. To deliver this experience, Zendesk runs a large Elasticsearch, Logstash, Kibana stack. The talk also discusses how Zendesk rearchitected their deployment to use m4s and also leverage the EBS Streaming Optimized HDD (st1) volume type. Tips for success will be shared throughout.
STG312 - Workshop: Working with AWS Snowball - Accelerating Data Ingest into the Cloud
This workshop will provide customers the opportunity to work hands-on with the AWS Snowball service, with attendees broken out into small teams to perform various on-prem to cloud data transfer scenarios using actual Snowball devices. These scenarios will include migrating backup archive data to S3-IA and Glacier, HDFS cluster migration to S3 for use with EMR/Redshift, and leveraging the Snowball API SDK to build AWS Snowball service integration into a custom application. The session will open with an overview of the service, objectives, and guidance on where to find resources. Attendees should bring their own laptops and should have a basic familiarity with AWS storage services (S3 and Glacier). Prerequisites: Participants should have an AWS account established and available for use during the workshop. Please bring your own laptop.
STG313 - Workshop: AWS S3 Deep-Dive Hands-On Workshop: Deploying and Managing a Global, Petabyte Scale Storage Infrastructure
In this session, well expand on the S3 re:Invent deep-dive session with a hands-on workshop on advanced S3 features and storage management capabilities. Well have AWS S3 and Glacier experts on-hand to deep-dive on S3 architecture, performance scalability optimization, how to analyze your content and leverage storage tiers (S3 Standard, S3 Standard Infrequent Access, Glacier) to balance cost and SLAs, security considerations, replication with Cross Region Replication (CRR), versioning for data protection and more. In the hands-on lab, well walk through a customer scenario: architecting a high-performance infrastructure for consumer applications. In the scenario, well use sample data sets on S3, analyze object retrieval patterns and design a complete solution using many of the features S3 offers including migrating objects to an appropriate tier. Prerequisites: Participants should have an AWS account established and available for use during the workshop. Please bring your own laptop.
WIN201 - Simplifying Microsoft Architectures with AWS services
Learn how to architect fully available and scalable Microsoft solutions and environments in AWS. Find out how Microsoft solutions can leverage various AWS services to achieve more resiliency, replace unnecessary complexity, simplify architecture, provide scalability, introduce DevOps concepts, automation and repeatability. Plan authentication and authorization, various hybrid scenarios with other cloud environment and on premise solutions/infrastructure. Learn about common architecture patterns for Active Directory and business productivity solutions like SharePoint, Exchange and Skype for Business, also common scenarios for SQL deployments and System Center.
WIN203 - How Pitney Bowes is transforming their business in the cloud
Pitney Bowes is reinventing its business based on a SaaS and a cloud-based model to deliver services for their clients globally centered on the Pitney Bowes Commerce Cloud. The Pitney Bowes Commerce Cloud is a commerce enabler, providing access to solutions, analytics and APIs across the full commerce continuum with speed and agility to help clients identify customers, locate opportunities, enable communications, power shipping from anywhere to everywhere, and manage payments. During this session, the Pitney Bowes team will discuss the strategy behind the Commerce Cloud, how we accelerated the speed of innovation and creation of new product solutions by taking full advantage of the AWS platform, from Cloud Infrastructure Services (Compute, Big Data, Storage and Content Delivery, Databases and Networking) , to Analytics and Internet of Things Services. Pitney Bowes applications are deployed in Docker containers using AWS Elastic Beanstalk, and are utilizing S3, Amazon RDS, SQS, SNS, and ElastiCache. CloudWatch and CloudFormation are being used to manage the solutions that Pitney Bowes brings to the market. Additionally, the Pitney Bowes Data and Analytics platform is powered by Elastic Map Reduce (EMR), Spark, Aurora, DynamoDB and Postgres. The session will also discuss Pitney Bowes partnership with AWS to deploy Pitney Bowes APIs for Location Intelligence via the AWS Marketplace, allowing developers, customers and partners to access innovative location intelligence data services in an easy to consume and highly reliable manner.
WIN204 - How to move 1,000 VMs and Biz Critical Apps to AWS in 6 months. Methodology and lessons learned from Edwards Lifesciences
Edwards Lifesciences partners with clinicians to develop innovative technologies in the areas of structural heart disease and critical care monitoring to help patients live longer, healthier and more productive lives. The Edwards Lifesciences IT team recently embarked on a journey to move many of their business critical corporate IT applications to AWS to better support the Edwards Lifesciences mission to help patients worldwide. In this session, the Edwards Lifesciences team would like to propose a different approach for Corporate IT teams to consider with respect to the cloud that can help fulfill key business outcomes while transforming IT infrastructure quickly and safely! An approach that starts with business critical applications to take full advantage of the scale and growth that were key requirements for the team. It worked for us and created a new vibe in our IT organization which we are currently enjoying.
WIN205 - NEW LAUNCH! Amazon EC2 Systems Manager for Hybrid Cloud Management at Scale
Today, we are announcing EC2 Systems Manager. Amazon EC2 Systems Manager is a management service that helps you automatically collect software inventory, apply OS patches, create system images, and configure Windows and Linux operating systems. These capabilities help you define and track system configurations, prevent drift, and maintain software compliance of your EC2 and on-premises configurations. This session provides an overview of these newly announced services and how they work together within the larger AWS ecosystem to provide comprehensive management capabilities.
WIN301 - Bring Microsoft Applications to AWS to Save Money and Stay License Compliant using PowerShell, Windows KMS, and Dedicated Hosts
Running Microsoft workloads on AWS is easy and can save you money. This session will cover how to bring your own Microsoft licenses to AWS, and then demonstrate using PowerShell to import your Windows Server image from Vmware or Hyper-V, configure Windows KMS with your license key, and launch an EC2Dedicated Host. We will discuss ways you can use AWS Config rules to manage license compliance.
WIN303 - How to Launch a 100K-User Corporate Back Office with Microsoft Servers and AWS
Learn how to build a scalable, compliance-ready, and automated deployment of the Microsoft backoffice servers for 100K users running on AWS. In this session, we show a reference architecture deployment of Exchange, SharePoint, Skype for Business, SQL Server and Active Directory in a single VPC. We discuss the following: (1) how the solution is automated for 100K users, (2) how the solution is enabled for compliance (e.g., FedRAMP, HIPAA, PCI), and (3) how the solution is built from modular 10K user blocks. Attendees should have knowledge of AWS CloudFormation, PowerShell, instance bootstrapping, VPCs, and Amazon Route53, as well as the relevant Microsoft technologies.
WIN304 - Design, Deploy, and Optimize Microsoft SharePoint on AWS
AWS can help you rapidly deploy and scale your Microsoft SharePoint environment to help you collaborate more efficiently and cost-effectively. This session reviews architectural considerations for building a SharePoint deployment on AWS, best practices to ensure optimal performance, how to leverage multiple Availability Zones for high availability and disaster recovery, and how to integrate with Microsoft Active Directory. We will also look at new Quick Start guides, AWS CloudFormation templates, and other tools that dramatically reduce the time to deployment.
WIN305 - Best Practices for Integrating Active Directory with AWS Workloads
Active Directory (AD) is essential for Windows workloads in the cloud. AWS offers customers multiple ways to integrate AD with cloud workloads like EC2, RDS, and AWS Enterprise Applications: AWS Directory Service for Microsoft Active Directory (Enterprise Edition) as a managed service and Active Directory running on AWS EC2 Windows instances. Which option is right for you? This session will discuss the key deployment considerations for each option to help you identify which best meets your project goals, and the effort involved. The session will cover options for integrating with your on-premises directory, port and security considerations, application considerations, and best practices.
WIN306 - Design, Deploy, and Optimize Microsoft SQL Server on AWS
This session goes deep on best practices and considerations for running Microsoft SQL Server on AWS. We cover best practices for deploying SQL Server, how to choose between Amazon EC2 and Amazon RDS, and ways to optimize the performance of your SQL Server deployment for different application types. We review in detail how to provision and monitor your SQL Server databases, and how to manage scalability, performance, availability, security, and backup and recovery in both Amazon RDS and Amazon EC2. In addition, we cover how you can set up a disaster recovery solution between an on-premises SQL Server environment and AWS, using native SQL Server features like log shipping, replication, and AlwaysOn Availability Groups.
WIN307 - Searching Inside Video at Petabyte Scale Using Spot
Video is a last-mile problem for search technology. Unlike webpages, documents, and email, content in videos has traditionally been impossible to search. Recent advances in automated speech and text recognition, however, let businesses and universities search inside video assets as easily as inside textual content. In this session, you'll learn how Panopto is using AWS to solve the video-search problem at scale, while saving over 50% in operating costs by taking advantage of Spot instances. We discuss the cross-platform architecture that combines Windows and Linux to provide cost-effective video processing and search indexing. We also dive deep into scaling Spot elastically based on user demand, handling fallback situations when instances are revoked, and using the Spot bidding process to optimize cost structure. Finally, we discuss future plans to reduce operating costs even further through Spot fleets and grid processing.
WIN401 - How to Manage Inventory, Patching, and System Images for Your Hybrid Cloud with AWS Management Capabilities
Whether you are a traditional enterprise making a shift to the cloud or are already all-in, AWS EC2 management capabilities enable you to perform common tasks of inventory collection, and patch and image management of your infrastructure at scale. In this session, we'll provide a brief overview of these powerful capabilities and bring it all together with a demo of real-world hybrid-cloud management scenario.
WIN402 - How I learned to embrace DevOps and Configure Infrastructure at Scale
In this session, you will learn how you can use EC2 management capabilities to perform repeatable automation of your infrastructure at scale, across platforms and hybrid environments. As you embark on your cloud journey and embrace a modern DevOps mindset, you not only want to deploy software quickly but also ensure configuration consistency at cloud scale. Many enterprises have successfully used services such as EC2 Run Command to perform administrative tasks, and we'll share some of those stories. In addition, we will demo new capabilities to ensure a desired state approach to software configuration either through predefined or easy to build custom configurations, and also how you can integrate with other AWS services to deliver enterprise IT and business value.
WIN403 - How to Migrate Microsoft Windows Applications to AWS Quickly, with Less Risk, Using Multisite Replication and SQL HA
Migrating on-premises workloads to AWS is a top priority for customers today. AWS has an array of tools and capabilities to make the move faster and less risky. This deep-dive session presents how to build a migration framework for your Windows-based applications by using AWS CloudFormation templates, along with multisite replication capabilities built into AWS and optimized for Microsoft Windows workloads built on SQL HA. This framework helps speed up your data center migration and minimize risk during the process.
GA101 - AWS Worldwide Government, Education, and Nonprofit Breakfast
Our AWS Worldwide Public Sector Breakfast on Wednesday, November 30, from 6:30 to 7:30 a.m. will feature Teresa Carlson, vice president of AWS Worldwide Public Sector, to share important public sector updates alongside Defense Digital Services, American Heart Association, and Ellucian. These public sector leaders at the forefront of innovation each take the stage to discuss the impact they are making with the cloud in the government, education, and the nonprofit communities. Changing the world is no easy task. From tackling heart disease to running national security systems to removing barriers to achieve student successall these missions have something in common: the cloud is propelling them to results no one thought possible. Join us for this hour-long breakfast dedicated to the unique challenges, needs, and successes of the public sector around the world.
GA201 - Diversify Tech: Building the Future IT Workforce
In this session, learn about efforts to diversify IT, from programs in high schools all the way up to diversity among top-level executives. Regardless of where you are in your career, this session dives deep to speak to you about the changing tech landscape. Through successes and failures, speakers candidly discuss how everyone can help in changing the ratio to attract a more diverse workforce and create a culture of inclusion. Bring your questions, network with peers, and be inspired by both women and men making an impact in non-profit, enterprise and government sectors.
WWPS301 - AWS GovCloud (US) for Highly Regulated Workloads
Learn how to architect for compliance in the AWS cloud and see how your organization can leverage the agility, cost savings, scalability, and flexibility of the cloud while meeting the most stringent regulatory and compliance requirements, including Federal Risk and Authorization Management Program (FedRAMP), ITAR, CJIS, HIPAA, and DoD Cloud Computing Security Requirements Guide (SRG) Levels 2 and 4. Hear best practices and practical use cases for using AWS GovCloud (US) to comply with a variety of regulatory regimes.
WWPS302 - Governance Strategies for Cloud Transformation
Cloud governance is critical to executing a systemic, supportable, and sustainable cloud transformation strategy. This session explores best practices to achieve that model, including standards policies, automation that consistently applies and enforces policies and controls, self-service capabilities that enable agility and speed, and automated monitoring and cost management that ensure operational integrity. A well-developed cloud governance model enables customers to improve operational integrity, reliability, performance, and transparency using the AWS cloud. This session guides you through a best practice governance model, including policy considerations and recommendations, self-service automation methods towards IT-as-a-Service, and use case examples.
WWPS303 - Modernizing Government in the Cloud in Highly Regulated Environments
Cloud computing can help government organizations increase innovation, efficiency, agility, and resiliencyall while reducing costs. This session highlights perspectives on how small, highly empowered United States Digital Service (USDS) teams within the Department of Defense (DoD), Department of Homeland Security (DHS), and Department of Veterans Affairs (VA) are breaking down innovation barriers, tackling mission-critical operations, and delivering more value with the cloud in highly regulated, unclassified environments.
WWPS304 - Using AWS to Meet Requirements for Education, Healthcare and Public Safety (HIPAA, FERPA, and CJIS)
With rich controls, auditing, and broad security accreditations, AWS enables its customers to be in compliance with CJIS, FERPA, and HIPAA. Come hear customers share their approaches to achieving compliance for these and other standards across many markets.
WWPS401 - Data Polygamy: The Many-Many Relationships among Urban Spatio-Temporal Datasets
In this session, learn how Data Polygamy, a scalable topology-based framework, can enable users to query for statistically significant relationships between spatio-temporal datasets. With the increasing ability to collect data from urban environments and a push toward openness by governments, we can analyze numerous spatio-temporal datasets covering diverse aspects of a city. Urban data captures the behavior of the citys citizens, existing infrastructure (physical and policies), and environment over space and time. Discovering relationships between these datasets can produce new insights by enabling domain experts to not only test but also generate hypotheses. However, discovery is difficult. A relationship between two datasets can occur only at locations or time periods that behave differently compared to the regions neighborhood. The size and number of datasets and diverse spatial and temporal scales at which the data is available presents computational challenges. Finally, of several thousand possible relationships, only a small fraction is actually informative. We have implemented the framework on Amazon EMR and show through an experimental evaluation using over 300 spatial-temporal urban datasets how our approach is scalable and effective at identifying significant relationships. Find details about the work at http://dl.acm.org/citation.cfm?id=2915245. The code and experiments are available at https://github.com/ViDA-NYU/data-polygamy.


Inspired by Rodney Haywood's index in 2012, I decided to do the same for 2013 (in Groovy) and 2014/205 (in Scala). This year I decided to recode it in Go. I borrowed from his Rodney's formatting. The code is this github project which a mix of chrome dev tools web scraping, google data V3 API (YouTube), Sean Watsons SlideShare golang lib (requires one fix I intend to upstream), Chris Broglie's mustache golang impl and Moovweb's "tagsoup" and XPath impl. The code took two evenings of hacking (mostly slowed down by the fact that Slideshare rate limits so aggressively). If you see any missing information feel free to issue a pull request to fix it. The Go impl was pretty straight forward with the Youtube and Slideshare libriaries. I felt like I was writing for loops over ranges way to much (as compared to Scale) - I miss the functional data operators. Given I've only used Go in anger a bit, if you see stupid problems in my Go coding please let me know.

40 comments:

  1. If you are looking for a solid contextual ad company, I recommend you check out PropellerAds.

    ReplyDelete
  2. Get daily suggestions and guides for earning THOUSANDS OF DOLLARS per day FROM HOME totally FREE.
    SUBSCRIBE NOW

    ReplyDelete
  3. Did you know that you can shorten your long links with Shortest and receive dollars for every click on your shortened links.

    ReplyDelete
  4. Do you love Pepsi or Coca-Cola?
    PARTICIPATE IN THE POLL and you could get a prepaid VISA gift card!

    ReplyDelete
  5. DreamHost is ultimately the best web-hosting provider for any hosting plans you require.

    ReplyDelete
  6. You might qualify for a new solar energy program.
    Click here to find out if you qualify now!

    ReplyDelete
  7. I'm using AVG protection for a couple of years, I would recommend this solution to everyone.

    ReplyDelete
  8. There is a chance you are eligible to get a free $1,000 Amazon Gift Card.

    ReplyDelete
  9. Searching for the Best Dating Website? Join to find your perfect match.

    ReplyDelete
  10. Cloud computing is at an early stage, with a motley crew of providers large and small delivering a slew of cloud-based services, from full-blown applications to storage services to spam filtering.

    Cloud computing architectures

    ReplyDelete
  11. Thank you for taking the time to provide us with your valuable information. We strive to provide our candidates with excellent care.As always, we appreciate you confidence and trust in us.

    Recruitment Consultancy in Bangalore

    ReplyDelete
  12. Really nice and definitely it will be useful for many people. Kindly keep update like this.

    Digital Marketing Company in Chennai

    ReplyDelete

  13. These information really worth saying, i think you are master of the content and thank you so much sharing that valuable information and get new skills after refer that post.


    Tooth Braces In Chennai

    ReplyDelete
  14. Superb. I really enjoyed very much with this article here. Really it is an amazing article I had ever read. I hope it will help a lot for all. Thank you so much for this amazing posts and please keep update like this excellent article.thank you for sharing such a greatblog with us. expecting for your..
    SEO Company in India
    Digital Marketing Company in India

    ReplyDelete
  15. Superb. I really enjoyed very much with this article here. Really it is an amazing article I had ever read. I hope it will help a lot
    for all. Thank you so much for this amazing posts and please keep update like this excellent article.thank you for sharing such a great blog with us. expecting for your..
    SEO Company in India
    Digital Marketing Company in India

    ReplyDelete
  16. This is really a very great blog. The information present in this very useful for us. thank you for sharing with us. Thanks for reading my comment.
    Hadoop Training in Chennai

    ReplyDelete
  17. This is an awesome post.Really very informative and creative contents. These concept is a good way to enhance the knowledge.I like it and help me to development very well.
    Android training in chennai

    ReplyDelete
  18. Finding the time and actual effort to create a superb article like this is great thing. I’ll learn many new stuff right here! Good luck for the next post buddy..
    SEO Company in Chennai

    ReplyDelete
  19. Hi I read your post very carefully and I think you are right that a well written post should be at least a 100 words and should capture the essence of your blog, book or article.

    Digital Marketing Company in India

    ReplyDelete
  20. It's interesting that many of the bloggers your tips helped to clarify a few things for me as well as giving.. very specific nice content. And tell people specific ways to live their lives.Sometimes you just have to yell at people and give them a good shake to get your point across.

    Hr Consultancy in Chennai

    ReplyDelete
  21. The way you have stated everything above is quite awesome. Keep blogging like this. Thanks a lot.

    Web Development Company in India

    ReplyDelete
  22. Great suggestions. Your blog is very helpful and your way of teaching is awesome.


    Digital Marketing Company in Chennai

    ReplyDelete
  23. Social Media icons are very important for a website .Mostly In two ways one is seo and other is design. your top followers instagram
    GMAT Coaching Institutes in Chennai

    ReplyDelete
  24. Great Article… I love to read your articles because your writing style is too good, its is very very helpful for all of us and I never get bored while reading your article because, they are becomes a more and more interesting from the starting lines until the end.
    Flats Cleaning in Chennai

    ReplyDelete
  25. This is an awesome post.Really very informative and creative contents. These concept is a good way to enhance the knowledge.I like it and help me to development very well.Thank you for this brief explanation and very nice information.Well, got a good knowledge.
    Isoft Innovations Facebook

    ReplyDelete
  26. Great Article… I love to read your articles because your writing style is too good, its is very very helpful for all of us and I never get bored while reading your article because, they are becomes a more and more interesting from the starting lines until the end.

    Dotnet Training in Chennai

    ReplyDelete
  27. Thank you for sharing such a nice and interesting blog with us. I have seen that all will say the same thing repeatedly. But in your blog, I had a chance to get some useful and unique information. I would like to suggest your blog in my dude circle.
    Isoft Innovations Company Address
    Isoft Innovations Adyar
    Isoft Innovations Reviews
    Isoft Innovation Chennai
    Isoft Innovation

    ReplyDelete
  28. This is a great article, I have been always to read something with specific tips! I will have to work on the time for scheduling my learning.
    Smile Designing Dental Clinic In Chennai

    ReplyDelete
  29. Thank you for sharing such a nice and interesting blog with us. I have seen that all will say the same thing repeatedly. But in your blog, I had a chance to get some useful and unique information. I would like to suggest your blog in my dude circle.
    Best Dental Clinic In Vellore

    ReplyDelete
  30. Thank you for sharing such a nice and interesting blog with us. I have seen that all will say the same thing repeatedly. But in your blog, I had a chance to get some useful and unique information. I would like to suggest your blog in my dude circle.
    Jobs in Chennai
    Jobs in Bangalore
    Jobs in Delhi
    Jobs in Hyderabad
    Jobs in Kolkata
    Jobs in Mumbai
    Jobs in Noida
    Jobs in Pune

    ReplyDelete
  31. This is an awesome post.Really very informative and creative contents. These concept is a good way to enhance the knowledge.I like it and help me to development very well.Thank you for this brief explanation and very nice information.Well, got a good knowledge.
    Fresher Jobs
    Fresher Openings

    ReplyDelete
  32. This comment has been removed by the author.

    ReplyDelete
  33. Pretty article! I found some useful information in your blog, it was awesome to read, thanks for sharing this great content to my vision, keep sharing..
    Home Automation in Chennai
    smart home in Chennai
    Home security in Chennai
    Burglar alarm in Chennai
    Door sensors Chennai

    ReplyDelete
  34. Finding the time and actual effort to create a superb article like this is great thing. I’ll learn many new stuff right here!
    Good luck for the next post buddy..
    Staffing Companies in Bangalore
    HR franchise in Bangalore

    ReplyDelete
  35. We found some useful information in your blog, it is awesome to good materials, thanks for sharing this great content on sap hana online training in USA, UK, CANADA,India.

    SAP HANA Training Hyderabad
    SAP HANA Online Training in Hyderabad

    ReplyDelete
  36. It's Really too Useful Blogs More Ideas Sharing There.Share That's More New Blogs Technologies.Hadoop Training We'll Waiting There New Technologies and Then Useful Information.Keep It Well.Hadoop Online Training

    ReplyDelete
  37. Thanks for sharing this blog. I really appreciate your efforts for such a useful post.

    sap hana online training in hyderabad
    sap hana online course

    ReplyDelete
  38. Our VPS Power servers are the ultimate in performance, security, and stability. We will make sure that you’re chosen E-Commerce such as Magneto run smoothly on our servers.

    Best Australian Based Web Hosting

    ReplyDelete
  39. Good Post! Thank you so much for sharing this pretty post, it was so good to read and useful to improve my knowledge as updated one, keep blogging…
    Seo Training in Chennai
    Best Seo Training
    SEO Company in India

    ReplyDelete