1. 程式人生 > >Addressing Data Residency with AWS

Addressing Data Residency with AWS

Whitepaper image

AWS has released a new whitepaper that has been requested by many AWS customers: AWS Policy Perspectives: Data Residency. Data residency is the requirement that all customer content processed and stored in an IT system must remain within a specific country’s borders, and it is one of the foremost concerns of governments that want to use commercial cloud services. General cybersecurity concerns and concerns about government requests for data have contributed to a continued focus on keeping data within countries’ borders. In fact, some governments have determined that

mandating data residency provides an extra layer of security.

This approach, however, is counterproductive to the data protection objectives and the IT modernization and global economic growth goals that many governments have set as milestones. This new whitepaper addresses the real and perceived security risks expressed by governments when they demand in-country data residency by identifying the most likely and prevalent IT vulnerabilities and security risks, explaining the native security embedded in cloud services, and highlighting the roles and responsibilities of cloud service providers (CSPs), governments, and customers in protecting data.

Large-scale, multinational CSPs, often called hyperscale CSPs, represent a transformational disruption in technology because of how they support their customers with high degrees of efficiency, agility, and innovation as part of world-class security offerings. The whitepaper explains how hyperscale CSPs, such as AWS, that might be located out of country provide their customers the ability to achieve high levels of data protection through safeguards on their own platform and with turnkey tooling for their customers. They do this while at the same time preserving nation-state regulatory sovereignty.

The whitepaper also considers the commercial, public-sector, and economic effects of data residency policies and offers considerations for governments to evaluate before enforcing requirements that can unintentionally limit public-sector digital transformation goals, in turn possibly leading to increased cybersecurity risk.

AWS continues to engage with governments around the world to hear and address their top-of-mind security concerns. We take seriously our commitment to advocate for our customers’ interests and enforce security from “ground zero.” This means that when customers use AWS, they can have the confidence that their data is protected with a level of assurance that meets, if not exceeds, their needs, regardless of where the data resides.

– Min Hyun, Cloud Security Policy Strategist

相關推薦

Addressing Data Residency with AWS

AWS has released a new whitepaper that has been requested by many AWS customers: AWS Policy Perspectives: Data Residency. Data residency

Amazon Kinesis Firehose Data Transformation with AWS Lambda

Shiva Narayanaswamy, Solution Architect Amazon Kinesis Firehose is a fully managed service for delivering real-time streaming data to des

Quickly build, test, and deploy your data lake with AWS and partner solutions

Performing data science workloads on data from disparate sources – data lake, data warehouse, streaming, and more – creates challenges f

Restrict access to your AWS Glue Data Catalog with resource

A data lake provides a centralized repository that you can use to store all your structured and unstructured data at any scale. A data lake can in

Data Lake on AWS with Talend

An out-of-the-box open data lake solution with AWS and Talend allows you to build, manage, and govern your cloud data lake in the AWS Cloud so tha

Predictive Data Science with Amazon SageMaker and a Data Lake on AWS

This Quick Start builds a data lake environment for building, training, and deploying machine learning (ML) models with Amazon SageMaker on the Am

Run federated queries to an AWS data lake with SAP HANA

Harpreet Singh is a Solution Architect at Amazon Web Services (AWS). An Aberdeen survey revealed that organizations who implemented a data

Amazon Redshift Data Warehouse with Matillion ETL on AWS

AWS offers a common architectural structure that enables you to leverage new and existing big data technologies and data warehouse methods. Throug

Store, Protect, Optimize Your Healthcare Data with AWS: Part 1

This blog post was co-authored by Ujjwal Ratan, a senior AI/ML solutions architect on the global life sciences team. Healthcare data is ge

Collect, Structure, and Search Industrial IoT data with AWS IoT SiteWise

AWS IoT SiteWise is a managed service that makes it easy to collect and organize data from industrial equipment at scale. You can easily monitor

Data Analysis with Python : Exercise- Titantic Survivor Analysis | packtpub.com

.com pub nal kaggle out conda anti vivo python kaggle-titantic, from: https://www.youtube.com/watch?v=siEPqQsPLKA install matplotlib: con

SDP(0):Streaming-Data-Processor - Data Processing with Akka-Stream

數據庫管理 新的 集成 部分 ont lock 感覺 sharding 數據源 再有兩天就進入2018了,想想還是要準備一下明年的工作方向。回想當初開始學習函數式編程時的主要目的是想設計一套標準API給那些習慣了OOP方式開發商業應用軟件的程序員們,使他們能用一種接近

Building Data Models with PowerPivot_進階篇2

5.1 使用 Userelationship 建立兩表之間的多個關係 USERELATIONSHIP(多端,一端) Measure_送貨數量 = CALCULATE(SUM([數量])),USERELATIONSHIP('銷售記錄'[實際送貨日期],'日曆年'[日期]) 5.2

Building Data Models with PowerPivot_進階篇

Building Data Models with PowerPivot_進階篇 2.3 使用連結回標進行RFM分析 R Recent近度 MIN([近度]); [近度]=TODAY()-[下單日期] 3.1 使用高階DAX函式 高階聚合函式SUMX SUMX函式

Change the default MySQL data directory with SELinux enabled

轉載:https://rmohan.com/?p=4605   Change the default MySQL data directory with SELinux enabled This is a short article that explains how you

DataCamp Data Scientist with Python track 學習筆記

Importing Data in Python:  Customizing your pandas import:  # Import matplotlib.pyplot as plt import matplotlib.pyplot as plt #

Modern Data Lake with Minio : Part 1

轉自:https://blog.minio.io/modern-data-lake-with-minio-part-1-716a49499533 Modern data lakes are now built on cloud storage, helping organizations lever

Modern Data Lake with Minio : Part 2

轉自: https://blog.minio.io/modern-data-lake-with-minio-part-2-f24fb5f82424 In the first part of this series, we saw why object storage systems like Min

Planar data classification with one hidden layer

From Logistic Regression with a Neural Network mindset, we achieved the Neural Network which use Logistic Regression to resolve the linear class

Chapter 6: Dimensionality Reduction: Squashing the Data Pancake with PCA

Suggestion it is best not to apply PCA to raw countss (word counts, music play counts, movie viewing counts, etc.)。 The reason for this is that such counts