Amazon EC2 FAQs
Q: What are Accelerated Computing instances?
Accelerated Computing instance family is a family of instances which use hardware accelerators, or co-processors, to perform some functions, such as floating-point number calculation and graphics processing, more efficiently than is possible in software running on CPUs. Amazon EC2 provides three types of Accelerated Computing instances – GPU compute instances for general-purpose computing, GPU graphics instances for graphics intensive applications, and FPGA programmable hardware compute instances for advanced scientific workloads.
Q. When should I use GPU Graphics and Compute instances?
GPU instances work best for applications with massive parallelism such as workloads using thousands of threads. Graphics processing is an example with huge computational requirements, where each of the tasks is relatively small, the set of operations performed form a pipeline, and the throughput of this pipeline is more important than the latency of the individual operations. To be able build applications that exploit this level of parallelism, one needs GPU device specific knowledge by understanding how to program against various graphics APIs (DirectX, OpenGL) or GPU compute programming models (CUDA, OpenCL).
Q: How are P3 instances different from G3 instances?
P3 instances are the next-generation of EC2 general-purpose GPU computing instances, powered by up to 8 of the latest-generation NVIDIA Tesla V100 GPUs. These new instances significantly improve performance and scalability, and add many new features, including new Streaming Multiprocessor (SM) architecture for machine learning (ML)/deep learning (DL) performance optimization, second-generation NVIDIA NVLink high-speed GPU interconnect, and highly tuned HBM2 memory for higher-efficiency.
G3 instances use NVIDIA Tesla M60 GPUs and provide a high-performance platform for graphics applications using DirectX or OpenGL. NVIDIA Tesla M60 GPUs support NVIDIA GRID Virtual Workstation features, and H.265 (HEVC) hardware encoding. Each M60 GPU in G3 instances supports 4 monitors with resolutions up to 4096x2160, and is licensed to use NVIDIA GRID Virtual Workstation for one Concurrent Connected User. Example applications of G3 instances include 3D visualizations, graphics-intensive remote workstation, 3D rendering, application streaming, video encoding, and other server-side graphics workloads.
Q: What are the benefits of NVIDIA Volta GV100 GPUs?
The new NVIDIA Tesla V100 accelerator incorporates the powerful new Volta GV100 GPU. GV100 not only builds upon the advances of its predecessor, the Pascal GP100 GPU, it significantly improves performance and scalability, and adds many new features that improve programmability. These advances will supercharge HPC, data center, supercomputer, and deep learning systems and applications.
Q: Who will benefit from P3 instances?
P3 instances with their high computational performance will benefit users in artificial intelligence (AI), machine learning (ML), deep learning (DL) and high performance computing (HPC) applications. Users includes data scientists, data architects, data analysts, scientific researchers, ML engineers, IT managers and software developers. Key industries include transportation, energy/oil & gas, financial services (banking, insurance), healthcare, pharmaceutical, sciences, IT, retail, manufacturing, high-tech, transportation, government, academia, among many others.
Q: What are some key use cases of P3 instances?
P3 instance use GPUs to accelerate numerous deep learning systems and applications including autonomous vehicle platforms, speech, image, and text recognition systems, intelligent video analytics, molecular simulations, drug discovery, disease diagnosis, weather forecasting, big data analytics, financial modeling, robotics, factory automation, real-time language translation, online search optimizations, and personalized user recommendations, to name just a few.
Q: Why should customers use GPU-powered Amazon P3 instances for AI/ML and HPC?
GPU-based compute instances provide greater throughput and performance because they are designed for massively parallel processing using thousands of specialized cores per GPU, versus CPUs offering sequential processing with a few cores. In addition, developers have built hundreds of GPU-optimized scientific HPC applications such as quantum chemistry, molecular dynamics, meteorology, among many others. Research indicates that over 70% of the most popular HPC applications provide built-in support for GPUs.
Q: Will P3 instances support EC2 Classic networking and Amazon VPC?
P3 instances will support VPC only.
Q. How are G3 instances different from P2 instances?
G3 instances use NVIDIA Tesla M60 GPUs and provide a high-performance platform for graphics applications using DirectX or OpenGL. NVIDIA Tesla M60 GPUs support NVIDIA GRID Virtual Workstation features, and H.265 (HEVC) hardware encoding. Each M60 GPU in G3 instances supports 4 monitors with resolutions up to 4096x2160, and is licensed to use NVIDIA GRID Virtual Workstation for one Concurrent Connected User. Example applications of G3 instances include 3D visualizations, graphics-intensive remote workstation, 3D rendering, application streaming, video encoding, and other server-side graphics workloads.
P2 instances use NVIDIA Tesla K80 GPUs and are designed for general purpose GPU computing using the CUDA or OpenCL programming models. P2 instances provide customers with high bandwidth 25 Gbps networking, powerful single and double precision floating-point capabilities, and error-correcting code (ECC) memory, making them ideal for deep learning, high performance databases, computational fluid dynamics, computational finance, seismic analysis, molecular modeling, genomics, rendering, and other server-side GPU compute workloads.
Q: How are P3 instances different from G2 instances?
P3 Instances are the next-generation of EC2 general-purpose GPU computing instances, powered by up to 8 of the latest-generation NVIDIA Volta GV100 GPUs. These new instances significantly improve performance and scalability and add many new features, including new Streaming Multiprocessor (SM) architecture, optimized for machine learning (ML)/deep learning (DL) performance, second-generation NVIDIA NVLink high-speed GPU interconnect, and highly tuned HBM2 memory for higher-efficiency.
P2 instances use NVIDIA Tesla K80 GPUs and are designed for general purpose GPU computing using the CUDA or OpenCL programming models. P2 instances provide customers with high bandwidth 25 Gbps networking, powerful single and double precision floating-point capabilities, and error-correcting code (ECC) memory.
Q. What APIs and programming models are supported by GPU Graphics and Compute instances?
P3 instances support CUDA 9 and OpenCL, P2 instances support CUDA 8 and OpenCL 1.2 and G3 instances support DirectX 12, OpenGL 4.5, CUDA 8, and OpenCL 1.2.
Q. Where do I get NVIDIA drivers for P3 and G3 instances?
There are two methods by which NVIDIA drivers may be obtained. There are listings on the AWS Marketplace which offer Amazon Linux AMIs and Windows Server AMIs with the NVIDIA drivers pre-installed. You may also launch 64-bit, HVM AMIs and install the drivers yourself. You must visit the NVIDIA driver website and search for the NVIDIA Tesla V100 for P3, NVIDIA Tesla K80 for P2, and NVIDIA Tesla M60 for G3 instances.
Q. Which AMIs can I use with P3, P2 and G3 instances?
You can currently use Windows Server, SUSE Enterprise Linux, Ubuntu, and Amazon Linux AMIs on P2 and G3 instances. P3 instances only support HVM AMIs. If you want to launch AMIs with operating systems not listed here, contact AWS Customer Support with your request or reach out through EC2 Forums.
Q. Does the use of G2 and G3 instances require third-party licenses?
Aside from the NVIDIA drivers and GRID SDK, the use of G2 and G3 instances does not necessarily require any third-party licenses. However, you are responsible for determining whether your content or technology used on G2 and G3 instances requires any additional licensing. For example, if you are streaming content you may need licenses for some or all of that content. If you are using third-party technology such as operating systems, audio and/or video encoders, and decoders from Microsoft, Thomson, Fraunhofer IIS, Sisvel S.p.A., MPEG-LA, and Coding Technologies, please consult these providers to determine if a license is required. For example, if you leverage the on-board h.264 video encoder on the NVIDIA GRID GPU you should reach out to MPEG-LA for guidance, and if you use mp3 technology you should contact Thomson for guidance.
Q. Why am I not getting NVIDIA GRID features on G3 instances using the driver downloaded from NVIDIA website?
The NVIDIA Tesla M60 GPU used in G3 instances requires a special NVIDIA GRID driver to enable all advanced graphics features, and 4 monitors support with resolution up to 4096x2160. You need to use an AMI with NVIDIA GRID driver pre-installed, or download and install the NVIDIA GRID driver following the AWS documentation.
Q. Why am I unable to see the GPU when using Microsoft Remote Desktop?
When using Remote Desktop, GPUs using the WDDM driver model are replaced with a non-accelerated Remote Desktop display driver. In order to access your GPU hardware, you need to utilize a different remote access tool, such as VNC.
Q. What is Amazon EC2 F1?
Amazon EC2 F1 is a compute instance with programmable hardware you can use for application acceleration. The new F1 instance type provides a high performance, easy to access FPGA for developing and deploying custom hardware accelerations.
Q. What are FPGAs and why do I need them?
FPGAs are programmable integrated circuits that you can configure using software. By using FPGAs you can accelerate your applications up to 30x when compared with servers that use CPUs alone. And, FPGAs are reprogrammable, so you get the flexibility to update and optimize your hardware acceleration without having to redesign the hardware.
Q. How does F1 compare with traditional FPGA solutions?
F1 is an AWS instance with programmable hardware for application acceleration. With F1, you have access to FPGA hardware in a few simple clicks, reducing the time and cost of full-cycle FPGA development and scale deployment from months or years to days. While FPGA technology has been available for decades, adoption of application acceleration has struggled to be successful in both the development of accelerators and the business model of selling custom hardware for traditional enterprises, due to time and cost in development infrastructure, hardware design, and at-scale deployment. With this offering, customers avoid the undifferentiated heavy lifting associated with developing FPGAs in on-premises data centers.
Q: What is an Amazon FPGA Image (AFI)?
The design that you create to program your FPGA is called an Amazon FPGA Image (AFI). AWS provides a service to register, manage, copy, query, and delete AFIs. After an AFI is created, it can be loaded on a running F1 instance. You can load multiple AFIs to the same F1 instance, and can switch between AFIs in runtime without reboot. This lets you quickly test and run multiple hardware accelerations in rapid sequence. You can also offer to other customers on the AWS Marketplace a combination of your FPGA acceleration and an AMI with custom software or AFI drivers.
Q. How do I list my hardware acceleration on the AWS Marketplace?
You would develop your AFI and the software drivers/tools to use this AFI. You would then package these software tools/drivers into an Amazon Machine Image (AMI) in an encrypted format. AWS manages all AFIs in the encrypted format you provide to maintain the security of your code. To sell a product in the AWS Marketplace, you or your company must sign up to be an AWS Marketplace reseller, you would then submit your AMI ID and the AFI ID(s) intended to be packaged in a single product. AWS Marketplace will take care of cloning the AMI and AFI(s) to create a product, and associate a product code to these artifacts, such that any end-user subscribing to this product code would have access to this AMI and the AFI(s).
Q. What is available with F1 instances?
For developers, AWS is providing a Hardware Development Kit (HDK) to help accelerate development cycles, a FPGA Developer AMI for development in the cloud, an SDK for AMIs running the F1 instance, and a set of APIs to register, manage, copy, query, and delete AFIs. Both developers and customers have access to the AWS Marketplace where AFIs can be listed and purchased for use in application accelerations.
Q. Do I need to be a FPGA expert to use an F1 instance?
AWS customers subscribing to a F1-optimized AMI from AWS Marketplace do not need to know anything about FPGAs to take advantage of the accelerations provided by the F1 instance and the AWS Marketplace. Simply subscribe to an F1-optimized AMI from the AWS Marketplace with an acceleration that matches the workload. The AMI contains all the software necessary for using the FPGA acceleration. Customers need only write software to the specific API for that accelerator and start using the accelerator.
Q. I’m a FPGA developer, how do I get started with F1 instances?
Developers can get started on the F1 instance by creating an AWS account and downloading the AWS Hardware Development Kit (HDK). The HDK includes documentation on F1, internal FPGA interfaces, and compiler scripts for generating AFI. Developers can start writing their FPGA code to the documented interfaces included in the HDK to create their acceleration function. Developers can launch AWS instances with the FPGA Developer AMI. This AMI includes the development tools needed to compile and simulate the FPGA code. The Developer AMI is best run on the latest C5, M5, or R5 instances. Developers should have experience in the programming languages used for creating FPGA code (i.e. Verilog or VHDL) and an understanding of the operation they wish to accelerate.
Q. I’m not an FPGA developer, how do I get started with F1 instances?
Customers can get started with F1 instances by selecting an accelerator from the AWS Marketplace, provided by AWS Marketplace sellers, and launching an F1 instance with that AMI. The AMI includes all of the software and APIs for that accelerator. AWS manages programming the FPGA with the AFI for that accelerator. Customers do not need any FPGA experience or knowledge to use these accelerators. They can work completely at the software API level for that accelerator.
Q. Does AWS provide a developer kit?
Yes. The Hardware Development Kit (HDK) includes simulation tools and simulation models for developers to simulate, debug, build, and register their acceleration code. The HDK includes code samples, compile scripts, debug interfaces, and many other tools you will need to develop the FPGA code for your F1 instances. You can use the HDK either in an AWS provided AMI, or in your on-premises development environment. These models and scripts are available publically with an AWS account.
Q. Can I use the HDK in my on-premises development environment?
Yes. You can use the Hardware Development Kit HDK either in an AWS-provided AMI, or in your on-premises development environment.
Q. Can I add an FPGA to any EC2 instance type?
No. F1 instances come in three instance sizes f1.2xlarge, f1.4xlarge, and f1.16 xlarge.
相關推薦
Amazon EC2 FAQs
Q: What are Accelerated Computing instances? Accelerated Computing instance family is a family of instances which use hardware accel
Amazon EC2 Auto Scaling FAQs
Q: What are the benefits of using Amazon EC2 Auto Scaling? Amazon EC2 Auto Scaling helps to maintain your Amazon EC2 instance avai
Eucalyptus企業雲計算(建立能夠和Amazon EC2兼容的雲)
title 創建者 一個 2009年 但是 聯合 網絡設備 環境 存儲 Eucalyptus是與一個在加利福尼亞大學的研究性項目,創建了一個使企業能夠使用它們內部IT資源(包括服務器、存儲系統、網絡設備)的開源界面,來建立能夠和Amazon EC2兼容的雲。 &ldquo
Amazon EC2雲端服務器的使用方法
mod ins windows zip 服務 post root 權限 name Amazon的EC2服務器可以理解為虛擬機,不過它是不需要安裝系統的,它是根據鏡像自動創建的。在申請EC2的時候,可以選擇操作系統的類型,如Redhat Enterprise 6或ubuntu
通過SecureCRT訪問亞馬遜Amazon EC2主機
pen 生成密鑰 class 舉例 免費 linux 需要 輸入 面板 亞馬遜推出了免費的雲主機服務器 Amazon EC2,它是通過安全密鑰來訪問主機的。問題是下載的密鑰在SecureCRT 上無法直接使用,需要轉換。 下面的方法可以在自己的linux主機上生成secur
使用密碼登陸Amazon EC2
restart fig star 登錄 blog inux 9.1 config 添加 用的是亞馬遜推薦的系統:Amazon Linux AMI 2015.09.1 (HVM)創建instance後,會有一個pem的key使用該密鑰登陸到服務器上後,默認用戶名為 ec2-
《Spark 官方文件》在Amazon EC2上執行Spark
在Amazon EC2上執行Spark Spark的ec2目錄下有一個spark-ec2指令碼,可以幫助你在Amazon EC2上啟動、管理、關閉Spark叢集。該指令碼能在EC2叢集上自動設定好Spark和HDFS。本文將會詳細描述如何利用spark-ec2指令碼來啟動和關閉叢集,以及如何在叢集提交作業。
亞馬遜雲伺服器VPS Amazon EC2 免費VPS主機配置CentOS及其它內容
Amazon目前提供為期一年的免費VPS服務,可到地址http://aws.amazon.com 進行申請。 現在對賬號申請成功後,對VPS主機配置CentOS的過程做個圖文介紹 1、建立例項(Instance):建立一個例項,就是建立一個VPS主機。VPS主機目前支援Windows和Linux。在建立例
Amazon EC2 Systems Manager*
AWS Systems Manager is a new way to manage your cloud and hybrid IT environments. AWS Systems Manager provides a unified user interface that sim
Amazon EC2 Pricing
With per-second billing, you pay for only what you use. It takes cost of unused minutes and seconds in an hour off of the bill, so you can focu
Amazon EC2競價型例項_Amazon EC2優勢_AWS例項
Amazon EC2 Spot 例項是 AWS 雲中的可用空閒計算容量,與按需例項的價格相比,這類例項可提供超低折扣。EC2 Spot 可幫助您優化 AWS 雲的成本,可在預算相同的情況下將應用程式的吞吐量提高到 10 倍。您只需在啟動 EC2 例項時選擇“Spot”,即可節省按需例項價格的
Amazon EC2 Instance Types
Amazon EC2 allows you to choose between Fixed Performance Instances (e.g. M5, C5, and R5) and Burstable Performance Instances (e.g. T3). Burstabl
在Amazon EC2上配置Thin+NginX+Rails全攻略 (Linux環境)
upstream domain1 { server 0.0.0.0:3000; server 0.0.0.0:3001; server 0.0.0.0:3002; } server { listen 80;
在Amazon EC2主機上安裝Zmodem協議傳輸工具lrzsz
Amazon EC2主機上預設沒有安裝支援Zmodem協議的傳輸工具lrzsz,很不方便,按照以下步驟安裝: 1、下載lrzsz-1.12.20.tar.gz wget http://down1.chinaunix.net/distfiles/lrzsz-0.12.20
Amazon EC2安裝mysql多例項並配置主從複製
1、MySQL安裝: EC2直接使用yum安裝mysql很方便: yum install mysql mysql-server mysql-libs 此處安裝了三個包 完事之後可以直接使用 service mysqld start 啟動mysql服務
Amazon EC2製作自定義AMI
測試系統:Windows server 2008 實現功能:1.製作自定義AMI,生成instance時密碼隨機,動態桌面桌布。 2.生成instance時自動執行powershell指令碼,實現自動設定DNS地址,加入Domain 1.製作AMI server2008中
Amazon EC2 Developer Resources
Amazon Web Services is Hiring. Amazon Web Services (AWS) is a dynamic, growing business unit within Amazon.com. We are currently hiring So
Amazon EC2 z1d Instances
Many workloads such as electronic design automation (EDA) and relational databases have high per core software licensing fees. Semiconductor firms
New P2 Instance Type for Amazon EC2 – Up to 16 GPUs
I like to watch long-term technology and business trends and watch as they shape the products and services that I get to use and to write about. A