Cloud Solutions


Run your

AWS

Cloud computing is ideal for running flexible, scalable applications on demand, in periodic bursts, or for fixed periods of time. UVA Research Computing works alongside researchers to design and run research applications and datasets into Amazon Web Services, the leader among public cloud vendors. This means that server, storage, and database needs do not have to be estimated or purchased beforehand – they can be scaled larger and smaller with your needs, or programmed to scale dynamically with your application.


Service Oriented Architecture

A key advantage of the cloud is that for many services you do not need to build or maintain the servers that support the service – you simply use it.

Here are some of the building blocks available using cloud infrastructure:

  •  Compute
  •  Storage
  •  Databases
  • Containers / Docker
  • Analytics / Data Management
  •  Continuous Integration
  •   Sensor / IoT Data Streaming
  • Message Queues / Brokers
  • SMS / Push Integration
  •   Alexa Skills / Speech Integration
  • Serverless Computing
  •   Code Build / Validation

Researchers Using the Cloud

Serverless Web UVA faculty and researchers can share data, findings, tools and other resources from static HTML content published to object storage. This simple method for publishing can cost only a few dollars a month and requires no server management.
Data Lakes A new paradigm in data storage and processing, data lakes help researchers by providing a central repository for both structured and unstructured data, of any type or size. These data can then be siphoned off for processing, either in real-time streams or in queues for later analysis.
Services in Support of HPC Users of HPC usually have more than enough computing power to run their jobs. But what if you need a relational or NoSQL database, a messaging service, or offsite storage? Researchers have begun integrating the cloud into their HPC jobs to create, use, and manage external services like these.
HIPAA-Compliant Computing Researchers working on clinical datasets use Ivy, our private virtualized platform to perform HIPAA compliant analytics and compute jobs. This platform offers virtual machines, an R/Python data analytics tool, and Hadoop/Spark for larger analytics projects. Many users in Ivy work with EPIC clinical data alongside other highly-sensitive datasets for their investigations.
Workflows & Pipeline Management Workflows Researchers need flexibility for where they run their data pipelines -- it might be on a personal computer, a lab server, an HPC cluster, or a cloud instance. We are working with faculty to extend some commonly-used pipeline tools so that they can create and push jobs to cloud-based resources, regardless of the cloud vendor.
Long-term Cold Storage AWS Glacier and Google Nearline/Coldline offer researchers "cold" offsite storage for long-term backups of infrequently-accessed data. Many researchers use Glacier to store terabytes of source data to fulfill grants and federal research project compliance.

Other Common Use Cases

  • Proofs of concept - To verify a system or design works, to benchmark processing speeds, we may use short-lived instances to learn from before building a production system.

  • Test / Development environments - For installing test packages, trying new ideas, and testing design patterns.

  • Dynamic / flexible / scaling application stacks - When future traffic or load cannot be determined beforehand, deploying into a dynamic environment means the infrastructure is not locked into any set type of CPU/RAM or scale.

  • Short-term or fast deployment projects - For almost immediate computing needs, existing users can create new instances as needed.

  • Container deployments - Run microservices (such as Docker containers) in an environment that can load-balance their traffic and maintain container health.


Reference Architecture

To get an idea of how public or private cloud resources are used in real-world and research scenarios, visit one of these Solution Architecture References:

Some examples from AWS:

AWS-batch-processing-thumb

Batch Processing
Build auto-scalable batch processing systems like video/image/datastream processing pipelines.

AWS-large-scale-processing-huge-data-thumb

Large Scale Processing and Huge Data sets
Build high-performance computing systems that involve Big Data.

AWS-large-scale-processing-huge-data-thumb

Time Series and Streaming Data Processing
Build elastic systems that process chronological data.


Cloud Services at UVA

Request an Account

Researchers or labs who would like to use AWS for their computing infrastructure should request an account through UVA Information Technology Services (ITS). They currently support deployments in Amazon Web Services and Microsoft Azure, and offer both managed and self-service options.

Training & Implementation

With an AWS account in hand, you will need some training. ITS can provide you with links to self-paced training for both AWS and Azure. Research Computing

If you need help with how to design your project in a cloud environment, or thinking through how to migrate your existing projects, contact us for a consultation.

Sensitive Data in the Cloud

If your cloud-based project involves any sensitive data (HIPAA, PHI, etc.) you must request approval from the Information Security office at UVA. You will be required to verify that your application, infrastructure, and staff can meet all minimum requirements for the secure transfer and handling of sensitive data.

Please note that while sensitive data projects are possible in the cloud, their approval is not automatic nor guaranteed.

Solution Architecture / Consulting

We have experience designing and delivering solutions to the public cloud using industry best practices. If you have a project and would like to discuss options, pricing, design, or implementation, we are available for consultation. Our staff includes an AWS certified solution architect, and the RC team uses AWS for our own internal systems and development.


Training

We also offer in-person, hands-on workshops and sessions on working with the cloud. Workshops cover a number of topics, from creating object storage buckets and simple compute instances to more complex data-driven workflows and Docker containers, If you have an idea for a workshop or would like to schedule training for your lab or group, please contact us.


AWS Educate Member
Certified Solution Architect - Professional