Category

HPC

Women in HPC launches mentoring program

By | Educational, General Interest, HPC, News

Women in High Performance Computing (WHPC) has launched a year-round mentoring program, providing a framework for women to provide or receive mentorship in high performance computing. Read more about the program at https://womeninhpc.org/2019/03/mentoring-programme-2019/

WHPC was created with the vision to encourage women to participate in the HPC community by providing fellowship, education, and support to women and the organizations that employ them. Through collaboration and networking, WHPC strives to bring together women in HPC and technical computing while encouraging women to engage in outreach activities and improve the visibility of inspirational role models.

The University of Michigan has been recognized as one of the first Chapters in the new Women in High Performance Computing (WHPC) Pilot Program. Read more about U-M’s chapter at https://arc.umich.edu/whpc/

ARC Director Sharon Broude Geva re-elected Chair of the Coalition for Academic Scientific Computation

By | HPC, News

Dr. Sharon Broude Geva, Director of Advanced Research Computing at the University of Michigan, has been re-elected as Chair of the Coalition for Academic Scientific Computation (CASC) for 2020.

Founded in 1989, CASC advocates for the use of advanced computing technology to accelerate scientific discovery for national competitiveness, global security, and economic success. The organization’s members represent 87 institutions of higher education and national labs.

The chair position is one of four elected CASC executive officers. The officers work closely as a team with the director of CASC. The Chair is responsible for arranging and presiding over general CASC meetings and acts as an official representative of CASC.

Geva served as CASC secretary in 2015 and 2016, and vice-chair in 2017 and 2018.

 

U-M selects Dell EMC, Mellanox and DDN to Supply New “Great Lakes” Computing Cluster

By | Flux, General Interest, Happenings, HPC, News

The University of Michigan has selected Dell EMC as lead vendor to supply its new $4.8 million Great Lakes computing cluster, which will serve researchers across campus. Mellanox Technologies will provide networking solutions, and DDN will supply storage hardware.

Great Lakes will be available to the campus community in the first half of 2019, and over time will replace the Flux supercomputer, which serves more than 2,500 active users at U-M for research ranging from aerospace engineering simulations and molecular dynamics modeling to genomics and cell biology to machine learning and artificial intelligence.

Great Lakes will be the first cluster in the world to use the Mellanox HDR 200 gigabit per second InfiniBand networking solution, enabling faster data transfer speeds and increased application performance.

“High-performance research computing is a critical component of the rich computing ecosystem that supports the university’s core mission,” said Ravi Pendse, U-M’s vice president for information technology and chief information officer. “With Great Lakes, researchers in emerging fields like machine learning and precision health will have access to a higher level of computational power. We’re thrilled to be working with Dell EMC, Mellanox, and DDN; the end result will be improved performance, flexibility, and reliability for U-M researchers.”

“Dell EMC is thrilled to collaborate with the University of Michigan and our technology partners to bring this innovative and powerful system to such a strong community of researchers,” said Thierry Pellegrino, vice president, Dell EMC High Performance Computing. “This Great Lakes cluster will offer an exceptional boost in performance, throughput and response to reduce the time needed for U-M researches to make the next big discovery in a range of disciplines from artificial intelligence to genomics and bioscience.”

The main components of the new cluster are:

  • Dell EMC PowerEdge C6420 compute nodes, PowerEdge R640 high memory nodes, and PowerEdge R740 GPU nodes
  • Mellanox HDR 200Gb/s InfiniBand ConnectX-6 adapters, Quantum switches and LinkX cables, and InfiniBand gateway platforms
  • DDN GRIDScaler® 14KX® and 100 TB of usable IME® (Infinite Memory Engine) memory

“HDR 200G InfiniBand provides the highest data speed and smart In-Network Computing acceleration engines, delivering HPC and AI applications with the best performance, scalability and efficiency,” said Gilad Shainer, vice president of marketing at Mellanox Technologies. “We are excited to collaborate with the University of Michigan, Dell EMC and DataDirect Networks, in building a leading HDR 200G InfiniBand-based supercomputer, serving the growing demands of U-M researchers.”

“DDN has a long history of working with Dell EMC and Mellanox to deliver optimized solutions for our customers. We are happy to be a part of the new Great Lakes cluster, supporting its mission of advanced research and computing. Partnering with forward-looking thought leaders as these is always enlightening and enriching,” said Dr. James Coomer, SVP Product Marketing and Benchmarks at DDN.

Great Lakes will provide significant improvement in computing performance over Flux. For example, each compute node will have more cores, higher maximum speed capabilities, and increased memory. The cluster will also have improved internet connectivity and file system performance, as well as NVIDIA Tensor GPU cores, which are very powerful for machine learning compared to prior generations of GPUs.

“Users of Great Lakes will have access to more cores, faster cores, faster memory, faster storage, and a more balanced network,” said Brock Palen, Director of Advanced Research Computing – Technology Services (ARC-TS).

The Flux cluster was created approximately 8 years ago, although many of the individual nodes have been added since then. Great Lakes represents an architectural overhaul that will result in better performance and efficiency. Based on extensive input from faculty and other stakeholders across campus, the new Great Lakes cluster will be designed to deliver similar services and capabilities as Flux, including the ability to accommodate faculty purchases of hardware, access to GPUs and large-memory nodes, and improved support for emerging uses such as machine learning and genomics.

ARC-TS will operate and maintain the cluster once it is built. Allocations of computing resources through ARC-TS include access to hundreds of software titles, as well as support and consulting from professional staff with decades of combined experience in research computing.

Updates on the progress of Great Lakes will be available at https://arc-ts.umich.edu/greatlakes/.

ARC-TS seeks pilot users for two new research storage services

By | General Interest, Happenings, HPC, News

Advanced Research Computing – Technology Services (ARC-TS) is seeking pilot users for two new research storage services.

The first, Locker, is group project storage focused on large data sets, and is available at a cost less than half that of current primary storage services. Locker still provides encryption, replication, snapshots, and workstation access. Example use cases for Locker are research projects in climate studies, genomics, imaging, and other data-intensive sciences.

The second service, Data Den, provides archive class storage for research data that is not actively used. As our lowest cost research storage offering, Data Den provides “cold storage” for massive amounts of data with 20 petabytes of encrypted and replicated capacity. Data Den allows researchers to preserve data between rounds of funding and management plans, and to free up space in more expensive primary storage by moving valuable, but not currently used, data.

Those interested in participating in the pilots should contact ARC-TS at hpc-support@umich.edu.

HPC training workshops begin Tuesday, Feb. 13

By | Educational, Events, General Interest, Happenings, HPC, News

series of training workshops in high performance computing will be held Feb. 12 through March 6, 2018, presented by CSCAR in conjunction with Advanced Research Computing – Technology Services (ARC-TS).

Introduction to the Linux command Line
This course will familiarize the student with the basics of accessing and interacting with Linux computers using the GNU/Linux operating system’s Bash shell, also known as the “command line.”
Location: East Hall, Room B254, 530 Church St.
Dates: (Please sign up for only one)
• Tuesday, Feb. 13, 1 – 4 p.m. (full descriptionregistration)
• Friday, Feb. 16, 9 a.m. – noon (full description | registration)

Introduction to the Flux cluster and batch computing
This workshop will provide a brief overview of the components of the Flux cluster, including the resource manager and scheduler, and will offer students hands-on experience.
Location: East Hall, Room B254, 530 Church St.
Dates: (Please sign up for only one)
• Monday, Feb. 19, 1 – 4 p.m. (full description | registration)
• Tuesday, March 6, 1 – 4 p.m. (full description | registration)

Advanced batch computing on the Flux cluster
This course will cover advanced areas of cluster computing on the Flux cluster, including common parallel programming models, dependent and array scheduling, and a brief introduction to scientific computing with Python, among other topics.
Location: East Hall, Room B250, 530 Church St.
Dates: (Please sign up for only one)
• Wednesday, Feb. 21, 1 – 5 p.m. (full description | registration)
• Friday, Feb. 23, 1 – 5 p.m. (full description | registration)

Hadoop and Spark workshop
Learn how to process large amounts (up to terabytes) of data using SQL and/or simple programming models available in Python, R, Scala, and Java.
Location: East Hall, Room B250, 530 Church St.
Dates: (Please sign up for only one)
• Thursday, Feb. 22, 1 – 5 p.m. (full description | registration)

U-M fosters thriving artificial intelligence and machine learning research

By | General Interest, HPC, News, Research

Research using machine learning and artificial intelligence — tools that allow computers to learn about and predict outcomes from massive datasets — has been booming at the University of Michigan. The potential societal benefits being explored on campus are numerous, from on-demand transportation systems to self-driving vehicles to individualized medical treatments to improved battery capabilities.

The ability of computers and machines generally to learn from their environments is having transformative effects on a host of industries — including finance, healthcare, manufacturing, and transportation — and could have an economic impact of $15 trillion globally according to one estimate.

But as these methods become more accurate and refined, and as the datasets needed become bigger and bigger, keeping up with the latest developments and identifying and securing the necessary resources — whether that means computing power, data storage services, or software development — can be complicated and time-consuming. And that’s not to mention complying with privacy regulations when medical data is involved.

“Machine learning tools have gotten a lot better in the last 10 years,” said Matthew Johnson-Roberson, Assistant Professor of Engineering in the Department of Naval Architecture & Marine Engineering and the Department of Electrical Engineering and Computer Science. “The field is changing now at such a rapid pace compared to what it used to be. It takes a lot of time and energy to stay current.”

Diagram representing the knowledge graph of an artificial intelligence system, courtesy of Jason Mars, assistant professor, Electrical Engineering and Computer Science, U-M

Johnson-Roberson’s research is focused on getting computers and robots to better recognize and adapt to the world, whether in driverless cars or deep-sea mapping robots.

“The goal in general is to enable robots to operate in more challenging environments with high levels of reliability,” he said.

Johnson-Roberson said that U-M has many of the crucial ingredients for success in this area — a deep pool of talented researchers across many disciplines ready to collaborate, flexible and personalized support, and the availability of computing resources that can handle storing the large datasets and heavy computing load necessary for machine learning.

“The people is one of the reasons I came here,” he said. “There’s a broad and diverse set of talented researchers across the university, and I can interface with lots of other domains, whether it’s the environment, health care, transportation or energy.”

“Access to high powered computing is critical for the computing-intensive tasks, and being able to leverage that is important,” he continued. “One of the challenges is the data. A major driver in machine learning is data, and as the datasets get more and more voluminous, so does the storage needs.”

Yuekai Sun, an assistant professor in the Statistics Department, develops algorithms and other computational tools to help researchers analyze large datasets, for example, in natural language processing. He agreed that being able to work with scientists from many different disciplines is crucial to his research.

“I certainly find the size of Michigan and the inherent diversity that comes with it very stimulating,” he said. “Having people around who are actually working in these application areas helps guide the direction and the questions that you ask.”

Sun is also working on analyzing the potential discriminatory effects of algorithms used in decisions like whether to give someone a loan or to grant prisoners parole.

“If you use machine learning, how do you hold an algorithm or the people who apply it accountable? What does it mean for an algorithm to be fair?” he said. “Can you check whether this notion of non-discrimination is satisfied?”

Jason Mars, an assistant professor in the Electrical Engineering and Computer Science department and co-founder of a successful spinoff called Clinc, is applying artificial intelligence to driverless car technology and a mobile banking app that has been adopted by several large financial institutions. The app, called Finie, provides a much more conversational interface between users and their financial information than other apps in the field.

“There is going to be an expansion of the number of problems solved and number of contributions that are AI-based,” Mars said. He predicted that more researchers at U-M will begin exploring AI and ML as they understand the potential.

“It’s going to require having the right partner, the right experts, the right infrastructure, and the best practices of how to use them,” he said.

He added that U-M does a “phenomenal job” in supporting researchers conducting AI and ML research.

“The level of support and service is awesome here,” he said. “Not to mention that the infrastructure is state of the art. We stay relevant to the best techniques and practices out there.”

Advanced Research Computing at U-M, in part through resources from the university-wide Data Science Initiative, provides computing infrastructure, consulting expertise, and support for interdisciplinary research projects to help scientists conducting artificial intelligence and machine learning research.

For example, Consulting for Statistics, Computing and Analytics Research, an ARC unit, has several consultants on staff with expertise in machine learning and predictive analysis with large, complex, and heterogeneous data. CSCAR recently expanded capacity to support very large-scale machine learning using tools such as Google’s TensorFlow.

CSCAR consultants are available by appointment or on a drop-in basis, free of charge. See cscar.research.umich.edu or email cscar@umich.edu for more information.

CSCAR also provides workshops on topics in machine learning and other areas of data science, including sessions on Machine Learning in Python, and an upcoming workshop in March titled “Machine Learning, Concepts and Applications.”

The computing resources available to machine learning and artificial intelligence researchers are significant and diverse. Along with the campus-wide high performance computing cluster known as Flux, the recently announced Big Data cluster Cavium ThunderX will give researchers a powerful new platform for hosting artificial intelligence and machine learning work. Both clusters are provided by Advanced Research Computing – Technology Services (ARC-TS).

All allocations on ARC-TS clusters include access to software packages that support AI/ML research, including TensorFlow, Torch, and Spark ML, among others.

ARC-TS also operates the Yottabyte Research Cloud (YBRC), a customizable computing platform that recently gained the capacity to host and analyze data governed by the HIPAA federal privacy law.

Also, the Michigan Institute for Data Science (MIDAS) (also a unit of ARC) has supported several AI/ML projects through its Challenge Initiative program, which has awarded more than $10 million in research support since 2015.

For example, the Analytics for Learners as People project is using sensor-based machine learning tools to translate data on academic performance, social media, and survey data into attributes that will form student profiles. Those profiles will help link academic performance and mental health with the personal attributes of students, including values, beliefs, interests, behaviors, background, and emotional state.

Another example is the Reinventing Public Urban Transportation and Mobility project, which is using predictive models based on machine learning to develop on-demand, multi-modal transportation systems for urban areas.

In addition, MIDAS supports student groups involved in this type of research such as the Michigan Student Artificial Intelligence Lab (MSAIL) and the Michigan Data Science Team (MDST).

(A version of this piece appeared in the University Record.)

Yottabyte Research Cloud able to accept HIPAA-aligned data

By | General Interest, HPC, News

Advanced Research Computing – Technology Services (ARC-TS) is pleased to announce that the Yottabyte Research Cloud (YBRC) computing platform is now HIPAA-compliant. This means that YBRC and its associated services can accept restricted data, enabling secure data analysis on Windows and Linux virtual desktops as well as secure hosting of databases and data ingestion.

The new capability ensures the security of restricted data through the creation of firewalled network enclaves, allowing HIPAA-aligned data to be analyzed safely and securely in YBRC’s flexible, robust and scalable environment.   Within each network enclave, researchers have access to Windows and Linux virtual desktops that can contain any software required for their analysis pipeline.

This capability also extends to our database and ingestion services:

  • Structured databases:  MySQL/MariaDB, and PostgreSQL.
  • Unstructured databases: Cassandra, MongoDB, InfluxDB, Grafana, and ElasticSearch.
  • Data ingestion: Redis, Kafka, RabbitMQ.
  • Data processing: Apache Flink, Apache Storm, Node.js and Apache NiFi.
  • Other data services are available upon request.

YBRC is supported by U-M’s Data Science Initiative launched in 2015. YBRC was created through a partnership between Yottabyte and ARC-TS announced last fall.

These tools are offered to all researchers at the University of Michigan free of charge, provided that certain usage restrictions are not exceeded. Large-scale users who outgrow the no-cost allotment may purchase additional YBRC resources. All interested parties should contact hpc-support@umich.edu.

U-M wraps up successful SC17 conference

By | General Interest, Happenings, HPC, News

Several University of Michigan researchers and professional IT staff attended the Supercomputing 17 (SC17) conference in Denver from Nov. 12-17, participating in a number of different ways, including demonstrations, presentations and tutorials.

U-M participation included:

  • Matt McLean, a Big Data systems administrator with ARC-TS, served as a panelist at a session titled “The ARM Software Ecosystem: Are We There Yet?” (Slides)
  • Jeff Sica, a research database administrator with ARC-TS, helped lead a Birds of a Feather session titled “Containers in HPC.” (Slides)
  • Quentin Stout (EECS) and Christiane Jablonowski (CLASP) taught the “Parallel Computing 101” tutorial.
  • Shawn McKee, U-M Department of Physics, and OSiRIS Principal Investigator, demonstrated Object Storage and Caching for Science (network topology diagrams)
  • Eric Boyd, Director of Research Networks, presented on Research Networking at the University of Michigan at the U-M exhibit booth.
  • Simon Adorf, Ph.D. Candidate, Chemical Engineering Department, U-M, presented on Simple Data and Workflow Management with Signac and GPU-Accelerated Predictive Material Design at the U-M exhibit booth.
  • ARC sponsored a networking and career networking reception put on by Women in HPC. ARC Director Sharon Broude Geva spoke at the event.
  • Amy Liebowitz, a network architect at ITS, worked on SCINet, a high-capacity network created every year for the conference. Liebowitz was on the routing team, which is responsible for installing, configuring and supporting the high performance conference network. The Routing Team also coordinated external connectivity with commodity Internet and R&E WAN service providers.

U-M partners with Cavium on Big Data computing platform

By | Feature, General Interest, Happenings, HPC, News

A new partnership between the University of Michigan and Cavium Inc., a San Jose-based provider of semiconductor products, will create a powerful new Big Data computing cluster available to all U-M researchers.

The $3.5 million ThunderX computing cluster will enable U-M researchers to, for example, process massive amounts of data generated by remote sensors in distributed manufacturing environments, or by test fleets of automated and connected vehicles.

The cluster will run the Hortonworks Data Platform providing Spark, Hadoop MapReduce and other tools for large-scale data processing.

“U-M scientists are conducting groundbreaking research in Big Data already, in areas like connected and automated transportation, learning analytics, precision medicine and social science. This partnership with Cavium will accelerate the pace of data-driven research and opening up new avenues of inquiry,” said Eric Michielssen, U-M associate vice president for advanced research computing and the Louise Ganiard Johnson Professor of Engineering in the Department of Electrical Engineering and Computer Science.

“I know from experience that U-M researchers are capable of amazing discoveries. Cavium is honored to help break new ground in Big Data research at one of the top universities in the world,” said Cavium founder and CEO Syed Ali, who received a master of science in electrical engineering from U-M in 1981.

Cavium Inc. is a leading provider of semiconductor products that enable secure and intelligent processing for enterprise, data center, wired and wireless networking. The new U-M system will use dual socket servers powered by Cavium’s ThunderX ARMv8-A workload optimized processors.

The ThunderX product family is Cavium’s 64-bit ARMv8-A server processor for next generation Data Center and Cloud applications, and features high performance custom cores, single and dual socket configurations, high memory bandwidth and large memory capacity.

Alec Gallimore, the Robert J. Vlasic Dean of Engineering at U-M, said the Cavium partnership represents a milestone in the development of the College of Engineering and the university.

“It is clear that the ability to rapidly gain insights into vast amounts of data is key to the next wave of engineering and science breakthroughs. Without a doubt, the Cavium platform will allow our faculty and researchers to harness the power of Big Data, both in the classroom and in their research,” said Gallimore, who is also the Richard F. and Eleanor A. Towner Professor, an Arthur F. Thurnau Professor, and a professor both of aerospace engineering and of applied physics.

Along with applications in fields like manufacturing and transportation, the platform will enable researchers in the social, health and information sciences to more easily mine large, structured and unstructured datasets. This will eventually allow, for example, researchers to discover correlations between health outcomes and disease outbreaks with information derived from socioeconomic, geospatial and environmental data streams.

U-M and Cavium chose to run the cluster on Hortonworks Data Platform, which is based on open source Apache Hadoop. The ThunderX cluster will deliver high performance computer services for the Hadoop analytics and, ultimately, a total of three petabytes of storage space.

“Hortonworks is excited to be a part of forward-leading research at the University of Michigan exploring low-powered, high-performance computing,” said Nadeem Asghar, vice president and global head of technical alliances at Hortonworks. “We see this as a great opportunity to further expand the platform and segment enablement for Hortonworks and the ARM community.”