뉴스레터

이메일로 Hortonworks의 새 업데이트를 받으세요.

한 달에 한 번 빅 데이터와 관련한 최신 인사이트, 동향, 분석 정보, 지식을 받아 보세요.

AVAILABLE NEWSLETTERS:

Sign up for the Developers Newsletter

한 달에 한 번 빅 데이터와 관련한 최신 인사이트, 동향, 분석 정보, 지식을 받아 보세요.

CTA

시작하기

클라우드

시작할 준비가 되셨습니까?

Sandbox 다운로드

어떤 도움이 필요하십니까?

* 저는 언제든지 구독을 해지할 수 있다는 점을 이해합니다. 또한 저는 Hortonworks이 개인정보 보호정책에 추가된 정보를 확인하였습니다.
닫기닫기 버튼
October 03, 2018
이전 슬라이드다음 슬라이드

How Johns Hopkins is Utilizing Apache Hadoop to Securely Access Log Events

작성자:
Matt Spillar

Yesterday was Customer Experience Day, a day where we had the opportunity to celebrate the people and companies that make great customer experiences happen and recognize great customer work! As part of this celebration, we had a customer webinar with Johns Hopkins University.

Johns Hopkins University is an American private research university, founded in 1876 and located in Baltimore, Maryland. It is considered the first research university in the United States, and is organized into 10 divisions on campuses in Maryland and Washington, D.C. These divisions include the Johns Hopkins School of Medicine and the Applied Physics Laboratory, among various others.

In this webinar we heard from Conrad Fernandes, a Cloud and Cyber Security Engineer at APL. Fernandes is a long time cyber security engineer and architect, having worked with US Defense agencies and the DoD since the early 2000’s while at Booz Allen Hamilton. He currently serves as a senior cyber security engineer at the Johns Hopkins Applied Physics Laboratory (APL), where he leads security and governance practices on emerging cloud technologies, including commercial and US GovCloud (e.g., Amazon web services) and Hadoop-based data science platforms).

Fernandes talks about the strategies used to collect, audit, and access log events from key Apache Hadoop services and forwarding to a central server for monitoring, analysis, and response to a suspected breach. This mission critical for precision medicine and HIPAA sensitive data, and unclassified information (CUI) for defense projects. Johns Hopkins needed a platform that was both robust and secure for housing all of this data. The modern data architecture that was put into place includes Hortonworks DataFlow, and security and governance with Apache Atlas, Apache Ranger, and Apache Knox.

The university is now able to achieve results that would never be possible without a big data platform. Now data can be ingested from disparate sources, transported from various sources to the Hadoop cluster, and used to help clinicians administer more targeted treatments.

This webinar describes the real business value that Johns Hopkins has been gaining from its data, and how the lab was able to put these solutions into place.

Be sure to check out the webinar on-demand here.

For more about our customers, visit: https://hortonworks.com/customers/

 

답변을 남기십시오

귀하의 이메일 주소는 공개되지 않을 것입니다. 필수 내용은 *로 표시되어 있습니다.