Data Scientist 3

Hewlett Packard EnterpriseSunnyvale, CA
1dOnsite

About The Position

Data Scientist 3 This role has been designed as ‘’Onsite’ with an expectation that you will primarily work from an HPE office. Who We Are: Hewlett Packard Enterprise is the global edge-to-cloud company advancing the way people live and work. We help companies connect, protect, analyze, and act on their data and applications wherever they live, from edge to cloud, so they can turn insights into outcomes at the speed required to thrive in today’s complex world. Our culture thrives on finding new and better ways to accelerate what’s next. We know varied backgrounds are valued and succeed here. We have the flexibility to manage our work and personal needs. We make bold moves, together, and are a force for good. If you are looking to stretch and grow your career our culture will embrace you. Open up opportunities with HPE. Job Description: Data Scientist 3 Location (3 days a week): Sunnyvale, CA Data Scientist/Software Engineer will be engaged in data science-related research and software application development and engineering duties related to AI Datacenter technology and autonomous platform to provide an unprecedented visibility and operational efficiency into the user experience. The Data scientist will collaborate with other engineers and data scientists to build the next generation of autonomous Datacenter networks leveraging big data and predictive models. The Software Engineer/Data Scientist will leverage the data collected from the network to empower the inference engine of our Mist based cloud platform. In addition, Data Scientist will use his/her knowledge of network communication, machine learning and software engineering to develop and implement scalable algorithms to process a large amount of streaming data to detect anomalies, predict problems, and classify them in real-time. He/She will also be responsible to develop the software and algorithms to enhance the cloud intelligent for Marvis and Apstra Cloud Services for Datacenter.

Requirements

  • Solid statistics and math background, good knowledge of machine learning methods like k-Nearest Neighbors, Naive Bayes, SVM, Decision Forests.
  • Excellent Communication Skills to articulate observations and use cases with PM and network domain experts who are not experienced in AI/ML through data visualization tool.
  • Have done time series data analysis, forecasting and correlation is preferrable.
  • Have utilized latest AI/ML techniques, such as Neural Networks, Transformer, etc. for time series data or interested to explore these techniques for time series data.
  • Require the knowledge of the multi-cloud production environment
  • Require the agility to troubleshoot open-source data processing engine, such as Apache Spark, Apache Storm and Apache Flink
  • Require good knowledge and experience of the big data tool sets and techniques of distributed storage and computation engine
  • Require the experience to develop the reusable and highly scalable data processing component
  • Require good knowledge and experience to work with cloud based CICD tools and cloud dev-ops teams to collect stats and create monitors for our data processing pipelines
  • Bachelor's degree in Computer Science/Engineering/Mathematics or equivalent experience
  • 5+ years of experience Search Indexing, Ranking, Information Retrieval and Querying.
  • Proficient in Python and Golang
  • Proficient in implementing NLP, Machine Learning models and algorithms into production at scale.

Nice To Haves

  • PhD or MS in Computer Science, Applied Mathematics, Physics or related quantitative field
  • 5+ years' experience with statistical data analysis, data mining, and querying.
  • Experience in deploying and leading ML platforms in AWS/GCP/Azure.
  • Cloud Architectures
  • Cross Domain Knowledge
  • Design Thinking
  • Development Fundamentals
  • DevOps
  • Distributed Computing
  • Microservices Fluency
  • Full Stack Development
  • Security-First Mindset
  • Solutions Design
  • Testing & Automation
  • User Experience (UX)

Responsibilities

  • Design and implement machine learning solutions which require to process terabytes of streaming data to detect anomalies in DC networks of our customers, predict problems and future trends, classify them in real-time (60%)
  • Analyze feature requirements from product manager, collaborate with engineers and data scientists to design the solutions.
  • Troubleshoot production environment and customer reported issues (20%)
  • Utilize analytical and programming skills and open-source systems, such as Hadoop, Hive, Spark, Elasticsearch, Redis, etc. develop data processing pipeline required efficacy and latency (20%)

Benefits

  • Health & Wellbeing We strive to provide our team members and their loved ones with a comprehensive suite of benefits that supports their physical, financial and emotional wellbeing.
  • Personal & Professional Development We also invest in your career because the better you are, the better we all are. We have specific programs catered to helping you reach any career goals you have — whether you want to become a knowledge expert in your field or apply your skills to another division.
  • Unconditional Inclusion We are unconditionally inclusive in the way we work and celebrate individual uniqueness. We know varied backgrounds are valued and succeed here. We have the flexibility to manage our work and personal needs. We make bold moves, together, and are a force for good.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service