Company

Laboratory Technician | St. Cloud ShinesSee more

addressAddressMinneapolis, MN
type Form of workContractor
CategoryInformation Technology

Job description

Preference is San Francisco or Cupertino Office in Hybrid Model. Remote OK for strong candidate with good skills below and great communication skills.
Please do strong Tech Screening Video and send with Submittal

Job Title: Data Engineer
Location: Cupertino,CA
Duration: 6 Months Contract
  • MLOps / Client Engineering => 8/10 => 6 to 8 experience
  • Platform Development / MicroServices / Arch => 7/10 => 8 to 10 experience
  • Docker/Containers/Kubernetes => 6/10 => 5 to 6 experience
  • Data Science / Machine Learning => 5/10 => 5 to 6 experience
  • Azure - Highly preferred to have the experience => 6 to 8 experience
  • Python - must have => 8 to 10 experience
  • Spark- Required => 4 to 5 experience
  • Client tools experience such as AzureML/MLFlow/Databricks/Kubeflow etc. - Deployed & worked on some of these tools

Kubernetes - Very Strong and #1 - 4 to 5 years
Data Pipelines - ETL Preferred - Bring data to send back to other team
Understanding of Python is good and will code in Python - Not working on API's
Good Understanding of Machine Learning Pipelines
Argo WorkFlow Experience
Docker and Jenkins - Would be good
Workflow Experience - Would be good
Job Description: Summary of the project/initiatives which describes what's being done:
o Build, modernize and maintain the U.S. Bank AI/Client Platform & related frameworks / solutions.
o Participate and contribute in architecture & design reviews.
o Build/Deploy AI/Client platform in Azure with open-source applications (Argo, Jupyter Hub/Kubeflow) and/or cloud/SaaS solutions (Azure Client, Databricks).
o You will design, develop, test, deploy, and maintain distributed & GPU-enabled Machine Learning Pipelines using K8s/AKS based Argo Workflow Orchestration solutions, while collaborating with Data Scientists.
o Enable/Support platform to do distributed data processing using Apache Spark and other distributed / scale technologies.
o Build ETL pipelines, ingress / egress methodologies in context to AIML use-cases.
o Build highly scalable backend REST APIs for metadata management and other misc. business needs.
o Deploy Application in Azure Kubernetes Service using GitLab, Jenkins, Docker, Kubectl, Helm and Manifest
o Experience in branching, tagging, and maintaining the versions across different environments in GitLab.
o Review code developed by other developers and provide feedback to ensure best practices (e.g., design patterns, accuracy, testability, efficiency etc.)
o Work with relevant engineering, operations, business lines, and infrastructure groups to ensure effective architectures and designs and communicate findings clearly to technical and non-technical partners.
o Perform functional, benchmark & performance testing and tuning to achieve performant AIML workflow(s), interactive notebook user experiences, and pipelines.
o Assess, design & optimize the resources capacities for Client based resource (GPU) intensive workloads.
o Communicate processes and results of the application with all parties involved in the product team, like engineers, product owner, scrum master and third-party vendors.
Top 5-10 responsibilities for this position:
o Experience developing AIML platforms & frameworks (including core offerings such as model training, inferencing, distributed/parallel programming), preferably on Kubernetes and native cloud.
o Highly skilled with Python or JAVA programming languages
o Highly skilled with database languages like SQL & NoSQL
o Experience designing, developing, and deploying highly maintainable, extensible, and testable distributed applications using Python and other languages.
o Experience developing ETL pipelines and REST APIs in Python using Flask or Django
o Experienced with technologies/frameworks including Kubernetes, Helm Charts, Notebooks, Workflow orchestration tools, and CI/CD & monitoring frameworks.
Basic Qualifications:
Bachelor's/master's degree in computer science or data science
6 - 8 years of experience in software development and with data structures/algorithms
Required Technical Qualifications / Skills:
Experience with AI/Client open-source projects in large datasets using Jupyter, Argo, Spark, Pytorch, TensorFlow
Experience creating Unit and Functional test cases using PyTest, UnitTest
Experience with training and tuning models in Machine Learning
Experience working with Jupyter Hub
Experience with DB management system like PostgreSQL
Experience in searching, monitoring, and analyzing logs using Splunk/Kibana
GraphQL/Swagger implementation knowledge
Strong understanding and experience with Kubernetes for availability and scalability of applications in Azure Kubernetes Service
Experience building CI/CD pipelines using Cloudbees Jenkins, Docker, Artifactory, Kubernetes, Helm Charts and Gitlab
Experience with tools like Jupyter Hub, Kubeflow, MLFlow, TensorFlow, Scikit, Apache Spark, Kafka
Experience with workflow orchestration tools such as Apache Airflow, Argo workflows
Familiarity with Conda, PyPi, and Node.js package builds
Preferred Qualifications / Skills:
Experience with AI/Client open-source projects in large datasets using Jupyter, Argo, Spark, Pytorch, TensorFlow
Experience creating Unit and Functional test cases using PyTest, UnitTest
Experience with training and tuning models in Machine Learning
Experience working with Jupyter Hub
Experience with DB management system like PostgreSQL
Experience in searching, monitoring, and analyzing logs using Splunk/Kibana
GraphQL/Swagger implementation knowledge
Strong understanding and experience with Kubernetes for availability and scalability of applications in Azure Kubernetes Service
Experience building CI/CD pipelines using Cloudbees Jenkins, Docker, Artifactory, Kubernetes, Helm Charts and Gitlab
Experience with tools like Jupyter Hub, Kubeflow, MLFlow, TensorFlow, Scikit, Apache Spark, Kafka
Experience with workflow orchestration tools such as Apache Airflow, Argo workflows
Familiarity with Conda, PyPi, and Node.js package builds
------------------------DO NOT EDIT BELOW THIS LINE, PLEASE INSERT JOB DESCRIPTION ABOVE---------------------------------
Is responsible for developing, implementing and maintaining knowledge-based or artificial intelligence application systems. The individual should ensure that information is converted into a format that is digestible and easy for end users to access the information and utilize it optimally.
ESSENTIAL FUNCTIONS:
? Designs and writes complex code in several languages relevant to our existing product stack, with a focus on automation
? Configures, tunes, maintains and installs applications systems and validates system functionality
? Monitors and fine tunes applications system to achieve optimum performance levels and works with hardware teams to resolve issues with hardware and software
? Develops and maintains department's knowledge database containing enterprise issues and possible resolutions.
? Develops models of task problem domain for which a system will be designed or built.
? Uses models, hypotheses, and cognitive analysis techniques to elicit real problem-solving knowledge from the experts
? Mediates between the expert and knowledge base; encodes for the knowledge base
? Acts as subject matter expert for difficult or complex application problems requiring interpretation of AI tools and principles
? Researches and prepares reports and studies on various aspects of knowledge acquisition, modeling, management, and presentation
? Develops and maintains processes, procedures, models, and templates for collecting and organizing knowledge into specialized knowledge representation programs
? Acts as vendor liaison for products and services to support development tools
? Maintains the definition, documentation, training, testing, and activation of Disaster Recovery/Business Continuity Planning to meet compliance standards
? Maintains a comprehensive operating system hardware and software configuration database/library of all supporting documentation to ensure data integrity
? Acts to improve the overall reliability of systems and to increase efficiency
? Works collaboratively with cross functional teams, using Agile / DevOps principles to bring products to life, achieve business objectives and serve customer needs
Comments for Vendors: Remote position
Refer code: 7095387. Laboratory Technician | St. Cloud Shines - The previous day - 2023-12-16 05:55

Laboratory Technician | St. Cloud Shines

Minneapolis, MN
Jobs feed

Vice President, Finance - Navy Business Group

Saic

Virginia, United States

VP, Financial Consultant- Leesburg, VA

Charles Schwab

Leesburg, VA

Staff Accountant

Woodfin-Your Home Team

Richmond, VA

Project Controller

St Engineering Idirect

Herndon, VA

Junior Accountant REMOTE

American Management Group, Llc

Manassas, VA

Grocery Store Manager

Gonzales Village Llc

Gonzales, LA

$39.3K - $49.7K a year

Accounting - Financial Accountant (Sr)

Quadrant Inc

Burke, VA

Project Accountant/Controller

Fusco Personnel Inc.

New York, NY

Senior Accountant

Fanduel

New York, NY

Share jobs with friends

Related jobs

Data Engineer

Data Engineer

Trissential

$120,000 - $132,000 a year

Minneapolis, MN

4 days ago - seen

Data Engineer

Synergis

Minneapolis, MN

4 weeks ago - seen

Senior Data Engineer

Sourcewell

Staples, MN

a month ago - seen

Principal Data Engineer (Remote)

Patterson Companies, Inc.

Saint Paul, MN

a month ago - seen

Sr. Data Engineer

Tekfortune Inc.

Minneapolis, MN

a month ago - seen

sr. data engineer

Skiltrek Llc

Minneapolis, MN

a month ago - seen

Senior Data Engineer

Medica

Hopkins, MN

a month ago - seen

Automation Engineer 3 - (Data Analytics and Automation)

U.s. Bank

Saint Paul, MN

a month ago - seen

Data Engineer

Ucare

$104K - $131K a year

Minneapolis, MN

2 months ago - seen

UI/UX Architect

Data Engineer - Aptask

Minneapolis, MN

2 months ago - seen

Lead Data Engineer

Land O'lakes

Hills, MN

2 months ago - seen

Data Engineer - Generative AI - Remote

Mayo Clinic

$97,884.80 - $137,030.41 a year

Rochester, MN

2 months ago - seen

React Lead Developer

Data Engineer - Aptask

Minneapolis, MN

2 months ago - seen

Java Springboot GraphQL Developer

Data Engineer - Aptask

Minneapolis, MN

2 months ago - seen

Ruby on Rails Developer

Data Engineer - Aptask

Minneapolis, MN

2 months ago - seen

Techno Functional Lead - Conga CPQ

Data Engineer - Aptask

Saint Paul, MN

2 months ago - seen

Senior Data Engineer (Remote)

Patterson

$120,000 - $140,000 a year

Saint Paul, MN

2 months ago - seen

Senior Quality Engineer - Data Scientist

Boston Scientific

Saint Paul, MN

3 months ago - seen