Data Engineer

Data Engineering - Chicago, IL

Company Description

rMark Bio helps life science companies solve for the complexities that come with digital transformation by developing end-to-end AI solutions that deliver personalized business intelligence through integrated applications and API accessible services.

Healthcare innovation is best served when individuals with diverse backgrounds come together with a common purpose and clear objectives to improve patient lives.

We are product strategists, engineers, data scientists and designers who are experts in our domain and passionate about our mission to accelerate innovation, collaboration and scientific discovery for life sciences

Job Description

rMark Bio, is searching for an experienced Data Engineer to work within an agile team of data scientists, software architects and developers to implement and maintain secure, scalable, cloud-based software solutions, build customer integration solutions, and streamline the team’s software delivery tools and processes. Technical aptitude is a must as well as the right team-minded attitude and the ability to work interchangeably with others. The core team is as a small SWAT-style team with everyone pulling their own weight, playing a variety of roles, and covering each others’ responsibilities when needed.  Applicants should be qualified in collecting, cleaning, and maintaining large datasets that are critical to customer and product success.

Job Responsibilities

  • Build automated ETL pipelines for cloud environments including, AWS, GCP, Azure, and Heroku.
  • Develop and support data integrity reporting and alerting for ETL pipelines.
  • Ability to work closely with Engineering, Product and Customer Success Teams. 
  • Build robust and deployable software in Python, C++ or C#.
  • Build, deploy, and maintain RESTful APIs to access datasets.
  • Parse and extract data from common formats including, XML, JSON, CSV, and Pipe delimited.
  • Integrate with customer provided APIs.
  • Build, organize, and maintain datamarts using any of SQL, JSON, Blob, or other databases as needed.  
  • Write and maintain excellent documentation of all work.

Experience and Qualifications

  • Minimum 3-5 years of experience is required
  • B.S. in Computer Science or closely related field preferred, but not required. Real-world experience and proven track records count as much, if not more.
  • Programming Languages: Python and C# (C++ and R are a plus).
  • Experience with databricks is a plus.
  • Established expertise developing ETL pipelines on serverless cloud solutions. (AWS Lambda, Azure App Services, etc..) 
  • Linux/Unix
  • Docker
  • Databases: SQL, JSON, object storage, etc.  Experience with graph databases like Neo4j and TigerGraph is a plus.
  • Strong technical writing skills will be heavily stressed


  • All candidates must successfully complete a background check
  • Candidates must be US Citizen or US Permanent Resident
  • Please e-mail resume, cover letter, and include job title applied for in subject line, to: