Job description
Your role and responsibilities will include:
- Developing data infrastructure to ingest, sanitize and normalize a broad range of medical data, such as electronics health records, journals, established medical ontologies, crowd-sourced labelling and other human inputs
- Building performant and expressive interfaces to the data
- Creating infrastructure to help us not only scale up data ingest, but large-scale cloud-based machine learning
We are looking for a teammate with:
- 2+ years of development experience in a company/production setting
- Experience building data pipelines from disparate sources
- Hands-on experience building and scaling up compute clusters
- A solid understanding of databases and large-scale data processing frameworks like Hadoop or Spark and the ability to evaluate which tools to use on the job
- A unique combination of creative and analytic skills apt of designing a system capable of pulling together, training, and testing dozens of data sources under a unified ontology
Bonus points if you have:
- Know-how of developing systems to do or support machine learning, including experience working with NLP toolkits like Stanford CoreNLP, OpenNLP, and/or Python's NLTK
- Expertise with wrangling healthcare data and/or HIPAA
- Experience with managing large-scale data labelling and acquisition, through tools such as through Amazon Turk or DeepDive
Salary range:
- $100 000 - $160 000 USD
About Fathom
Revenue: Unknown / Non-Applicable
Size: 1 to 50 Employees
Type: Company - Private
Website: fathomhealth.com