P\S\L Group is a global organization dedicated to putting information at the service of medicine. The companies and people of the P\S\L Group aim to improve medical care by serving those who need it, those who provide it, and those who seek to improve it.
Our primary purpose is to help clients increase the effectiveness of activities pertaining to scientific communication, medical education, and product/service marketing. To this end, we want our information services to contribute to the goals we share with our clients, namely: to accelerate the advancement of medicine and help people enjoy better, longer lives.
Position Summary
As a Data Engineer, you will be a key resource for creating and supporting organizational data engineering efforts. This role will work closely with the engineering teams to proactively maintain and help create a high-quality data ecosystem. This role will require technical expertise to drive innovation and evolve our data integration capabilities.
You will work on collecting, storing, processing, and analyzing huge sets of data. The focus will be on working with the Data Management Team to design technologies that will wrangle, standardize, and enhance our master data and transactional data repositories, then build operational and monitoring processes to govern that data.
What you will do
- Build the infrastructure required for optimal data pipelines from a wide variety of data sources using Python, AWS services, and Snowflake.
- Create and maintain enterprise-wide data pipelines leveraging Fargate, Kinesis, Glue, Lambda, and general microservices/microbatch architecture best practices.
- Manage databases running on PostgreSQL, Snowflake, Redshift, and ElasticSearch.
- Monitor performance using Cloudwatch, Cloudtrail, and advise on necessary infrastructure changes as needed.
- Support Data Quality and data completeness by implementing monitoring dashboards for business-critical data.
- Identify problems in data assets and take appropriate corrective action to fix within defined SLAs.
Who you are
- Minimum of 3 years experience implementing production systems in the cloud (preferably AWS/Snowflake).
- Understanding of database design (both SQL and NoSQL).
- Experience with object-oriented/object function scripting languages: Python.
- Excellent analytical and problem-solving skills.
- Experience with data cleansing, data wrangling, data quality, standardization, transformations, etc.
- Experience with data pipeline and workflow management tools.
- Experience with relational SQL, including PostgreSQL and MSSQL.
- Experience with build systems: GitHub, Bitbucket.
- Advanced working SQL knowledge and experience working with relational databases - both operational DBs and data warehouses.
- Strong analytic skills related to working with unstructured datasets.
- Prior Experience with Master Data Management is a plus.
- BS/MS in Math, Computer Science, or equivalent experience.