Job Description
Our Purpose
P\S\L Group is a global organisation dedicated to putting information at the service of medicine. The companies and people of the P\S\L Group aim to improve medical care by serving those who need it, those who provide it and those who seek to improve it.
Our primary purpose is to help clients increase the effectiveness of activities pertaining to scientific communication, medical education and product/service marketing. To this end, we want our information services to contribute to the goals we share with our clients, namely: to accelerate the advancement of medicine and help people enjoy better, longer lives.
Objective
If you are a Sr. Data Engineer with a craving for making sense out of structured and unstructured data with the goal of affecting people's lives in a positive manner, please read on!
We are looking for a Data Engineer that will work on collecting, storing, processing, and analyzing huge sets of data. The focus will be on working with the Data Engineering Team to design technologies that will wrangle, standardize and enhance our master data and transactional data repositories, then build operational and monitoring processes to govern that data. You will also be responsible for federation of this data across the enterprise using batch, streaming and microservices architectures.
Unique skills expected for this job are the ability to write clean, high-quality Python libraries that can be re-used within our platform; ability to create orchestration workflows that ingest structured and unstructured data in both streaming and batch modes; enrich and make it available for use throughout the enterprise.
What you will do
- Build the infrastructure required for optimal data pipelines from a wide variety of data sources using Python, AWS services and big data tech
- Create and maintain enterprise-wide data pipelines leveraging Kinesis, Glue, Lambda, and general microservices\microbatch architecture best practices
- Manage databases running on PostgreSQL, Snowflake, Redshift and ElasticSearch
- Monitor performance using Cloudwatch, Cloudtrail and advise on necessary infrastructure changes as needed
- Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, redesigning for greater scalability, etc.
- Create data tools for analytics and data scientist team members that assist them in building and optimizing our enterprise data hub into an innovative industry leader.
Who you are
- Minimum of 4 years experience implementing production systems in the cloud (preferably AWS)
- Understanding of database design (both SQL and noSQL)
- Experience with object-oriented/object function scripting languages: Python
- Excellent analytical and problem solving skills
- Experience with data cleansing, data wrangling, data quality, standardization, transformations etc
- Experience with data pipeline and workflow management tools: Streamsets
- Experience with relational SQL, including PostgreSQL, and MSSQL
- Experience with build systems: github, bitbucket
- Advanced working SQL knowledge and experience working with relational databases - both operational DBs and data warehouses
- Strong analytic skills related to working with unstructured datasets
- Prior Experience with Master Data Management is a plus
- BS/MS in Math, Computer Science, or equivalent experience
Visit Original Source:
http://ca.indeed.com/viewjob