Zortech Solutions – Data Architect-Canada – Toronto, ON

Company: Zortech Solutions

Location: Toronto, ON

Expected salary:

Job date: Tue, 21 Jan 2025 23:15:06 GMT

Job description: Role: Data ArchitectLocation: Remote-CanadaDuration: 6-12+ MonthsJob Description:QualificationsWhat you bring:

  • Build Data pipelines required for optimal extraction, anonymization, and transformation of data from a wide variety of data sources using SQL, NoSQL and AWS ‘big data’ technologies.
  • Streaming
  • Batch
  • Work with stakeholders including the Product Owners, Developers and Data scientists to assist
  • with data-related technical issues and support their data infrastructure needs.
  • Ensure that data is secure and separated following corporate compliance and data governance
  • policies
  • Take ownership of existing ETL scripts, maintain and rewrite them in modern data
  • transformation tools whenever needed.
  • Being an automation advocate for data transformation, cleaning and reporting tools.
  • You are proficient in developing software from idea to production
  • You can write automated test suites for your preferred language
  • You have frontend development experience with frameworks such as React.js/Angular
  • You have backend development experience building and integrating with REST APIs and Databases using languages such as Java Spring, JavaScript on Node.js, Flask on Python
  • You have experience with cloud-native technologies, such as Cloud Composer, Dataflow, Dataproc, BigQuery, GKE, Cloud run, Docker, Kubernetes, and Terraform
  • You have used cloud platforms such as Google Cloud/AWS for application hosting
  • You have used and understand CI/CD best practices with tools such as GitHub Actions, GCP Cloud Build
  • You have experience with YAML and JSON for configuration
  • You are up-to-date on the latest trends in AI Technology

Great-to-haves

  • 3+ years of experience as a data or software architect
  • 3+ years of experience in SQL and Python
  • 2+ years of experience with ELT/ETL platforms (Airflow, DBT, Apache Beam, PySpark, Airbyte)
  • 2+ years of experience with BI reporting tools (Looker, Metabase, Quicksight, PowerBI, Tableau)
  • Extensive knowledge of the Google Cloud Platform, specifically the Google Kubernetes Engine
  • Experience with GCP cloud data related services ( Dataflow, GCS, Datastream, Data Fusion, Data Application, BigQuery, Data Flow, Data Proc, Dataplex, PubSub, CloudSQL, BigTable)
  • Experience in health industry an asset
  • Expertise in Python, Java
  • Interest in PaLM, LLM usage and LLMOps
  • Familiarity with LangFuse or Backstage plugins or GitHub Actions
  • Strong experience with GitHub beyond source control
  • Familiarity with monitoring, alerts, and logging solutions
  • Join us on this exciting journey to make Generative AI accessible to all and create a positive impact with technology

#L!-CEIPAL