Xgrid is a cloud services tech firm, specializing in providing expert consultancy and professional services in building secure, complex, large-scale, production grade distributed systems in cloud environments (AWS Google Cloud Platform Azure).
The architects of the Xgrid cloud business group have 20+ years of combined experience building production-grade hyperscale systems for top silicon valley tech companies including Amazon, Salesforce, VMware, Cisco, and multiple top startups with successful exits.
Our mission is to help small and big businesses grow their business by equipping them with state-of-the-art tools and technologies.
This not only helps their business but also modernizes and builds a robust and cutting-edge cloud tech stack satisfying all the requirements for their use cases.
We are looking for passionate data engineers who have an understanding of building, maintaining, and optimizing data pipelines in cloud environments (AWS GCP Azure).
Our cloud business group works with top-tier silicon valley and European customers and helps them build production grade systems using cutting-edge cloud software technology.
At Xgrid, you will enjoy being part of an extremely talented, motivated team interacting with both business and development counterparts to capture key marketing requirements and translate them to products and services designed for our customers.
The ideal Xgrid Cloud Data Engineer will have experience developing and delivering complex enterprise-grade mission-critical automated software systems for our cloud customers.
If you are passionate about working with the world’s top tech silicon valley companies and highly experienced software engineers and are excited about solving interesting and complex problems and challenges in a fast-paced environment, where you get to not only grow yourself but also help build a team and culture, Xgrid has a place for you and we would love to have you onboard!
What you will do
Build data pipelines ingesting and integrating datasets from multiple data sources, while designing and developing solutions for data integration, data modeling, and data inference and insights.
Design, monitor, automate, and improve development, test, and production infrastructure for data pipelines and data stores.
Troubleshoot and performance tune data pipelines and processes for data ingestion, merging, and integration across multiple technologies and architectures including ETL, ELT, API, and SQL.
Build pipelines and solutions on multiple cloud platforms (AWS GCP Azure)
Work with highly experienced silicon valley engineers
Experience with public cloud data services (AWS GCP Azure)
At least 3 Plus years of Experience in dealing with BigData on cloud (AWS GCP Azure)
Knowledge of building data pipelines on AWS (using Glue, Lambda, S3, Athena, Kinesis, etc.) or GCP (Big Query, Big Table, Google Cloud Storage, PubSub, Data Fusion, Dataflow) or Azure ( Azure Data Factory, Data Lake Analytics, Data Bricks, Stream Analytics )
Solid understanding and hands-on experience of Data Engineering principles Data warehousing, ETL process, SQL, and handling data in JSON and other semi-structured formats.
Understanding of object-oriented programming, and proficiency and experience with at least one object-oriented programming language (Go, Python, C++, Java) (Python and Go preferred)
Good knowledge of operating systems (Linux power user)
Thorough knowledge of software development best practices, including coding standards, code reviews, source control management, build process, continuous integration and continuous delivery.
Strong skills in information gathering and requirement documentation.
Quick learner and ability to adapt to customer-driven fast-paced development environments.
Aptitude to learn new bleeding-edge technologies.
Team player with outstanding collaboration and teamwork attitude.
Excellent written and verbal communication skills.
Excellent analytical and problem-solving skills.
Hands-on experience with building large scale, complex distributed data processing systems and data pipelines
Solid understanding of SQL (Postgres, MySQL, RDS, etc.) and NoSQL (Mongo, Elasticsearch, Cassandra, DynamoDB, etc.) technologies
Experience with open source data processing and workflow orchestration technologies like Airflow, Apache Spark, Apache Beam, etc.
Any cloud (AWS GCP Azure) certifications