One stop solution to your remote job hunt!

By signing up you get access to highly customizable remote jobs newsletter, An app which helps you in your job hunt by providing you all the necessary tools.

OR
Subscribe to our highly customizable newsletter to get remote jobs from top remote job boards delivered to your inbox.
mod.io about 2 years ago
australiadata
Apply Now

mod.io is a fast-growing, early-stage startup backed by leading gaming venture capital firms, working to bring user-generated content (UGC) to games and their players by offering a ready-to-go digital logistics solution that includes community tools, a web UI, and a REST API for any developer to integrate into their game.

The mod.io platform supports over 550,000 daily active users, and we are seeking a Data Engineer to help us scale the industry's best solution for exploring and installing UGC that will be embedded in some of the largest games in the world.

The mod.io service is centered around the REST API which allows developers to bring the modding and UGC experience into a game's UI rather than requiring players to run an external tool or integrate mods manually.

Why mod.io

The popularity of in-game content is exploding, it's in more games and on more platforms than ever before. We have experienced up to 20x growth in our key metrics, and have a database with over a billion rows, and plans to grow even larger.

So if you enjoy solving scaling challenges for a company expecting significant growth, and you want to work at a company and in an industry where data and data insights matter, then mod.io is the place for you.

Requirements

We are looking for a savvy Data Engineer to join our growing team. You will be responsible for expanding and optimizing our data and data pipeline architecture, as well as optimizing data flow and collection across 1,000s of data inputs.

The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up. The Data Engineer is also a hands-on software developer, able to write and modify the code to capture relevant data, massage it into meaningful insights and make these insights available via APIs. You will also ensure we have an optimal data delivery architecture that is consistent throughout all our projects.

You must be self-directed and comfortable supporting the data needs of multiple project teams.

You will be excited by the prospect of designing and optimizing our company’s data architecture to support our next generation of growth, products and data initiatives.

 

Your Responsibilities:

  • Understand mod.io’s data needs and objectives
  • Provide architectural and implementation leadership for data architectures, data warehouses, data lakes and other cloud-related data initiatives based on mod.io’s needs and objectives
  • Build and maintain an optimal data ingestion & ETL solution
  • Assemble large, complex data sets that meet functional and non-functional business requirements
  • Explore and implement ways to enhance data quality and reliability
  • Identify opportunities for (relevant) data acquisition
  • Consider legal and regulatory obligations in regard to data capture and usage
  • Identify, design, and implement internal process improvements: automate manual processes, optimize data delivery, (re)design infrastructure for greater scalability, etc.
  • Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and AWS ‘big data’ technologies
  • Build or implement BI and query/reporting tools that utilize the data pipeline to provide actionable insights for key business performance metrics for presentation internally and externally (eg: customer dashboards etc)
  • Work with project teams to assist with data-related technical issues and support their data infrastructure needs.
  • Keep our data separated and secure across international boundaries through multiple data centers and AWS regions.
  • Provide data and analytics expertise as we strive for greater functionality in our data systems.
  • In relation to our data pipeline architecture:
    • Deploy and maintain production environments that require high availability.
    • Monitor data servers to proactively identify performance issues, problematic trends and troubleshoot/escalate to resolve as appropriate.
    • Drive the product towards higher availability and reliability & assist with on-call support on a rotating schedule for incident escalations (24x7).
    • Ensure our data services meet stability, performance, and availability requirements.
    • Monitor backups, usage, capacity, and performance of servers; liaise with users and/or vendors to address problems and changes in requirements.
    • Build robust, self-healing features and automation that reduce operational effort and improve service up-time.
  • Self-starter mindset with a strong drive to learn and own engineering initiatives to promote a culture of continuous improvement, and engineering excellence.

 

Qualifications

  • Extensive experience working with relational databases, query authoring (SQL) and designing database backup and replication approaches
  • Familiarity with a variety of database technologies
  • Experience building and optimizing ‘big data’ data pipelines, architectures and data sets.
  • Advanced SQL knowledge
  • Expert ability to tune databases and optimise query performance
  • Extensive experience in our core database technologies: MySQL, Redshift (PostgreSQL) and Memcached
  • Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
  • Strong analytic skills related to working with unstructured datasets.
  • Build processes supporting data transformation, data structures, metadata, dependency and workload management.
  • A successful history of manipulating, processing and extracting value from large disconnected datasets.
  • Working knowledge of message queuing, stream processing, and highly scalable ‘big data’ data stores.
  • Strong project management and organizational skills.
  • Experience supporting and working with cross-functional teams in a dynamic environment.
  • Experience with big data tools: Hadoop, Spark, Kafka, etc.
  • Experience with relational SQL and NoSQL databases, including Postgres and MySQL
  • Experience with data pipeline and workflow management tools: Azkaban, Luigi, Airflow, etc.
  • Experience with AWS cloud services: EC2, EMR, RDS, Redshift
  • Experience with stream-processing systems: Storm, Spark-Streaming, etc.
  • Experience with object-oriented/object function scripting languages: Python, Java, C++, Scala, etc.
  • Experience in one of the following Scripting languages: Python, PowerShell, Bash, Shell Script.
  • Experience with monitoring and logging services (e.g. Elasticsearch, Wavefront, Uptime, Solarwinds, or similar).

It would be awesome if you also have:

  • Data engineering certification (e.g AWS Certified Data Engineer)
  • Experience with cloud-based infrastructure and services (AWS)
  • Experience working in an agile environment
  • Experience with Jenkins or similar build automation tools.
  • Experience with Machine learning and AI over large data sets
  • Experience with Trello and G-Suite
  • A passion for video games.

Benefits

  • Remote working is actively supported.
  • Competitive salary plus equity.
  • Remote working is actively supported.
  • Flexible working hours and family-friendly considerations.
  • Sit-stand desks, 27” monitor, ergonomic chairs.
  • Regular social events.
  • Experience new games, digital and tabletop.
  • Attend international gaming conferences.
  • Contributing to open-source on Github