About me

Hey! I’m Jose Andres 🚀

B.S in Computer Science. I work in Data Engineering at Arkose Labs, a cybersecurity startup focused on bot management and fraud prevention, where I now own the end-to-end of data workflows, building scalable pipelines and data models that transform event and raw data into insights and business value. I work very closely with business intelligence products and reports that serve enterprise customers like Microsoft, Adobe, Meta, and Sony.

I believe that strong teamwork and open communication lead to the best solutions. I enjoy collaborating across teams, sharing knowledge and supporting others in their growth. Seeing the impact of good collaboration is something I truly value.

I’m genuinely excited about the future of tech, and AI especially. I enjoy connecting the dots, testing, iterating, building. I’m endlessly curious about these topics, and as much as I enjoy learning, I value sharing what I learn: helping teammates create better Jira tickets, manage their Asana to-do lists, keep documentation clean in Confluence, and build sharper visualizations for decks and custom reports. Sharing knowledge is the best way I’ve found to solidify it, and the curiosity it sparks is what keeps me learning and feeling really optimistic about what’s ahead.

What I Do

  • 🗄️

    End-to-End Pipelines

    I design and build pipelines that move data from source systems to gold tables ready for analytics, with reliability and clear monitoring.

  • 📊

    BI & Reporting

    I build the data layer that powers BI platforms, turning raw events into the insights teams use to drive business value.

  • 🤖

    AI Integration

    Bringing LLMs and agentic workflows into data and analytics workflows to automate, accelerate, and personalize at scale.

  • 🧠

    Knowledge Sharing

    Running upskilling programs, leading study cohorts, and helping teammates level up their data, security, and AI skills.

Currently Building & Exploring

  • ❄️

    Migrating to Snowflake

    Leading the migration of ~70 tables from Athena/Postgres to Snowflake, scoping dependencies, redesigning models, and shaping the next chapter of our data platform.

  • Bringing AI Into Analytics

    Integrating Claude and agentic workflows into our reporting stack to deliver hyper-personalized customer insights at scale.

  • 🎓

    Always Learning

    Exploring agents, Claude Code, and the tools shaping the future of data and analytics. Sharing what I learn with my team along the way.

Resume

Education

  1. Universidad Cenfotec

    2024 — 2026

    Master's Degree in Databases and Analytics

    In Progress

    Balancing work, a Master's Degree and personal life has been a challenge, some days I win, some days I just try again tomorrow.

  2. Unversidad Fidélitas

    2018 — 2023

    Bachelor's Degree in Systems Computer Enginnering/Computer Science.

    Click here to see the Degree!

Experience

  1. Associate Data Engineer

    Arkose Labs · Escazú, Costa Rica

    01/2025 — Present 

    As a Data Engineer in the Customer Success team, I own the data backend that turns raw event data into clean datasets powering an automated BI reporting platform built in Looker Studio. CS teams use it to present value to enterprise customers, and executives use it to monitor portfolio health and renewals. I drive cross-team technical alignment with Professional Services, Account Management, and Solutions Engineering to maintain a single source of truth across reporting use cases, shared definitions, and KPI calculations. I lead the planning and scoping for incoming requests, prioritizing what delivers the most business value and turning it into production-ready datasets, fast.

    Key achievements

    • Owned and scaled the CS BI platform data backend, the source of truth for product usage and value, adoption and engagement, and customer portfolio intelligence and health. The platform centralizes data from multiple sources (product event data lakes, Zendesk, Gainsight, Pylon, Slack, Jira, Harvest) into a unified PostgreSQL warehouse through dedicated pipelines, serving CS, executives, and stakeholder teams.
    • Cut manual reporting time across the team by optimizing queries and schedules; the BI platform now refreshes twice daily (up from once):
      • QBRs: from 2 hours to 10 minutes per customer
      • Monthly Security Reports: from 2 hours to a fully automated Slack-delivered report
      • Post-attack reports: from 2–3 hours to 10 minutes per incident
      • Threat Intel reports for Marketing: from 6–8 hours of manual exports to 30 minutes of centralized data. Recent examples: Threat Actor Patterns Q2 2025 and 2026 Agentic AI Security Report
    • Leading my team's AI adoption strategy, primarily on the Claude stack, defining how AI integrates into our current tools and processes. The vision: build intelligence layers that produce hyper-personalized, standardized customer reports that today require heavy manual customization. Early wins are already showing in daily tasks (ticket creation, to-do lists, documentation, coding). I'm excited to lead this transformation and turn frontier technology into real business value. As part of this, I designed and ran an 8-lesson AI upskilling program for a 6-person team covering LLMs, how tokens work, prompting, RAG, evaluation, and agentic workflows, delivered with slide decks and hands-on labs.
    • Led end-to-end data engineering for two churn and contraction prevention initiatives, from source system discovery and API integrations to transforming raw → clean → gold datasets in PostgreSQL:
      • Customer Health Score & Quadrant Classification: blends adoption and engagement signals across 20+ metrics into a standardized health score, with custom recommendations sent up to 6 months before renewal so teams can act on risks and opportunities early.
    • Planned, scoped, and executed the migration of Python ETL code from ECS Fargate/ECR to Airflow (Amazon MWAA), achieving more reliable runs, better isolation of data processing, stronger QA checks, and less manual work. Designed a phased migration plan with full pipeline inventory and dependencies. In Airflow, DAG pipelines run under one scheduler with clear retries and simple monitoring; heavy jobs were split into smaller tasks to isolate failures. 7 pipelines migrated successfully.
    • Currently leading the migration of the Customer Success data backend to Snowflake (~70 tables): scoping the migration plan, mapping dependencies, redesigning data models, and shaping the next chapter of our CS data platform. Aligning with the broader Data team that owns the company-wide migration to ensure consistency in patterns and standards.
    • Launched a Security+ certification cohort for my team: scoped phases, led weekly study sessions, gathered materials, and scheduled practice tests to prepare teammates for certification. Sharing knowledge and watching teammates grow their careers is one of the parts of this work I enjoy most.
  2. Data Engineer Intern

    Arkose Labs · Escazú, Costa Rica · 04/2024 — 01/2025

    I worked in a fast-paced and iterative startup environment where adaptability and problem-solving were critical. I was given significant responsibility early on and exposed to real business challenges, allowing me to grow quickly. This was an invaluable learning experience that prepared me for delivering data solutions with real impact.

Skills

Data Engineering

  • Python Python
  • PostgreSQL SQL
  • Airflow Airflow
  • Docker Docker
  • Git Git
  • Terraform Terraform
  • Linux Linux

Cloud & Data Platforms

  • AWS AWS
  • Databricks Databricks
  • Snowflake Snowflake

AWS: S3 · Glue · Athena · RDS · ECS · CloudWatch · EventBridge

Databricks: Delta Lake · Spark SQL · Lakeflow Jobs · Lakeflow Spark Declarative Pipelines · Lakeflow Connect · Unity Catalog · Auto Loader

Snowflake: Architecture · Snowpipe · Streams & Tasks · Dynamic Tables · Time Travel · Iceberg Tables · Cortex

AI & LLMs

  • Claude Claude
  • Claude Code Claude Code
  • RAG RAG
  • 🤖 Agents

Collaboration & Tools

  • Jira Jira
  • Confluence Confluence
  • Slack Slack
  • Salesforce Salesforce
  • Zendesk Zendesk
  • Looker Studio Looker Studio
  • Gainsight Gainsight Gainsight
  • Pylon Pylon

Certifications

I enjoy taking certifications because they give me a structured way to learn while keeping things interesting :)

Portfolio

Here is where my portfolio would be... if I had one!

Jokes aside, one of my personal goals is to start adding projects here as I explore new technologies. Hope to begin soon! 🚀