I am Abhijeet Karmakar, a dedicated software and data engineer with experience in developing and maintaining scalable solutions. My expertise lies in creating robust data pipelines, software development and leveraging cloud technologies to manage and process large datasets. I hold a Master's degree in Electrical and Computer Engineering from the University of Calgary, where I graduated with a GPA of 3.88. Throughout my career, I have demonstrated strong problem-solving skills and a collaborative team spirit. My technical proficiency includes extensive experience with Django, Airflow, PostgreSQL, PySpark, GCP, Apache Superset, Tableau, and Azure. I am also well-versed in code quality, agile methodologies, version control, testing, CI/CD, design patterns, code reviews, and documentation.
I have been leading the development of a cutting-edge micro-learning platform designed for the energy community. My role involves building a dynamic and publicly accessible platform using MongoDB, Express, ReactJS, and NodeJS, while prioritizing software security by addressing and resolving code vulnerabilities. I also enforce Test Driven Development (TDD) practices using Jest and leverage OpenAPI for designing and documenting backend API specifications, ensuring robust and scalable solutions.
Led the development of scalable microservices architecture using Java Spring and Maven, significantly enhancing application performance and scalability. Utilized Docker to create multi-stage container images for efficient deployment on Kubernetes within cloud environments. Engineered robust MySQL database designs with encryption, ensuring data security and optimal performance. Designed and implemented event-driven architectures on AWS, boosting system reliability and responsiveness. Additionally, integrated OpenAPI with Spring Boot, automating API documentation and improving overall development efficiency.
At Untapped Energy, I play a pivotal role in orchestrating the high-profile Datathon event in Calgary, showcasing my exceptional teamwork and unwavering dedication. I am at the forefront of developing dynamic Python notebooks and PowerBI dashboards, transforming intricate datasets into accessible and user-friendly tools for event participants. My work not only simplifies data complexity but also enhances the overall experience, sparking curiosity and enthusiasm among attendees
As a Graduate Teaching Assistant at the University of Calgary, I facilitated engaging lab discussions in Practical Data Structures and Algorithms, fostering a collaborative and inclusive learning environment. I supported student progress through effective communication and prompt responses, ensuring that each student received the guidance they needed to succeed. My dedication to promoting participation and inclusivity made a significant impact on the students' academic journey.
During my tenure at Tata Consultancy Services, I automated territory assignments with batch data pipelines using Python and SQL, significantly improving efficiency. I enhanced sales assignment accuracy by analyzing territory trends and reduced processing time by 70% by converting Apex code for Salesforce CRM into data pipelines on Heroku Cloud. By contributing to design discussions and all lifecycle stages, I enhanced team productivity by 20%. I also managed client data challenges, increasing decision-making efficiency by 30%.
During my tenure at Tata Consultancy Services, I tackled business challenges to enhance client information assessment and decision-making. I designed and implemented a robust backend system for efficient data processing and compliance, integrating advanced functionality for handling sensitive client information. My proactive problem-solving skills led to the resolution of 15 production issues, ensuring timely support for client success. By improving sales operations, I increased efficiency by 30% and expanded the system’s reach into the LATAM region, growing the user base by 20%.
I secured:
I collaborated with a six-member team to develop an innovative reservation-based parking system. My role involved designing batch data pipelines to create a user-friendly analytics dashboard, utilizing core data engineering principles such as data modeling, ETL, and warehousing. I also deployed REST APIs built with Django Rest Framework on AWS, ensuring robust and scalable cloud integration. The project leveraged technologies including Python, MySQL, DRF, and AWS.
View ProjectI improved data processing speed by 50% by implementing scalable pipelines on Google Cloud Platform for Uber Eats data. This was achieved using Google Cloud Storage, Airflow, and Python for efficient data transformation. Additionally, I increased data accessibility by 40% by leveraging Google BigQuery and Apache Superset to create intuitive dashboards. The technologies used in this project included Google Cloud Storage, Python, SQL, Airflow, Google BigQuery, and Apache Superset.
View ProjectI designed and implemented a data pipeline using Python and Apache Airflow. This involved developing real-time data ingestion with Google Pub/Sub and processing the data with Apache Beam for scalable handling. The processed data was then stored in BigQuery, enabling high-speed querying and data analysis. Additionally, I created Tableau dashboards for visualizing stock trends. The technologies utilized in this project included Python, Apache Airflow, Google Pub/Sub, Apache Beam, Google BigQuery, and Tableau.
View ProjectI am currently developing a user-friendly platform for seamless clock-in/out and work hour tracking, utilizing ReactJS, Python Fast API, PostgreSQL, and Redis for secure authentication and real-time updates. The project aims to deliver a robust shift management system, reducing administrative overhead and improving scheduling transparency.
View Project Frontend