Job Description

Key Responsibilities

  • Collect, clean, and analyze large datasets from multiple sources to support business decisions.
  • Develop and maintain dashboards, reports, and data visualizations to track key performance indicators (KPIs).
  • Use statistical methods to identify trends, correlations, and anomalies in business data.
  • Apply machine learning techniques (supervised, unsupervised, reinforcement learning) to solve business problems.
  • Develop and maintain complex database queries and optimize database performance.
  • Utilize Python libraries such as Pandas, NumPy, Scikit-learn, TensorFlow, and PyTorch for analysis and development of complex ML/DL models .
  • Utilize data visualization libraries such as Matplotlib and Seaborn for advanced visual representation.
  • Develop and maintain efficient SQL queries and optimize database performance as a DB SQL Expert.
  • Build and maintain ETL pipelines using Python (Pandas, PySpark, etc.) to streamline data workflows.
  • Implement data engineering best practices for data processing, transformation, and storage.
  • Work with data science and analytics tools such as Anaconda, Spyder, and Jupyter Notebook for development and experimentation.
  • Present findings and recommendations to stakeholders through reports and presentations.
  • Work with cloud-based AI/ML services (AWS, Azure, GCP) for model deployment and optimization.
  • Work cross-functionally with engineering, product, and business teams to ensure data-driven decision-making.
  • Leverage Natural Language Processing (NLP), computer vision, and other AI techniques where applicable.

Required Skills & Qualifications

  • Bachelor’s or master’s degree in data science, Computer Science, Statistics, Mathematics, or a related field.
  • 5+ years of experience in data analytics or a related field
  • Expertise in building and maintaining ETL pipelines using Python (Pandas, PySpark, etc.)
  • Proficiency in SQL, Python (3.x and above) for data analysis.
  • Experience with data science and analytics tools such as Anaconda, Spyder, and Jupyter Notebook.
  • Experience with data visualization tools such as Tableau, Power BI, or Looker.
  • Experience to develop ML/DL Models using Pandas, NumPy, Scikit-learn, TensorFlow, PyTorch etc.
  • Develop and fine-tune predictive models using Python, R, or other statistical tools.
  • Work on feature engineering, model selection, and hyperparameter tuning to improve accuracy.
  • Deploy, monitor, and retrain ML models as necessary to ensure continued performance.
  • Experience working with relational and non-relational databases such as MySQL, PostgreSQL, MongoDB, or Snowflake.
  • Expertise in data visualization using Matplotlib ,Seaborn , Google Data Studio etc.
  • Excellent communication skills and ability to present data-driven insights to stakeholders.
  • Strong experience in statistical analysis, hypothesis testing, and A/B testing.
  • Experience with big data technologies (Spark, Hadoop) and cloud platforms (AWS,Azure,GCP).

(ref:hirist.tech)