Location: 100% Remote within the United States

 

Job Overview:
As a Data Scientist, you will be responsible for managing the complete Model Development Life Cycle (MDLC), from problem definition to model deployment and monitoring. You will work closely with cross-functional teams to deliver machine learning models that support business objectives and drive innovation. The ideal candidate should have a strong background in data analysis, feature engineering, and model selection, along with a deep understanding of model deployment and ongoing model maintenance.

Key Responsibilities:

  • Problem Definition: Collaborate with business stakeholders to define and structure data-driven problems. Translate business objectives into machine learning tasks (e.g., classification, regression, clustering).
  • Data Collection & Preprocessing: Gather, clean, and preprocess data from multiple sources (e.g., databases, APIs, publicly available datasets). Handle missing data, outliers, and apply normalization techniques.
  • Exploratory Data Analysis (EDA): Use statistical analysis and data visualization techniques to identify key patterns, trends, and correlations in the data.
  • Feature Engineering: Create, extract, and transform features to improve model performance. Apply techniques such as feature extraction, selection, and transformation.
  • Model Selection & Training: Select the appropriate machine learning models based on the problem at hand (e.g., supervised learning, unsupervised learning, deep learning). Train models using tools like Scikit-learn, TensorFlow, or PyTorch. Evaluate model performance using relevant metrics (e.g., RMSE, accuracy, F1-score, ROC-AUC) and optimize hyperparameters to ensure robustness. Deploy models in a production environment using tools like Flask, FastAPI, Docker, and Kubernetes. Ensure scalability and integration with existing systems.
  • Model Monitoring & Maintenance: Monitor model performance post-deployment, address model drift, and retrain models as needed. Ensure continuous accuracy and relevance of models in real-world scenarios.
  • Model Interpretation & Communication: Provide clear and actionable insights through model interpretation techniques such as feature importance and SHAP values. Present results to both technical and non-technical stakeholders.

Qualifications:

  • PhD degree in Computer Science, Data Science, Statistics, Engineering, or a related field.
  • 3+ years of experience in machine learning, statistical modeling, and data science.
  • Proficiency in Python, SQL, and experience with libraries such as Pandas, NumPy, Scikit-learn, TensorFlow, and Keras.
  • Hands-on experience with model deployment tools such as Flask, Docker, Kubernetes, and cloud platforms like AWS, Azure, or Google Cloud.
  • Strong knowledge of data preprocessing techniques, feature engineering, and exploratory data analysis.
  • Experience with hyperparameter tuning techniques (e.g., Grid Search, Bayesian Optimization).
  • Familiarity with model monitoring tools such as MLflow, Prometheus, or Grafana.
  • Excellent communication skills, with the ability to translate technical results into actionable insights for stakeholders.
  • Strong problem-solving skills and the ability to work on complex, data-driven projects.

Preferred Qualifications:

  • Experience with deep learning models (e.g., CNNs, RNNs, LSTMs).
  • Familiarity with NLP and time-series analysis.
  • Knowledge of big data tools like Spark or Hadoop.
  • Experience in sectors such as healthcare, finance, or e-commerce.

Location

United States - Remote

Remote Job

Job Overview
Job Posted:
3 months ago
Job Expires:
Job Type
Full Time

Share This Job: