Hypothesis Testing, T-Test, Z-Test, Regression (Linear, Logistic), Python/PySpark, SAS/SPSS, Statistical analysis and computing, Probabilistic Graph Models, Great Expectation, Evidently AI, Forecasting (Exponential Smoothing, ARIMA, ARIMAX), Tools(KubeFlow, BentoML), Classification (Decision Trees, SVM), ML Frameworks (TensorFlow, PyTorch, Sci-Kit Learn, CNTK, Keras, MXNet), Distance (Hamming Distance, Euclidean Distance, Manhattan Distance), R/ R Studio
Specialization
Data Science Advanced: AI/ML Engineer
Job requirements
Qualifications
Bachelor's or Master's degree in Computer Science, Data Science, or a related field with 3-5 years’ experience o Proven experience as a Data Engineer or similar role, with a focus on AI/ML projects.
Strong proficiency in Python programming and experience with relevant libraries and frameworks (e.g., pandas, NumPy, scikit-learn, TensorFlow, PyTorch).
Solid understanding of data engineering concepts, data modeling, and database systems (e.g., SQL, NoSQL).
Experience with data integration and ETL tools (e.g., Apache Airflow, Apache Spark, Talend). o Familiarity with cloud platforms (e.g., AWS, Azure, GCP) and related services (e.g., S3, EC2, BigQuery).
Strong problem-solving skills and ability to work in a fast-paced, collaborative environment.
Job Description & Responsibilities Data Pipeline Development:
Design, develop, and maintain scalable and efficient data pipelines to collect, clean, and transform large volumes of data.
Collaborate with software engineers, and other stakeholders to understand data requirements and implement effective solutions.
Ensure data pipelines are robust, reliable, and optimized for performance. Data Modeling and Integration:
Design and implement data models that support the storage, retrieval, and analysis of structured and unstructured data.
Integrate and consolidate data from various sources, both internal and external, to create a unified and comprehensive data ecosystem.
Ensure data integrity and accuracy through data quality assessments, cleansing, and validation techniques. Machine Learning Implementation:
Optimize and enhance machine learning algorithms for performance, scalability, and accuracy.
Implement data preprocessing, feature engineering, and model training workflows using Python and relevant libraries (e.g., scikit-learn, TensorFlow, PyTorch). Data Infrastructure Management
Configure and maintain cloud-based infrastructure for data storage, processing, and analysis.
Monitor and troubleshoot data-related issues, ensuring high availability and reliability of data systems.
Stay up-to-date with emerging technologies, tools, and best practices in data engineering, AI, and ML.
Tailor your resume for each Brillio role by matching the exact technology keywords from the job description — Azure Data Factory, Microsoft Fabric, Power BI, Snowflake, and PySpark appear across many current openings, and Lever's search algorithms rely on precise keyword matches.
Position yourself as a consultant, not just a technician — highlight client-facing experience, stakeholder communication, and business outcome delivery in your resume and interview responses, since Brillio's model demands both technical depth and consulting soft skills.
You're here early and that means a lot! ResumeGeni is in Beta which means it isn't 100% finished yet, but every day I'm working towards the vision. It's my goal to help you land your next job.
During the Beta, you'll get access to the state of the art AI to help find your perfect match, and when you do ResumeGeni will prepare your resume to get you to the interview.
There's still a lot to do. Please reach out if you have feedback, a request, or just want to send me a note.